file_name
large_stringlengths
4
140
prefix
large_stringlengths
0
12.1k
suffix
large_stringlengths
0
12k
middle
large_stringlengths
0
7.51k
fim_type
large_stringclasses
4 values
BratsCRFNeighborWhole.py
Counter from CRFUtils import * from BratsCheckPredictions import * import math from math import * train_path="/media/bmi/MyPassport/new_n4/Recon_2013_data/N4_zscore_training_t1_t1c_hist_match" test_path="/media/bmi/MyPassport/new_n4/Recon_2013_data/N4_zscore_testing_t1_t1c_hist_match" #train_path="/media/bmi/MyPassport/n4_entire/Recon_2013_data/training_longitudnal" #test_path="/media/bmi/MyPassport/n4_entire/Recon_2013_data/testing_longitudnal" #train_path="/home/rsk/Documents/PyStruct/data/train" #test_path="/home/rsk/Documents/PyStruct/data/test" #%% ################################################################################################# #Training the model def trainModel_Neighbor(num_iter=5,inference="qpbo",trainer="NSlack",num_train=2,num_test=1,C=0.1,edges="180x180_dist1_diag0",inputs=[1,1,1,1,1,1],features="post+img+pred",neighbor=-1,directed=False,savePred=False):
[trainLayers,trainTruth,sliceShape] = train print "Training slices loaded in %f" % (train_load_time) n_features= len(trainLayers[0][0,0]) print "Layer shape is : " print trainLayers[0].shape print "Training the model" edges= np.load("/home/bmi/CRF/edges/"+edges+".npy") G = [edges for x in trainLayers] print trainLayers[0].shape trainLayers = np.array( [x.reshape((sliceShape[0]*sliceShape[1],n_features)) for x in trainLayers] ) trainTruth = np.array( [x.reshape((sliceShape[0]*sliceShape[1],)).astype(int) for x in trainTruth] ) if inference=='ogm': crf = GraphCRF(inference_method=('ogm',{'alg':'fm'}),directed=directed) else: crf = GraphCRF(inference_method=inference,directed=directed) if trainer=="Frank": svm = FrankWolfeSSVM(model = crf,max_iter=num_iter,C=C,n_jobs=6,verbose=1) elif trainer=="NSlack": svm = NSlackSSVM(model = crf,max_iter=num_iter,C=C,n_jobs=-1,verbose=1) else: svm = OneSlackSSVM(model = crf,max_iter=num_iter,C=C,n_jobs=-1,verbose=1) start = time.clock() asdf = zip(trainLayers,G) svm.fit(asdf,trainTruth) end = time.clock() train_time = (end-start)/60.0 print "The training took %f" % (train_time) print "Model parameter size :" print svm.w.shape print "making predictions on train data" predTrain = svm.predict(asdf) trainDice=[] for i in range(len(trainLayers)): diceScore = accuracy(predTrain[i],trainTruth[i]) trainDice.append(diceScore) meanTrainDice = sum(trainDice)/len(trainLayers) del trainLayers,trainTruth ################################################################################################ overallDicePerPatient=[] # For overall test Dice extDicePerPatient=[] PatientTruthLayers=[] PatientPredLayers=[] PREC=[] RECALL=[] F1=[] LayerwiseDiceTotal=[] testResultFile = open(os.getcwd()+"/CRFResults/"+filename+".csv",'a') testResultFile.write("folderName,numLayers, Overall Dice, precision , recall, extDice"+"\n") counter=0 print "Loading the test slices" for folder in os.listdir(test_path): path = test_path + "/" + folder layerDiceScores='' data = extractTestSlices2(path,padding,neighbor=neighbor,inputs=inputs) if data!=0: [testLayers,testTruth,sliceShape,startSlice,endSlice] = data # trueTestLayers=testLayers GTest = [edges for x in testLayers] testLayers = np.array( [x.reshape((sliceShape[0]*sliceShape[1],n_features)) for x in testLayers] ) testTruth = np.array( [x.reshape((sliceShape[0]*sliceShape[1],)).astype(int) for x in testTruth] ) asdfTest = zip(testLayers,GTest) predTest = svm.predict(asdfTest) LayerwiseDice=[] for i in range(len(testLayers)): diceScore = accuracy(predTest[i],testTruth[i]) layerDiceScores+=","+str(diceScore) if math.isnan(diceScore): if sum(predTest[i])==0 and sum(testTruth[i])==0: LayerwiseDice.append(1.0) continue LayerwiseDice.append(diceScore) LayerwiseDiceTotal.append(LayerwiseDice) ### Imputing the predicted pixels into full volume if savePred==True: finalPatientPred = np.zeros((240,240,150)) finalPatientTruth = np.zeros((240,240,150)) predInsert = np.dstack(tuple([x.reshape(180,180) for x in predTest])) truthInsert = np.dstack(tuple([x.reshape(180,180) for x in testTruth])) finalPatientPred[30:(240-30),30:(240-30),startSlice:endSlice] = predInsert finalPatientTruth[30:(240-30),30:(240-30),startSlice:endSlice] = truthInsert finalPatientPred = finalPatientPred.astype('int') # print "saving at "+ path+"/"+filename+"whole" np.save(path+"/"+folder+filename+"whole",finalPatientPred) # print "predInsert shape" # print predInsert.shape # finalPatientPred = np.reshape(finalPatientPred,(240*240*150,)).astype('int') # finalPatientTruth = np.reshape(finalPatientTruth,(240*240*150,)).astype('int') # # print "Counters" # print Counter(list(np.hstack(testTruth))) # print Counter(list(finalPatientTruth)) # print confusion_matrix(np.hstack(predTest),np.hstack(testTruth)) # print confusion_matrix(finalPatientPred,finalPatientTruth) overallTestDice = accuracy(np.hstack(predTest),np.hstack(testTruth)) extDice = np.mean ( np.array(LayerwiseDice)[ range(10) + range(len(LayerwiseDice)-10, len(LayerwiseDice)) ] ) prec,recall,f1 = precision_score(np.hstack(testTruth),np.hstack(predTest)) , recall_score(np.hstack(testTruth),np.hstack(predTest)) , f1_score(np.hstack(testTruth),np.hstack(predTest)) print "Patient %d : Overall test DICE for %s is : %f and extDice is %f"%(counter,folder,overallTestDice,extDice) print "Precision : %f Recall : %f F1 : %f " %(prec,recall,f1) print "__________________________________________" # testResultFile.write(folder+","+str(len(testLayers))+","+str(meanTestDice)+","+str(overallTestDice) ","+str(np.max(testDice)) +","+ str(np.min(testDice))+"\n" ) testResultFile.write(folder+","+str(len(testLayers)) + ","+ str(overallTestDice) + ","+str(prec)+","+str(recall)+","+str(extDice)+layerDiceScores+"\n" ) overallDicePerPatient.append(overallTestDice) extDicePerPatient.append(extDice) PREC.append(prec), RECALL.append(recall) , F1.append(f1) PatientTruthLayers.append(testTruth) PatientPredLayers.append(predTest) counter+=1 if counter==num_test and num_test!=-1: break ###################################################################################################### print "Done testing slices" overallDice = sum(overallDicePerPatient)/len(PatientTruthLayers) overallPrec = sum(PREC)/len(PatientTruthLayers) overallRecall = sum(RECALL)/len(PatientTruthLayers) overallExtDice = np.mean(extDicePer
padding=(30,30,30,30) if directed==True: features +="+directed" resultsDir = os.getcwd()+'/CRFResults' nameLen = len(os.listdir(resultsDir)) edgeFeature = edges filename=str(nameLen)+"_CRF_iter_"+str(num_iter)+"_"+inference+"_"+trainer+"_"+features+"_"+str(num_train)+"_"+str(num_test)+"_"+edgeFeature print "Loading training slices" start = time.clock() train =extractSlices2(train_path,num_train,padding,neighbor=neighbor,inputs=inputs) end= time.clock() train_load_time = (end-start)/60.0
identifier_body
BratsCRFNeighborWhole.py
Counter from CRFUtils import * from BratsCheckPredictions import * import math from math import * train_path="/media/bmi/MyPassport/new_n4/Recon_2013_data/N4_zscore_training_t1_t1c_hist_match" test_path="/media/bmi/MyPassport/new_n4/Recon_2013_data/N4_zscore_testing_t1_t1c_hist_match" #train_path="/media/bmi/MyPassport/n4_entire/Recon_2013_data/training_longitudnal" #test_path="/media/bmi/MyPassport/n4_entire/Recon_2013_data/testing_longitudnal" #train_path="/home/rsk/Documents/PyStruct/data/train" #test_path="/home/rsk/Documents/PyStruct/data/test" #%% ################################################################################################# #Training the model def
(num_iter=5,inference="qpbo",trainer="NSlack",num_train=2,num_test=1,C=0.1,edges="180x180_dist1_diag0",inputs=[1,1,1,1,1,1],features="post+img+pred",neighbor=-1,directed=False,savePred=False): padding=(30,30,30,30) if directed==True: features +="+directed" resultsDir = os.getcwd()+'/CRFResults' nameLen = len(os.listdir(resultsDir)) edgeFeature = edges filename=str(nameLen)+"_CRF_iter_"+str(num_iter)+"_"+inference+"_"+trainer+"_"+features+"_"+str(num_train)+"_"+str(num_test)+"_"+edgeFeature print "Loading training slices" start = time.clock() train =extractSlices2(train_path,num_train,padding,neighbor=neighbor,inputs=inputs) end= time.clock() train_load_time = (end-start)/60.0 [trainLayers,trainTruth,sliceShape] = train print "Training slices loaded in %f" % (train_load_time) n_features= len(trainLayers[0][0,0]) print "Layer shape is : " print trainLayers[0].shape print "Training the model" edges= np.load("/home/bmi/CRF/edges/"+edges+".npy") G = [edges for x in trainLayers] print trainLayers[0].shape trainLayers = np.array( [x.reshape((sliceShape[0]*sliceShape[1],n_features)) for x in trainLayers] ) trainTruth = np.array( [x.reshape((sliceShape[0]*sliceShape[1],)).astype(int) for x in trainTruth] ) if inference=='ogm': crf = GraphCRF(inference_method=('ogm',{'alg':'fm'}),directed=directed) else: crf = GraphCRF(inference_method=inference,directed=directed) if trainer=="Frank": svm = FrankWolfeSSVM(model = crf,max_iter=num_iter,C=C,n_jobs=6,verbose=1) elif trainer=="NSlack": svm = NSlackSSVM(model = crf,max_iter=num_iter,C=C,n_jobs=-1,verbose=1) else: svm = OneSlackSSVM(model = crf,max_iter=num_iter,C=C,n_jobs=-1,verbose=1) start = time.clock() asdf = zip(trainLayers,G) svm.fit(asdf,trainTruth) end = time.clock() train_time = (end-start)/60.0 print "The training took %f" % (train_time) print "Model parameter size :" print svm.w.shape print "making predictions on train data" predTrain = svm.predict(asdf) trainDice=[] for i in range(len(trainLayers)): diceScore = accuracy(predTrain[i],trainTruth[i]) trainDice.append(diceScore) meanTrainDice = sum(trainDice)/len(trainLayers) del trainLayers,trainTruth ################################################################################################ overallDicePerPatient=[] # For overall test Dice extDicePerPatient=[] PatientTruthLayers=[] PatientPredLayers=[] PREC=[] RECALL=[] F1=[] LayerwiseDiceTotal=[] testResultFile = open(os.getcwd()+"/CRFResults/"+filename+".csv",'a') testResultFile.write("folderName,numLayers, Overall Dice, precision , recall, extDice"+"\n") counter=0 print "Loading the test slices" for folder in os.listdir(test_path): path = test_path + "/" + folder layerDiceScores='' data = extractTestSlices2(path,padding,neighbor=neighbor,inputs=inputs) if data!=0: [testLayers,testTruth,sliceShape,startSlice,endSlice] = data # trueTestLayers=testLayers GTest = [edges for x in testLayers] testLayers = np.array( [x.reshape((sliceShape[0]*sliceShape[1],n_features)) for x in testLayers] ) testTruth = np.array( [x.reshape((sliceShape[0]*sliceShape[1],)).astype(int) for x in testTruth] ) asdfTest = zip(testLayers,GTest) predTest = svm.predict(asdfTest) LayerwiseDice=[] for i in range(len(testLayers)): diceScore = accuracy(predTest[i],testTruth[i]) layerDiceScores+=","+str(diceScore) if math.isnan(diceScore): if sum(predTest[i])==0 and sum(testTruth[i])==0: LayerwiseDice.append(1.0) continue LayerwiseDice.append(diceScore) LayerwiseDiceTotal.append(LayerwiseDice) ### Imputing the predicted pixels into full volume if savePred==True: finalPatientPred = np.zeros((240,240,150)) finalPatientTruth = np.zeros((240,240,150)) predInsert = np.dstack(tuple([x.reshape(180,180) for x in predTest])) truthInsert = np.dstack(tuple([x.reshape(180,180) for x in testTruth])) finalPatientPred[30:(240-30),30:(240-30),startSlice:endSlice] = predInsert finalPatientTruth[30:(240-30),30:(240-30),startSlice:endSlice] = truthInsert finalPatientPred = finalPatientPred.astype('int') # print "saving at "+ path+"/"+filename+"whole" np.save(path+"/"+folder+filename+"whole",finalPatientPred) # print "predInsert shape" # print predInsert.shape # finalPatientPred = np.reshape(finalPatientPred,(240*240*150,)).astype('int') # finalPatientTruth = np.reshape(finalPatientTruth,(240*240*150,)).astype('int') # # print "Counters" # print Counter(list(np.hstack(testTruth))) # print Counter(list(finalPatientTruth)) # print confusion_matrix(np.hstack(predTest),np.hstack(testTruth)) # print confusion_matrix(finalPatientPred,finalPatientTruth) overallTestDice = accuracy(np.hstack(predTest),np.hstack(testTruth)) extDice = np.mean ( np.array(LayerwiseDice)[ range(10) + range(len(LayerwiseDice)-10, len(LayerwiseDice)) ] ) prec,recall,f1 = precision_score(np.hstack(testTruth),np.hstack(predTest)) , recall_score(np.hstack(testTruth),np.hstack(predTest)) , f1_score(np.hstack(testTruth),np.hstack(predTest)) print "Patient %d : Overall test DICE for %s is : %f and extDice is %f"%(counter,folder,overallTestDice,extDice) print "Precision : %f Recall : %f F1 : %f " %(prec,recall,f1) print "__________________________________________" # testResultFile.write(folder+","+str(len(testLayers))+","+str(meanTestDice)+","+str(overallTestDice) ","+str(np.max(testDice)) +","+ str(np.min(testDice))+"\n" ) testResultFile.write(folder+","+str(len(testLayers)) + ","+ str(overallTestDice) + ","+str(prec)+","+str(recall)+","+str(extDice)+layerDiceScores+"\n" ) overallDicePerPatient.append(overallTestDice) extDicePerPatient.append(extDice) PREC.append(prec), RECALL.append(recall) , F1.append(f1) PatientTruthLayers.append(testTruth) PatientPredLayers.append(predTest) counter+=1 if counter==num_test and num_test!=-1: break ###################################################################################################### print "Done testing slices" overallDice = sum(overallDicePerPatient)/len(PatientTruthLayers) overallPrec = sum(PREC)/len(PatientTruthLayers) overallRecall = sum(RECALL)/len(PatientTruthLayers) overallExtDice = np.mean(extDicePer
trainModel_Neighbor
identifier_name
BratsCRFNeighborWhole.py
Counter from CRFUtils import * from BratsCheckPredictions import * import math from math import * train_path="/media/bmi/MyPassport/new_n4/Recon_2013_data/N4_zscore_training_t1_t1c_hist_match" test_path="/media/bmi/MyPassport/new_n4/Recon_2013_data/N4_zscore_testing_t1_t1c_hist_match" #train_path="/media/bmi/MyPassport/n4_entire/Recon_2013_data/training_longitudnal" #test_path="/media/bmi/MyPassport/n4_entire/Recon_2013_data/testing_longitudnal" #train_path="/home/rsk/Documents/PyStruct/data/train" #test_path="/home/rsk/Documents/PyStruct/data/test" #%% ################################################################################################# #Training the model def trainModel_Neighbor(num_iter=5,inference="qpbo",trainer="NSlack",num_train=2,num_test=1,C=0.1,edges="180x180_dist1_diag0",inputs=[1,1,1,1,1,1],features="post+img+pred",neighbor=-1,directed=False,savePred=False): padding=(30,30,30,30) if directed==True: features +="+directed" resultsDir = os.getcwd()+'/CRFResults' nameLen = len(os.listdir(resultsDir)) edgeFeature = edges filename=str(nameLen)+"_CRF_iter_"+str(num_iter)+"_"+inference+"_"+trainer+"_"+features+"_"+str(num_train)+"_"+str(num_test)+"_"+edgeFeature print "Loading training slices" start = time.clock() train =extractSlices2(train_path,num_train,padding,neighbor=neighbor,inputs=inputs) end= time.clock() train_load_time = (end-start)/60.0 [trainLayers,trainTruth,sliceShape] = train print "Training slices loaded in %f" % (train_load_time) n_features= len(trainLayers[0][0,0]) print "Layer shape is : " print trainLayers[0].shape print "Training the model" edges= np.load("/home/bmi/CRF/edges/"+edges+".npy") G = [edges for x in trainLayers] print trainLayers[0].shape trainLayers = np.array( [x.reshape((sliceShape[0]*sliceShape[1],n_features)) for x in trainLayers] ) trainTruth = np.array( [x.reshape((sliceShape[0]*sliceShape[1],)).astype(int) for x in trainTruth] ) if inference=='ogm': crf = GraphCRF(inference_method=('ogm',{'alg':'fm'}),directed=directed) else:
if trainer=="Frank": svm = FrankWolfeSSVM(model = crf,max_iter=num_iter,C=C,n_jobs=6,verbose=1) elif trainer=="NSlack": svm = NSlackSSVM(model = crf,max_iter=num_iter,C=C,n_jobs=-1,verbose=1) else: svm = OneSlackSSVM(model = crf,max_iter=num_iter,C=C,n_jobs=-1,verbose=1) start = time.clock() asdf = zip(trainLayers,G) svm.fit(asdf,trainTruth) end = time.clock() train_time = (end-start)/60.0 print "The training took %f" % (train_time) print "Model parameter size :" print svm.w.shape print "making predictions on train data" predTrain = svm.predict(asdf) trainDice=[] for i in range(len(trainLayers)): diceScore = accuracy(predTrain[i],trainTruth[i]) trainDice.append(diceScore) meanTrainDice = sum(trainDice)/len(trainLayers) del trainLayers,trainTruth ################################################################################################ overallDicePerPatient=[] # For overall test Dice extDicePerPatient=[] PatientTruthLayers=[] PatientPredLayers=[] PREC=[] RECALL=[] F1=[] LayerwiseDiceTotal=[] testResultFile = open(os.getcwd()+"/CRFResults/"+filename+".csv",'a') testResultFile.write("folderName,numLayers, Overall Dice, precision , recall, extDice"+"\n") counter=0 print "Loading the test slices" for folder in os.listdir(test_path): path = test_path + "/" + folder layerDiceScores='' data = extractTestSlices2(path,padding,neighbor=neighbor,inputs=inputs) if data!=0: [testLayers,testTruth,sliceShape,startSlice,endSlice] = data # trueTestLayers=testLayers GTest = [edges for x in testLayers] testLayers = np.array( [x.reshape((sliceShape[0]*sliceShape[1],n_features)) for x in testLayers] ) testTruth = np.array( [x.reshape((sliceShape[0]*sliceShape[1],)).astype(int) for x in testTruth] ) asdfTest = zip(testLayers,GTest) predTest = svm.predict(asdfTest) LayerwiseDice=[] for i in range(len(testLayers)): diceScore = accuracy(predTest[i],testTruth[i]) layerDiceScores+=","+str(diceScore) if math.isnan(diceScore): if sum(predTest[i])==0 and sum(testTruth[i])==0: LayerwiseDice.append(1.0) continue LayerwiseDice.append(diceScore) LayerwiseDiceTotal.append(LayerwiseDice) ### Imputing the predicted pixels into full volume if savePred==True: finalPatientPred = np.zeros((240,240,150)) finalPatientTruth = np.zeros((240,240,150)) predInsert = np.dstack(tuple([x.reshape(180,180) for x in predTest])) truthInsert = np.dstack(tuple([x.reshape(180,180) for x in testTruth])) finalPatientPred[30:(240-30),30:(240-30),startSlice:endSlice] = predInsert finalPatientTruth[30:(240-30),30:(240-30),startSlice:endSlice] = truthInsert finalPatientPred = finalPatientPred.astype('int') # print "saving at "+ path+"/"+filename+"whole" np.save(path+"/"+folder+filename+"whole",finalPatientPred) # print "predInsert shape" # print predInsert.shape # finalPatientPred = np.reshape(finalPatientPred,(240*240*150,)).astype('int') # finalPatientTruth = np.reshape(finalPatientTruth,(240*240*150,)).astype('int') # # print "Counters" # print Counter(list(np.hstack(testTruth))) # print Counter(list(finalPatientTruth)) # print confusion_matrix(np.hstack(predTest),np.hstack(testTruth)) # print confusion_matrix(finalPatientPred,finalPatientTruth) overallTestDice = accuracy(np.hstack(predTest),np.hstack(testTruth)) extDice = np.mean ( np.array(LayerwiseDice)[ range(10) + range(len(LayerwiseDice)-10, len(LayerwiseDice)) ] ) prec,recall,f1 = precision_score(np.hstack(testTruth),np.hstack(predTest)) , recall_score(np.hstack(testTruth),np.hstack(predTest)) , f1_score(np.hstack(testTruth),np.hstack(predTest)) print "Patient %d : Overall test DICE for %s is : %f and extDice is %f"%(counter,folder,overallTestDice,extDice) print "Precision : %f Recall : %f F1 : %f " %(prec,recall,f1) print "__________________________________________" # testResultFile.write(folder+","+str(len(testLayers))+","+str(meanTestDice)+","+str(overallTestDice) ","+str(np.max(testDice)) +","+ str(np.min(testDice))+"\n" ) testResultFile.write(folder+","+str(len(testLayers)) + ","+ str(overallTestDice) + ","+str(prec)+","+str(recall)+","+str(extDice)+layerDiceScores+"\n" ) overallDicePerPatient.append(overallTestDice) extDicePerPatient.append(extDice) PREC.append(prec), RECALL.append(recall) , F1.append(f1) PatientTruthLayers.append(testTruth) PatientPredLayers.append(predTest) counter+=1 if counter==num_test and num_test!=-1: break ###################################################################################################### print "Done testing slices" overallDice = sum(overallDicePerPatient)/len(PatientTruthLayers) overallPrec = sum(PREC)/len(PatientTruthLayers) overallRecall = sum(RECALL)/len(PatientTruthLayers) overallExtDice = np.mean(extDicePerPatient
crf = GraphCRF(inference_method=inference,directed=directed)
conditional_block
playlist.py
.get('placement_filter') ## audio settings ## # true if content responds to audio input self.sound_reactive = kwargs.get('sound_reactive', False) # if sound_reactive, false if content can still be shown without audio input available self.sound_required = kwargs.get('sound_required', True) # relative volume adjustment for audio input (to give more/less responsiveness) self.volume_adjust = kwargs.get('volume_adjust', 1.) # true if sketch has audio out (that we actually want to hear) self.has_audio = kwargs.get('has_audio', False) # true if content can use kinect self.kinect_enabled = kwargs.get('kinect_enabled', False) # if kinect_enabled, false if content still works without connect self.kinect_required = kwargs.get('kinect_required', True) self.server_side_parameters = kwargs.get('server_side_parameters', []) ## sketch-dependent parameters ## # video # length of video in seconds -- set automatically self.duration = self.get_video_duration() # how to play the video: # - 'shuffle': play a random excerpt for the specific runtime # - 'full': play the video start to finish self.play_mode = kwargs.get('play_mode', 'shuffle') # screencast # command to launch program to be cast self.cmdline = kwargs.get('cmdline') # a hook to further configure/interact with the program after it's launched self.post_launch = kwargs.get('post_launch') if set(kwargs.keys()) - set(self.__dict__): assert False, 'unrecognized arg' def get_video_duration(self): if not self.sketch == 'video': return vid = self.params['path'] try: duration = float(os.popen('mediainfo --Inform="Video;%%Duration%%" "%s"' % vid).readlines()[0].strip())/1000. except RuntimeError: print 'could not read duration of %s' % vid duration = 0 return duration def to_json_info(self): info = dict((k, getattr(self, k)) for k in ('name', 'sound_reactive', 'has_audio', 'kinect_enabled')) if self.play_mode == 'full': info['duration'] = self.duration return info # placement filter that ensures crisp alignment with lsdome panel/pixel geometry def pixel_exact(p): # forcing zero-rotation is sufficient for lsdome, and achieves the same spirit for prometheus (which # doesn't have a concept of 'pixel exact') while still allowing for some variation in wing overlap return getattr(p, 'rot', 0) == 0 and p.is_1to1 # like pixel_exact, but stretch to fit full canvas (so not 'exact', but still 'aligned') def align_but_stretch(p): return getattr(p, 'rot', 0) == 0 and p.stretch _all_content = None def all_content(): global _all_content if not _all_content: _all_content = [ Content('black', '[util] black (note: keeps running and using cpu)', manual=True), Content('gridtest', '[util] uvw grid test', geometries=['lsdome'], manual=True), Content('fctest', '[util] fc topology test', params={'fcconfig': fadecandy_config()}), Content('layouttest', '[util] cartesian test (mouse)', manual=True, placement_filter=pixel_exact), Content('binary', '[util] binary decomp', manual=True), Content('cloud'), Content('dontknow'), Content('harmonics', geometries=['lsdome']), Content('moire'), Content('rings'), Content('tube'), Content('twinkle'), Content('fft', sound_reactive=True), Content('pixelflock', sound_reactive=True, sound_required=False, kinect_enabled=True, kinect_required=False), Content('kinectdepth', 'kinectdepth', kinect_enabled=True, placement_filter=align_but_stretch), Content('screencast', 'projectm', cmdline='projectM-pulseaudio', sound_reactive=True, volume_adjust=1.5, server_side_parameters=projectm_parameters(), post_launch=lambda manager: projectm_control(manager, 'next'), # get off the default pattern ), Content('stream', 'hdmi-in', manual=True, stretch_aspect=True, params={ 'camera': 'FHD Capture: FHD Capture', }), Content('kaleidoscope', geometries=['lsdome'], placement_filter=pixel_exact, params={'scale': 2.}), Content('kaleidoscope', geometries=['prometheus'], params={'scale': 3.2}), Content('imgkaleidoscope', 'hearts', geometries=['lsdome'], placement_filter=pixel_exact, params={ 'image': "res/img/hearts.jpg", 'scale': 1., 'source_scale': 1.3, 'speed': .25, }), Content('video', 'video:chrissy_poi_zoom', geometries=['lsdome'], params={ 'path': os.path.join(VIDEO_DIR, 'hayley_chrissy_fire_spinning.mp4'), }, placement_filter=lambda p: p.name == 'poi (01-10 21:44)'), ] _all_content.extend(load_videos()) _all_content = [c for c in _all_content if not c.geometries or settings.geometry in c.geometries] _all_content = [c for c in _all_content if not (c.kinect_enabled and c.kinect_required) or settings.kinect] for c in _all_content: if c.kinect_enabled and settings.kinect and not c.placement_filter: # when kinect used, ensure display lines up with camera c.placement_filter = pixel_exact assert len(set(c.name for c in _all_content)) == len(_all_content), 'content names not unique' _all_content = dict((c.name, c) for c in _all_content) return _all_content def load_videos(): vids = [f.strip() for f in os.popen('find "%s" -type f' % VIDEO_DIR).readlines()] for vid in vids: # TODO placement restrictions? joan of arc require mirror mode? # do special things for certain videos -- should probably make this more maintainable args = {} if 'knife' in vid: args['play_mode'] = 'full' if any(k in vid for k in ('knife', 'flood')): args['has_audio'] = True yield Content('video', 'video:%s' % os.path.relpath(vid, VIDEO_DIR), stretch_aspect=True, params={ 'path': vid, }, **args) def fadecandy_config(): if settings.geometry == 'lsdome': fcconfig = 'lsdome_%spanel.json' % settings.num_panels elif settings.geometry == 'prometheus': fcconfig = 'prometheus_wing.json' return os.path.join(settings.repo_root, 'src/config/fadecandy', fcconfig) def projectm_control(mgr, command): interaction = { 'next': 'key r', 'toggle-lock': 'key l', }[command] launch.gui_interaction(mgr.content.window_id, interaction) def projectm_parameters():
def game_content(rom): try: args = launch.launch_emulator(rom) except: return None name = os.path.splitext(os.path.relpath(os.path.abspath(rom), settings.roms_path))[0] return Content('screencast', name, cmdline=args['cmd'], params=args.get('params', {}), stretch_aspect=True, has_audio=True) _games_content = None def load_games(filt): def all_roms_path_files(): for dirpath, _, filenames in os.walk(settings.roms_path): for f in filenames: yield os.path.join(dirpath, f) global _games_content if not _games_content: _games_content = filter(None, map(game_content, all_roms_path_files())) _games_content = dict((c.name, c) for c in _games_content) print len(_games_content), 'roms' if filt == 'favs': return filter_games_favorites(_games_content) else: return filter_games(_games_content, filt) def filter_games(all_games, filt): def name_to_search_key(name): name = os.path.split(name)[1] name = name.split('(')[0] words = name.lower().split() words = [re.sub('[^a-z0-9]', '', w) for w in words] return filter(None, words) def match_key(query, key): return all(any(kw.startswith(qw) for
import animations class ProjectMNextPatternAction(animations.Parameter): def param_def(self): return { 'name': 'next pattern', 'isAction': True, } def handle_input_event(self, type, val): if type != 'press': return projectm_control(self.manager, 'next') def _update_value(self, val): pass return [ProjectMNextPatternAction]
identifier_body
playlist.py
.get('placement_filter') ## audio settings ## # true if content responds to audio input self.sound_reactive = kwargs.get('sound_reactive', False) # if sound_reactive, false if content can still be shown without audio input available self.sound_required = kwargs.get('sound_required', True) # relative volume adjustment for audio input (to give more/less responsiveness) self.volume_adjust = kwargs.get('volume_adjust', 1.) # true if sketch has audio out (that we actually want to hear) self.has_audio = kwargs.get('has_audio', False) # true if content can use kinect self.kinect_enabled = kwargs.get('kinect_enabled', False) # if kinect_enabled, false if content still works without connect self.kinect_required = kwargs.get('kinect_required', True) self.server_side_parameters = kwargs.get('server_side_parameters', []) ## sketch-dependent parameters ## # video # length of video in seconds -- set automatically self.duration = self.get_video_duration() # how to play the video: # - 'shuffle': play a random excerpt for the specific runtime # - 'full': play the video start to finish self.play_mode = kwargs.get('play_mode', 'shuffle') # screencast # command to launch program to be cast self.cmdline = kwargs.get('cmdline') # a hook to further configure/interact with the program after it's launched self.post_launch = kwargs.get('post_launch') if set(kwargs.keys()) - set(self.__dict__): assert False, 'unrecognized arg' def get_video_duration(self): if not self.sketch == 'video': return vid = self.params['path'] try: duration = float(os.popen('mediainfo --Inform="Video;%%Duration%%" "%s"' % vid).readlines()[0].strip())/1000. except RuntimeError: print 'could not read duration of %s' % vid duration = 0 return duration def to_json_info(self): info = dict((k, getattr(self, k)) for k in ('name', 'sound_reactive', 'has_audio', 'kinect_enabled')) if self.play_mode == 'full': info['duration'] = self.duration return info # placement filter that ensures crisp alignment with lsdome panel/pixel geometry def pixel_exact(p): # forcing zero-rotation is sufficient for lsdome, and achieves the same spirit for prometheus (which # doesn't have a concept of 'pixel exact') while still allowing for some variation in wing overlap return getattr(p, 'rot', 0) == 0 and p.is_1to1 # like pixel_exact, but stretch to fit full canvas (so not 'exact', but still 'aligned') def align_but_stretch(p): return getattr(p, 'rot', 0) == 0 and p.stretch _all_content = None def all_content(): global _all_content if not _all_content: _all_content = [ Content('black', '[util] black (note: keeps running and using cpu)', manual=True), Content('gridtest', '[util] uvw grid test', geometries=['lsdome'], manual=True), Content('fctest', '[util] fc topology test', params={'fcconfig': fadecandy_config()}), Content('layouttest', '[util] cartesian test (mouse)', manual=True, placement_filter=pixel_exact), Content('binary', '[util] binary decomp', manual=True), Content('cloud'), Content('dontknow'), Content('harmonics', geometries=['lsdome']), Content('moire'), Content('rings'), Content('tube'), Content('twinkle'), Content('fft', sound_reactive=True), Content('pixelflock', sound_reactive=True, sound_required=False, kinect_enabled=True, kinect_required=False), Content('kinectdepth', 'kinectdepth', kinect_enabled=True, placement_filter=align_but_stretch), Content('screencast', 'projectm', cmdline='projectM-pulseaudio', sound_reactive=True, volume_adjust=1.5, server_side_parameters=projectm_parameters(), post_launch=lambda manager: projectm_control(manager, 'next'), # get off the default pattern ),
Content('kaleidoscope', geometries=['lsdome'], placement_filter=pixel_exact, params={'scale': 2.}), Content('kaleidoscope', geometries=['prometheus'], params={'scale': 3.2}), Content('imgkaleidoscope', 'hearts', geometries=['lsdome'], placement_filter=pixel_exact, params={ 'image': "res/img/hearts.jpg", 'scale': 1., 'source_scale': 1.3, 'speed': .25, }), Content('video', 'video:chrissy_poi_zoom', geometries=['lsdome'], params={ 'path': os.path.join(VIDEO_DIR, 'hayley_chrissy_fire_spinning.mp4'), }, placement_filter=lambda p: p.name == 'poi (01-10 21:44)'), ] _all_content.extend(load_videos()) _all_content = [c for c in _all_content if not c.geometries or settings.geometry in c.geometries] _all_content = [c for c in _all_content if not (c.kinect_enabled and c.kinect_required) or settings.kinect] for c in _all_content: if c.kinect_enabled and settings.kinect and not c.placement_filter: # when kinect used, ensure display lines up with camera c.placement_filter = pixel_exact assert len(set(c.name for c in _all_content)) == len(_all_content), 'content names not unique' _all_content = dict((c.name, c) for c in _all_content) return _all_content def load_videos(): vids = [f.strip() for f in os.popen('find "%s" -type f' % VIDEO_DIR).readlines()] for vid in vids: # TODO placement restrictions? joan of arc require mirror mode? # do special things for certain videos -- should probably make this more maintainable args = {} if 'knife' in vid: args['play_mode'] = 'full' if any(k in vid for k in ('knife', 'flood')): args['has_audio'] = True yield Content('video', 'video:%s' % os.path.relpath(vid, VIDEO_DIR), stretch_aspect=True, params={ 'path': vid, }, **args) def fadecandy_config(): if settings.geometry == 'lsdome': fcconfig = 'lsdome_%spanel.json' % settings.num_panels elif settings.geometry == 'prometheus': fcconfig = 'prometheus_wing.json' return os.path.join(settings.repo_root, 'src/config/fadecandy', fcconfig) def projectm_control(mgr, command): interaction = { 'next': 'key r', 'toggle-lock': 'key l', }[command] launch.gui_interaction(mgr.content.window_id, interaction) def projectm_parameters(): import animations class ProjectMNextPatternAction(animations.Parameter): def param_def(self): return { 'name': 'next pattern', 'isAction': True, } def handle_input_event(self, type, val): if type != 'press': return projectm_control(self.manager, 'next') def _update_value(self, val): pass return [ProjectMNextPatternAction] def game_content(rom): try: args = launch.launch_emulator(rom) except: return None name = os.path.splitext(os.path.relpath(os.path.abspath(rom), settings.roms_path))[0] return Content('screencast', name, cmdline=args['cmd'], params=args.get('params', {}), stretch_aspect=True, has_audio=True) _games_content = None def load_games(filt): def all_roms_path_files(): for dirpath, _, filenames in os.walk(settings.roms_path): for f in filenames: yield os.path.join(dirpath, f) global _games_content if not _games_content: _games_content = filter(None, map(game_content, all_roms_path_files())) _games_content = dict((c.name, c) for c in _games_content) print len(_games_content), 'roms' if filt == 'favs': return filter_games_favorites(_games_content) else: return filter_games(_games_content, filt) def filter_games(all_games, filt): def name_to_search_key(name): name = os.path.split(name)[1] name = name.split('(')[0] words = name.lower().split() words = [re.sub('[^a-z0-9]', '', w) for w in words] return filter(None, words) def match_key(query, key): return all(any(kw.startswith(qw) for kw
Content('stream', 'hdmi-in', manual=True, stretch_aspect=True, params={ 'camera': 'FHD Capture: FHD Capture', }),
random_line_split
playlist.py
kwargs.get('placement_filter') ## audio settings ## # true if content responds to audio input self.sound_reactive = kwargs.get('sound_reactive', False) # if sound_reactive, false if content can still be shown without audio input available self.sound_required = kwargs.get('sound_required', True) # relative volume adjustment for audio input (to give more/less responsiveness) self.volume_adjust = kwargs.get('volume_adjust', 1.) # true if sketch has audio out (that we actually want to hear) self.has_audio = kwargs.get('has_audio', False) # true if content can use kinect self.kinect_enabled = kwargs.get('kinect_enabled', False) # if kinect_enabled, false if content still works without connect self.kinect_required = kwargs.get('kinect_required', True) self.server_side_parameters = kwargs.get('server_side_parameters', []) ## sketch-dependent parameters ## # video # length of video in seconds -- set automatically self.duration = self.get_video_duration() # how to play the video: # - 'shuffle': play a random excerpt for the specific runtime # - 'full': play the video start to finish self.play_mode = kwargs.get('play_mode', 'shuffle') # screencast # command to launch program to be cast self.cmdline = kwargs.get('cmdline') # a hook to further configure/interact with the program after it's launched self.post_launch = kwargs.get('post_launch') if set(kwargs.keys()) - set(self.__dict__): assert False, 'unrecognized arg' def get_video_duration(self): if not self.sketch == 'video': return vid = self.params['path'] try: duration = float(os.popen('mediainfo --Inform="Video;%%Duration%%" "%s"' % vid).readlines()[0].strip())/1000. except RuntimeError: print 'could not read duration of %s' % vid duration = 0 return duration def to_json_info(self): info = dict((k, getattr(self, k)) for k in ('name', 'sound_reactive', 'has_audio', 'kinect_enabled')) if self.play_mode == 'full': info['duration'] = self.duration return info # placement filter that ensures crisp alignment with lsdome panel/pixel geometry def pixel_exact(p): # forcing zero-rotation is sufficient for lsdome, and achieves the same spirit for prometheus (which # doesn't have a concept of 'pixel exact') while still allowing for some variation in wing overlap return getattr(p, 'rot', 0) == 0 and p.is_1to1 # like pixel_exact, but stretch to fit full canvas (so not 'exact', but still 'aligned') def align_but_stretch(p): return getattr(p, 'rot', 0) == 0 and p.stretch _all_content = None def all_content(): global _all_content if not _all_content:
), Content('stream', 'hdmi-in', manual=True, stretch_aspect=True, params={ 'camera': 'FHD Capture: FHD Capture', }), Content('kaleidoscope', geometries=['lsdome'], placement_filter=pixel_exact, params={'scale': 2.}), Content('kaleidoscope', geometries=['prometheus'], params={'scale': 3.2}), Content('imgkaleidoscope', 'hearts', geometries=['lsdome'], placement_filter=pixel_exact, params={ 'image': "res/img/hearts.jpg", 'scale': 1., 'source_scale': 1.3, 'speed': .25, }), Content('video', 'video:chrissy_poi_zoom', geometries=['lsdome'], params={ 'path': os.path.join(VIDEO_DIR, 'hayley_chrissy_fire_spinning.mp4'), }, placement_filter=lambda p: p.name == 'poi (01-10 21:44)'), ] _all_content.extend(load_videos()) _all_content = [c for c in _all_content if not c.geometries or settings.geometry in c.geometries] _all_content = [c for c in _all_content if not (c.kinect_enabled and c.kinect_required) or settings.kinect] for c in _all_content: if c.kinect_enabled and settings.kinect and not c.placement_filter: # when kinect used, ensure display lines up with camera c.placement_filter = pixel_exact assert len(set(c.name for c in _all_content)) == len(_all_content), 'content names not unique' _all_content = dict((c.name, c) for c in _all_content) return _all_content def load_videos(): vids = [f.strip() for f in os.popen('find "%s" -type f' % VIDEO_DIR).readlines()] for vid in vids: # TODO placement restrictions? joan of arc require mirror mode? # do special things for certain videos -- should probably make this more maintainable args = {} if 'knife' in vid: args['play_mode'] = 'full' if any(k in vid for k in ('knife', 'flood')): args['has_audio'] = True yield Content('video', 'video:%s' % os.path.relpath(vid, VIDEO_DIR), stretch_aspect=True, params={ 'path': vid, }, **args) def fadecandy_config(): if settings.geometry == 'lsdome': fcconfig = 'lsdome_%spanel.json' % settings.num_panels elif settings.geometry == 'prometheus': fcconfig = 'prometheus_wing.json' return os.path.join(settings.repo_root, 'src/config/fadecandy', fcconfig) def projectm_control(mgr, command): interaction = { 'next': 'key r', 'toggle-lock': 'key l', }[command] launch.gui_interaction(mgr.content.window_id, interaction) def projectm_parameters(): import animations class ProjectMNextPatternAction(animations.Parameter): def param_def(self): return { 'name': 'next pattern', 'isAction': True, } def handle_input_event(self, type, val): if type != 'press': return projectm_control(self.manager, 'next') def _update_value(self, val): pass return [ProjectMNextPatternAction] def game_content(rom): try: args = launch.launch_emulator(rom) except: return None name = os.path.splitext(os.path.relpath(os.path.abspath(rom), settings.roms_path))[0] return Content('screencast', name, cmdline=args['cmd'], params=args.get('params', {}), stretch_aspect=True, has_audio=True) _games_content = None def load_games(filt): def all_roms_path_files(): for dirpath, _, filenames in os.walk(settings.roms_path): for f in filenames: yield os.path.join(dirpath, f) global _games_content if not _games_content: _games_content = filter(None, map(game_content, all_roms_path_files())) _games_content = dict((c.name, c) for c in _games_content) print len(_games_content), 'roms' if filt == 'favs': return filter_games_favorites(_games_content) else: return filter_games(_games_content, filt) def filter_games(all_games, filt): def name_to_search_key(name): name = os.path.split(name)[1] name = name.split('(')[0] words = name.lower().split() words = [re.sub('[^a-z0-9]', '', w) for w in words] return filter(None, words) def match_key(query, key): return all(any(kw.startswith(qw) for kw in
_all_content = [ Content('black', '[util] black (note: keeps running and using cpu)', manual=True), Content('gridtest', '[util] uvw grid test', geometries=['lsdome'], manual=True), Content('fctest', '[util] fc topology test', params={'fcconfig': fadecandy_config()}), Content('layouttest', '[util] cartesian test (mouse)', manual=True, placement_filter=pixel_exact), Content('binary', '[util] binary decomp', manual=True), Content('cloud'), Content('dontknow'), Content('harmonics', geometries=['lsdome']), Content('moire'), Content('rings'), Content('tube'), Content('twinkle'), Content('fft', sound_reactive=True), Content('pixelflock', sound_reactive=True, sound_required=False, kinect_enabled=True, kinect_required=False), Content('kinectdepth', 'kinectdepth', kinect_enabled=True, placement_filter=align_but_stretch), Content('screencast', 'projectm', cmdline='projectM-pulseaudio', sound_reactive=True, volume_adjust=1.5, server_side_parameters=projectm_parameters(), post_launch=lambda manager: projectm_control(manager, 'next'), # get off the default pattern
conditional_block
playlist.py
sdome, and achieves the same spirit for prometheus (which # doesn't have a concept of 'pixel exact') while still allowing for some variation in wing overlap return getattr(p, 'rot', 0) == 0 and p.is_1to1 # like pixel_exact, but stretch to fit full canvas (so not 'exact', but still 'aligned') def align_but_stretch(p): return getattr(p, 'rot', 0) == 0 and p.stretch _all_content = None def all_content(): global _all_content if not _all_content: _all_content = [ Content('black', '[util] black (note: keeps running and using cpu)', manual=True), Content('gridtest', '[util] uvw grid test', geometries=['lsdome'], manual=True), Content('fctest', '[util] fc topology test', params={'fcconfig': fadecandy_config()}), Content('layouttest', '[util] cartesian test (mouse)', manual=True, placement_filter=pixel_exact), Content('binary', '[util] binary decomp', manual=True), Content('cloud'), Content('dontknow'), Content('harmonics', geometries=['lsdome']), Content('moire'), Content('rings'), Content('tube'), Content('twinkle'), Content('fft', sound_reactive=True), Content('pixelflock', sound_reactive=True, sound_required=False, kinect_enabled=True, kinect_required=False), Content('kinectdepth', 'kinectdepth', kinect_enabled=True, placement_filter=align_but_stretch), Content('screencast', 'projectm', cmdline='projectM-pulseaudio', sound_reactive=True, volume_adjust=1.5, server_side_parameters=projectm_parameters(), post_launch=lambda manager: projectm_control(manager, 'next'), # get off the default pattern ), Content('stream', 'hdmi-in', manual=True, stretch_aspect=True, params={ 'camera': 'FHD Capture: FHD Capture', }), Content('kaleidoscope', geometries=['lsdome'], placement_filter=pixel_exact, params={'scale': 2.}), Content('kaleidoscope', geometries=['prometheus'], params={'scale': 3.2}), Content('imgkaleidoscope', 'hearts', geometries=['lsdome'], placement_filter=pixel_exact, params={ 'image': "res/img/hearts.jpg", 'scale': 1., 'source_scale': 1.3, 'speed': .25, }), Content('video', 'video:chrissy_poi_zoom', geometries=['lsdome'], params={ 'path': os.path.join(VIDEO_DIR, 'hayley_chrissy_fire_spinning.mp4'), }, placement_filter=lambda p: p.name == 'poi (01-10 21:44)'), ] _all_content.extend(load_videos()) _all_content = [c for c in _all_content if not c.geometries or settings.geometry in c.geometries] _all_content = [c for c in _all_content if not (c.kinect_enabled and c.kinect_required) or settings.kinect] for c in _all_content: if c.kinect_enabled and settings.kinect and not c.placement_filter: # when kinect used, ensure display lines up with camera c.placement_filter = pixel_exact assert len(set(c.name for c in _all_content)) == len(_all_content), 'content names not unique' _all_content = dict((c.name, c) for c in _all_content) return _all_content def load_videos(): vids = [f.strip() for f in os.popen('find "%s" -type f' % VIDEO_DIR).readlines()] for vid in vids: # TODO placement restrictions? joan of arc require mirror mode? # do special things for certain videos -- should probably make this more maintainable args = {} if 'knife' in vid: args['play_mode'] = 'full' if any(k in vid for k in ('knife', 'flood')): args['has_audio'] = True yield Content('video', 'video:%s' % os.path.relpath(vid, VIDEO_DIR), stretch_aspect=True, params={ 'path': vid, }, **args) def fadecandy_config(): if settings.geometry == 'lsdome': fcconfig = 'lsdome_%spanel.json' % settings.num_panels elif settings.geometry == 'prometheus': fcconfig = 'prometheus_wing.json' return os.path.join(settings.repo_root, 'src/config/fadecandy', fcconfig) def projectm_control(mgr, command): interaction = { 'next': 'key r', 'toggle-lock': 'key l', }[command] launch.gui_interaction(mgr.content.window_id, interaction) def projectm_parameters(): import animations class ProjectMNextPatternAction(animations.Parameter): def param_def(self): return { 'name': 'next pattern', 'isAction': True, } def handle_input_event(self, type, val): if type != 'press': return projectm_control(self.manager, 'next') def _update_value(self, val): pass return [ProjectMNextPatternAction] def game_content(rom): try: args = launch.launch_emulator(rom) except: return None name = os.path.splitext(os.path.relpath(os.path.abspath(rom), settings.roms_path))[0] return Content('screencast', name, cmdline=args['cmd'], params=args.get('params', {}), stretch_aspect=True, has_audio=True) _games_content = None def load_games(filt): def all_roms_path_files(): for dirpath, _, filenames in os.walk(settings.roms_path): for f in filenames: yield os.path.join(dirpath, f) global _games_content if not _games_content: _games_content = filter(None, map(game_content, all_roms_path_files())) _games_content = dict((c.name, c) for c in _games_content) print len(_games_content), 'roms' if filt == 'favs': return filter_games_favorites(_games_content) else: return filter_games(_games_content, filt) def filter_games(all_games, filt): def name_to_search_key(name): name = os.path.split(name)[1] name = name.split('(')[0] words = name.lower().split() words = [re.sub('[^a-z0-9]', '', w) for w in words] return filter(None, words) def match_key(query, key): return all(any(kw.startswith(qw) for kw in key) for qw in query) return dict((k, v) for k, v in all_games.iteritems() if match_key(name_to_search_key(filt), name_to_search_key(k))) def filter_games_favorites(all_games): with open(settings.rom_favorites) as f: favs = set(os.path.splitext(g.strip())[0] for g in f.readlines()) return dict((k, v) for k, v in all_games.iteritems() if k in favs) class Playlist(object): def __init__(self, name, choices): self.name = name # a mapping of content to relative likelihood self.choices = choices if type(choices) == type({}) else dict((c, 1.) for c in choices) self.last_played = None def _all_choices_except_last_played(self): for choice in self.choices.keys(): if choice == self.last_played and len(self.choices) > 1: continue yield choice def get_likelihood(self, choice): return self.choices[choice] # TODO reduce likelihood of previous N selections def get_next(self): total_likelihood = sum(self.get_likelihood(choice) for choice in self._all_choices_except_last_played()) rand = random.uniform(0, total_likelihood) cumulative_likelihood = 0 choice = None for ch in self._all_choices_except_last_played(): cumulative_likelihood += self.get_likelihood(ch) if cumulative_likelihood > rand: choice = ch break self.last_played = choice return choice def to_json(self): return { 'name': self.name, 'items': sorted(c.name for c in self.choices.keys()), } def load_playlists(): base = Playlist('(almost) everything', (c for c in all_content().values() if not c.manual)) nosound = Playlist('no sound-reactive', (c for c in base.choices.keys() if not c.sound_reactive or not c.sound_required)) playlists = [base, nosound] playlists.extend(load_playlist_files()) return dict((pl.name, pl) for pl in playlists) def load_playlist_files(): playlist_files = os.listdir(settings.playlists_dir) for filename in playlist_files: name, ext = os.path.splitext(filename) if ext != '.playlist': continue path = os.path.join(settings.playlists_dir, filename) with open(path) as f: entries = filter(None, (ln.strip() for ln in f.readlines())) def
parse_entry
identifier_name
rcu.rs
let (valid, pr) = match pllclk { 48_000_000 => (true, 0b01), // pllclk / 1 72_000_000 => (true, 0b00), // pllclk / 1.5 96_000_000 => (true, 0b11), // pllclk / 2 _ => (false, 0), }; usbclk_valid = valid; // Configure USB prescaler rcu.cfg0.modify(|_, w| unsafe { w.usbfspsc().bits(pr) }); } else { usbclk_valid = false; } let clocks = Clocks { sysclk: Hertz(target_sysclk), apb1_psc, apb2_psc, usbclk_valid }; Rcu { clocks, regs: self.regs } } } #[derive(Copy, Clone)] pub struct Clocks { sysclk: Hertz, apb1_psc: u8, apb2_psc: u8, usbclk_valid: bool, } impl Clocks { /// Returns the system (core) frequency pub const fn sysclk(&self) -> Hertz { self.sysclk } /// Returns the frequency of the AHB pub const fn hclk(&self) -> Hertz { self.sysclk } /// Returns the frequency of the APB1 pub const fn pclk1(&self) -> Hertz { Hertz(self.sysclk.0 / self.apb1_psc as u32) } /// Returns the frequency of the APB2 pub const fn pclk2(&self) -> Hertz { Hertz(self.sysclk.0 / self.apb2_psc as u32) } /// Returns the frequency of the SysTick timer pub const fn systick(&self) -> Hertz { Hertz(self.sysclk.0 / 4) } /// Returns the frequency of the TIMER0 base clock pub fn timer0(&self) -> Hertz { let pclk2 = self.pclk2(); if self.apb2_psc == 1 { pclk2 } else { Hertz(pclk2.0 * 2) } } /// Returns the frequency of the TIMER1..6 base clock pub fn timerx(&self) -> Hertz { let pclk1 = self.pclk1(); if self.apb1_psc == 1 { pclk1 } else { Hertz(pclk1.0 * 2) } } /// Returns whether the USBCLK clock frequency is valid for the USB peripheral pub const fn usbclk_valid(&self) -> bool { self.usbclk_valid } } macro_rules! base_freq { ($($PER:ident => $func:ident,)+) => { $( impl BaseFrequency for crate::pac::$PER { #[inline(always)] fn base_frequency(rcu: &Rcu) -> Hertz { rcu.clocks.$func() } } )+ } } base_freq! { ADC0 => pclk2, ADC1 => pclk2, I2C0 => pclk1, I2C1 => pclk1, SPI0 => pclk2, SPI1 => pclk1, SPI2 => pclk1, TIMER0 => timer0, TIMER1 => timerx, TIMER2 => timerx, TIMER3 => timerx, TIMER4 => timerx, TIMER5 => timerx, TIMER6 => timerx, UART3 => pclk1, UART4 => pclk1, USART0 => pclk2, USART1 => pclk1, USART2 => pclk1, } pub(crate) mod closed_traits { use super::Rcu; use crate::time::Hertz; /// Enable/disable peripheral pub trait Enable { fn enable(rcu: &mut Rcu); fn disable(rcu: &mut Rcu); } /// Reset peripheral pub trait Reset { fn reset(rcu: &mut Rcu); } pub trait BaseFrequency { fn base_frequency(rcu: &Rcu) -> Hertz; } } pub(crate) use closed_traits::*; macro_rules! bus_enable { ($PER:ident => ($apben:ident, $peren:ident)) => { impl Enable for crate::pac::$PER { #[inline(always)] fn enable(rcu: &mut Rcu) { interrupt::free(|_| { rcu.regs.$apben.modify(|_, w| w.$peren().set_bit()); }); } #[inline(always)] fn disable(rcu: &mut Rcu) { interrupt::free(|_| { rcu.regs.$apben.modify(|_, w| w.$peren().clear_bit()); }); } } } } macro_rules! bus { ($($PER:ident => ($apben:ident, $apbrst:ident, $peren:ident, $perrst:ident),)+) => { $( bus_enable!($PER => ($apben, $peren)); impl Reset for crate::pac::$PER { #[inline(always)] fn reset(rcu: &mut Rcu) { interrupt::free(|_| { rcu.regs.$apbrst.modify(|_, w| w.$perrst().set_bit()); rcu.regs.$apbrst.modify(|_, w| w.$perrst().clear_bit()); }); } } )+ } } bus! { ADC0 => (apb2en, apb2rst, adc0en, adc0rst), ADC1 => (apb2en, apb2rst, adc1en, adc1rst), AFIO => (apb2en, apb2rst, afen, afrst), BKP => (apb1en, apb1rst, bkpien, bkpirst), CAN0 => (apb1en, apb1rst, can0en, can0rst), CAN1 => (apb1en, apb1rst, can1en, can1rst), DAC => (apb1en, apb1rst, dacen, dacrst), GPIOA => (apb2en, apb2rst, paen, parst), GPIOB => (apb2en, apb2rst, pben, pbrst), GPIOC => (apb2en, apb2rst, pcen, pcrst), GPIOD => (apb2en, apb2rst, pden, pdrst), GPIOE => (apb2en, apb2rst, peen, perst), I2C0 => (apb1en, apb1rst, i2c0en, i2c0rst), I2C1 => (apb1en, apb1rst, i2c1en, i2c1rst), PMU => (apb1en, apb1rst, pmuen, pmurst), SPI0 => (apb2en, apb2rst, spi0en, spi0rst), SPI1 => (apb1en, apb1rst, spi1en, spi1rst), SPI2 => (apb1en, apb1rst, spi2en, spi2rst), TIMER0 => (apb2en, apb2rst, timer0en, timer0rst), TIMER1 => (apb1en, apb1rst, timer1en, timer1rst), TIMER2 => (apb1en, apb1rst, timer2en, timer2rst), TIMER3 => (apb1en, apb1rst, timer3en, timer3rst), TIMER4 => (apb1en, apb1rst, timer4en, timer4rst), TIMER5 => (apb1en, apb1rst, timer5en, timer5rst), TIMER6 => (apb1en, apb1rst, timer6en, timer6rst), UART3 => (apb1en, apb1rst, uart3en, uart3rst), UART4 => (apb1en, apb1rst, uart4en, uart4rst), USART0 => (apb2en, apb2rst, usart0en, usart0rst), USART1 => (apb1en, apb1rst, usart1en, usart1rst), USART2 => (apb1en, apb1rst, usart2en, usart2rst), USBFS_GLOBAL => (ahben, ahbrst, usbfsen, usbfsrst), WWDGT => (apb1en, apb1rst, wwdgten, wwdgtrst),
} bus_enable!(CRC => (ahben, crcen)); bus_enable!(DMA0 => (ahben, dma0en));
random_line_split
rcu.rs
clk => (0b01, false), (None, sysclk) if IRC8M == sysclk => (0b00, false), _ => (0b10, true), }; let pllsel_bit; let predv0_bits; let pllmf_bits; if use_pll { let pllmf; if let Some(hxtal_freq) = self.hxtal { // Use external clock + divider pllsel_bit = true; let calculate_pll = |source: u32, target: u32| -> Option<(u8, u8)> { const PLL_IN_MIN: u32 = 600_000; let div_max = cmp::min(16, source / PLL_IN_MIN); for d in 1..=div_max { let pllsource = source / d; let pllm = target / pllsource; if pllm < 2 || pllm == 15 || pllm > 32{ continue; } let actual_freq = pllsource * pllm; if actual_freq == target { return Some((d as u8, pllm as u8)); } } None }; let (d, m) = calculate_pll(hxtal_freq, target_sysclk).expect("invalid sysclk value"); predv0_bits = d - 1; pllmf = m; } else { // IRC8M/2 is used as an input clock pllsel_bit = false; let pllsource = IRC8M / 2; let m = target_sysclk / pllsource; let m = cmp::max(2, cmp::min(m, 32)); assert_ne!(m, 15, "invalid sysclk value"); let actual_sysclk = pllsource * m; assert_eq!(target_sysclk, actual_sysclk, "invalid sysclk value"); predv0_bits = 0; pllmf = m as u8; } pllmf_bits = match pllmf { 2..=14 => pllmf - 2, 16..=32 => pllmf - 1, _ => unreachable!("invalid pll multiplier"), }; } else { pllsel_bit = false; predv0_bits = 0; pllmf_bits = 0; } // Switch to the internal clock let rcu = unsafe { &*crate::pac::RCU::ptr() }; rcu.ctl.modify(|_, w| w.irc8men().set_bit()); // Enable IRC8M oscillator while rcu.ctl.read().irc8mstb().bit_is_clear() {} // Wait for oscillator to stabilize rcu.cfg0.modify(|_, w| unsafe { w.scs().bits(0b00) }); // Switch to the internal oscillator rcu.ctl.modify(|_, w| w.pllen().clear_bit()); // Disable PLL // Set bus prescalers rcu.cfg0.modify(|_, w| unsafe { w.ahbpsc().bits(0b0000) }); // CK_SYS rcu.cfg0.modify(|_, w| unsafe { w.apb1psc().bits(0b100) }); // CK_AHB / 2 rcu.cfg0.modify(|_, w| unsafe { w.apb2psc().bits(0b000) }); // CK_AHB let apb1_psc = 2; let apb2_psc = 1; if self.hxtal.is_some() { // Enable external oscillator rcu.ctl.modify(|_, w| w.hxtalen().set_bit()); // Wait for oscillator to stabilize while rcu.ctl.read().hxtalstb().bit_is_clear() {} // Select HXTAL as prescaler input source clock rcu.cfg1.modify(|_, w| w.predv0sel().clear_bit()); // Configure the prescaler rcu.cfg1.modify(|_, w| unsafe { w.predv0().bits(predv0_bits) }); } if use_pll { // Configure PLL input selector rcu.cfg0.modify(|_, w| w.pllsel().bit(pllsel_bit)); // Configure PLL multiplier rcu.cfg0.modify(|_, w| unsafe { w .pllmf_4().bit(pllmf_bits & 0x10 != 0) .pllmf_3_0().bits(pllmf_bits & 0xf) }); // Enable PLL rcu.ctl.modify(|_, w| w.pllen().set_bit()); // Wait for PLL to stabilize while rcu.ctl.read().pllstb().bit_is_clear() {} } else { // Disable PLL rcu.ctl.modify(|_, w| w.pllen().clear_bit()); } // Switch to the configured clock source rcu.cfg0.modify(|_, w| unsafe { w.scs().bits(scs_bits) }); let usbclk_valid; if use_pll { let pllclk = target_sysclk; let (valid, pr) = match pllclk { 48_000_000 => (true, 0b01), // pllclk / 1 72_000_000 => (true, 0b00), // pllclk / 1.5 96_000_000 => (true, 0b11), // pllclk / 2 _ => (false, 0), }; usbclk_valid = valid; // Configure USB prescaler rcu.cfg0.modify(|_, w| unsafe { w.usbfspsc().bits(pr) }); } else { usbclk_valid = false; } let clocks = Clocks { sysclk: Hertz(target_sysclk), apb1_psc, apb2_psc, usbclk_valid }; Rcu { clocks, regs: self.regs } } } #[derive(Copy, Clone)] pub struct Clocks { sysclk: Hertz, apb1_psc: u8, apb2_psc: u8, usbclk_valid: bool, } impl Clocks { /// Returns the system (core) frequency pub const fn sysclk(&self) -> Hertz { self.sysclk } /// Returns the frequency of the AHB pub const fn hclk(&self) -> Hertz { self.sysclk } /// Returns the frequency of the APB1 pub const fn pclk1(&self) -> Hertz { Hertz(self.sysclk.0 / self.apb1_psc as u32) } /// Returns the frequency of the APB2 pub const fn pclk2(&self) -> Hertz { Hertz(self.sysclk.0 / self.apb2_psc as u32) } /// Returns the frequency of the SysTick timer pub const fn systick(&self) -> Hertz { Hertz(self.sysclk.0 / 4) } /// Returns the frequency of the TIMER0 base clock pub fn timer0(&self) -> Hertz
/// Returns the frequency of the TIMER1..6 base clock pub fn timerx(&self) -> Hertz { let pclk1 = self.pclk1(); if self.apb1_psc == 1 { pclk1 } else { Hertz(pclk1.0 * 2) } } /// Returns whether the USBCLK clock frequency is valid for the USB peripheral pub const fn usbclk_valid(&self) -> bool { self.usbclk_valid } } macro_rules! base_freq { ($($PER:ident => $func:ident,)+) => { $( impl BaseFrequency for crate::pac::$PER { #[inline(always)] fn base_frequency(rcu: &Rcu) -> Hertz { rcu.clocks.$func() } } )+ } } base_freq! { ADC0 => pclk2, ADC1 => pclk2, I2C0 => pclk1, I2C1 => pclk1, SPI0 => pclk2, SPI1 => pclk1, SPI2 => pclk1, TIMER0 => timer0, TIMER1 => timerx, TIMER2 => timerx, TIMER3 => timerx, TIMER4 => timerx, TIMER5 => timerx, TIMER6 => timerx, UART3 => pclk1, UART4 => pclk1, USART0 => pclk2, USART1 => pclk1, USART2 => pclk1, } pub(crate) mod closed_traits { use super::Rcu; use crate::time::Hertz; /// Enable/disable peripheral
{ let pclk2 = self.pclk2(); if self.apb2_psc == 1 { pclk2 } else { Hertz(pclk2.0 * 2) } }
identifier_body
rcu.rs
clk => (0b01, false), (None, sysclk) if IRC8M == sysclk => (0b00, false), _ => (0b10, true), }; let pllsel_bit; let predv0_bits; let pllmf_bits; if use_pll { let pllmf; if let Some(hxtal_freq) = self.hxtal { // Use external clock + divider pllsel_bit = true; let calculate_pll = |source: u32, target: u32| -> Option<(u8, u8)> { const PLL_IN_MIN: u32 = 600_000; let div_max = cmp::min(16, source / PLL_IN_MIN); for d in 1..=div_max { let pllsource = source / d; let pllm = target / pllsource; if pllm < 2 || pllm == 15 || pllm > 32{ continue; } let actual_freq = pllsource * pllm; if actual_freq == target { return Some((d as u8, pllm as u8)); } } None }; let (d, m) = calculate_pll(hxtal_freq, target_sysclk).expect("invalid sysclk value"); predv0_bits = d - 1; pllmf = m; } else { // IRC8M/2 is used as an input clock pllsel_bit = false; let pllsource = IRC8M / 2; let m = target_sysclk / pllsource; let m = cmp::max(2, cmp::min(m, 32)); assert_ne!(m, 15, "invalid sysclk value"); let actual_sysclk = pllsource * m; assert_eq!(target_sysclk, actual_sysclk, "invalid sysclk value"); predv0_bits = 0; pllmf = m as u8; } pllmf_bits = match pllmf { 2..=14 => pllmf - 2, 16..=32 => pllmf - 1, _ => unreachable!("invalid pll multiplier"), }; } else { pllsel_bit = false; predv0_bits = 0; pllmf_bits = 0; } // Switch to the internal clock let rcu = unsafe { &*crate::pac::RCU::ptr() }; rcu.ctl.modify(|_, w| w.irc8men().set_bit()); // Enable IRC8M oscillator while rcu.ctl.read().irc8mstb().bit_is_clear() {} // Wait for oscillator to stabilize rcu.cfg0.modify(|_, w| unsafe { w.scs().bits(0b00) }); // Switch to the internal oscillator rcu.ctl.modify(|_, w| w.pllen().clear_bit()); // Disable PLL // Set bus prescalers rcu.cfg0.modify(|_, w| unsafe { w.ahbpsc().bits(0b0000) }); // CK_SYS rcu.cfg0.modify(|_, w| unsafe { w.apb1psc().bits(0b100) }); // CK_AHB / 2 rcu.cfg0.modify(|_, w| unsafe { w.apb2psc().bits(0b000) }); // CK_AHB let apb1_psc = 2; let apb2_psc = 1; if self.hxtal.is_some() { // Enable external oscillator rcu.ctl.modify(|_, w| w.hxtalen().set_bit()); // Wait for oscillator to stabilize while rcu.ctl.read().hxtalstb().bit_is_clear() {} // Select HXTAL as prescaler input source clock rcu.cfg1.modify(|_, w| w.predv0sel().clear_bit()); // Configure the prescaler rcu.cfg1.modify(|_, w| unsafe { w.predv0().bits(predv0_bits) }); } if use_pll { // Configure PLL input selector rcu.cfg0.modify(|_, w| w.pllsel().bit(pllsel_bit)); // Configure PLL multiplier rcu.cfg0.modify(|_, w| unsafe { w .pllmf_4().bit(pllmf_bits & 0x10 != 0) .pllmf_3_0().bits(pllmf_bits & 0xf) }); // Enable PLL rcu.ctl.modify(|_, w| w.pllen().set_bit()); // Wait for PLL to stabilize while rcu.ctl.read().pllstb().bit_is_clear() {} } else { // Disable PLL rcu.ctl.modify(|_, w| w.pllen().clear_bit()); } // Switch to the configured clock source rcu.cfg0.modify(|_, w| unsafe { w.scs().bits(scs_bits) }); let usbclk_valid; if use_pll { let pllclk = target_sysclk; let (valid, pr) = match pllclk { 48_000_000 => (true, 0b01), // pllclk / 1 72_000_000 => (true, 0b00), // pllclk / 1.5 96_000_000 => (true, 0b11), // pllclk / 2 _ => (false, 0), }; usbclk_valid = valid; // Configure USB prescaler rcu.cfg0.modify(|_, w| unsafe { w.usbfspsc().bits(pr) }); } else { usbclk_valid = false; } let clocks = Clocks { sysclk: Hertz(target_sysclk), apb1_psc, apb2_psc, usbclk_valid }; Rcu { clocks, regs: self.regs } } } #[derive(Copy, Clone)] pub struct Clocks { sysclk: Hertz, apb1_psc: u8, apb2_psc: u8, usbclk_valid: bool, } impl Clocks { /// Returns the system (core) frequency pub const fn sysclk(&self) -> Hertz { self.sysclk } /// Returns the frequency of the AHB pub const fn hclk(&self) -> Hertz { self.sysclk } /// Returns the frequency of the APB1 pub const fn pclk1(&self) -> Hertz { Hertz(self.sysclk.0 / self.apb1_psc as u32) } /// Returns the frequency of the APB2 pub const fn pclk2(&self) -> Hertz { Hertz(self.sysclk.0 / self.apb2_psc as u32) } /// Returns the frequency of the SysTick timer pub const fn
(&self) -> Hertz { Hertz(self.sysclk.0 / 4) } /// Returns the frequency of the TIMER0 base clock pub fn timer0(&self) -> Hertz { let pclk2 = self.pclk2(); if self.apb2_psc == 1 { pclk2 } else { Hertz(pclk2.0 * 2) } } /// Returns the frequency of the TIMER1..6 base clock pub fn timerx(&self) -> Hertz { let pclk1 = self.pclk1(); if self.apb1_psc == 1 { pclk1 } else { Hertz(pclk1.0 * 2) } } /// Returns whether the USBCLK clock frequency is valid for the USB peripheral pub const fn usbclk_valid(&self) -> bool { self.usbclk_valid } } macro_rules! base_freq { ($($PER:ident => $func:ident,)+) => { $( impl BaseFrequency for crate::pac::$PER { #[inline(always)] fn base_frequency(rcu: &Rcu) -> Hertz { rcu.clocks.$func() } } )+ } } base_freq! { ADC0 => pclk2, ADC1 => pclk2, I2C0 => pclk1, I2C1 => pclk1, SPI0 => pclk2, SPI1 => pclk1, SPI2 => pclk1, TIMER0 => timer0, TIMER1 => timerx, TIMER2 => timerx, TIMER3 => timerx, TIMER4 => timerx, TIMER5 => timerx, TIMER6 => timerx, UART3 => pclk1, UART4 => pclk1, USART0 => pclk2, USART1 => pclk1, USART2 => pclk1, } pub(crate) mod closed_traits { use super::Rcu; use crate::time::Hertz; /// Enable/disable peripheral
systick
identifier_name
rcu.rs
clk => (0b01, false), (None, sysclk) if IRC8M == sysclk => (0b00, false), _ => (0b10, true), }; let pllsel_bit; let predv0_bits; let pllmf_bits; if use_pll
} } None }; let (d, m) = calculate_pll(hxtal_freq, target_sysclk).expect("invalid sysclk value"); predv0_bits = d - 1; pllmf = m; } else { // IRC8M/2 is used as an input clock pllsel_bit = false; let pllsource = IRC8M / 2; let m = target_sysclk / pllsource; let m = cmp::max(2, cmp::min(m, 32)); assert_ne!(m, 15, "invalid sysclk value"); let actual_sysclk = pllsource * m; assert_eq!(target_sysclk, actual_sysclk, "invalid sysclk value"); predv0_bits = 0; pllmf = m as u8; } pllmf_bits = match pllmf { 2..=14 => pllmf - 2, 16..=32 => pllmf - 1, _ => unreachable!("invalid pll multiplier"), }; } else { pllsel_bit = false; predv0_bits = 0; pllmf_bits = 0; } // Switch to the internal clock let rcu = unsafe { &*crate::pac::RCU::ptr() }; rcu.ctl.modify(|_, w| w.irc8men().set_bit()); // Enable IRC8M oscillator while rcu.ctl.read().irc8mstb().bit_is_clear() {} // Wait for oscillator to stabilize rcu.cfg0.modify(|_, w| unsafe { w.scs().bits(0b00) }); // Switch to the internal oscillator rcu.ctl.modify(|_, w| w.pllen().clear_bit()); // Disable PLL // Set bus prescalers rcu.cfg0.modify(|_, w| unsafe { w.ahbpsc().bits(0b0000) }); // CK_SYS rcu.cfg0.modify(|_, w| unsafe { w.apb1psc().bits(0b100) }); // CK_AHB / 2 rcu.cfg0.modify(|_, w| unsafe { w.apb2psc().bits(0b000) }); // CK_AHB let apb1_psc = 2; let apb2_psc = 1; if self.hxtal.is_some() { // Enable external oscillator rcu.ctl.modify(|_, w| w.hxtalen().set_bit()); // Wait for oscillator to stabilize while rcu.ctl.read().hxtalstb().bit_is_clear() {} // Select HXTAL as prescaler input source clock rcu.cfg1.modify(|_, w| w.predv0sel().clear_bit()); // Configure the prescaler rcu.cfg1.modify(|_, w| unsafe { w.predv0().bits(predv0_bits) }); } if use_pll { // Configure PLL input selector rcu.cfg0.modify(|_, w| w.pllsel().bit(pllsel_bit)); // Configure PLL multiplier rcu.cfg0.modify(|_, w| unsafe { w .pllmf_4().bit(pllmf_bits & 0x10 != 0) .pllmf_3_0().bits(pllmf_bits & 0xf) }); // Enable PLL rcu.ctl.modify(|_, w| w.pllen().set_bit()); // Wait for PLL to stabilize while rcu.ctl.read().pllstb().bit_is_clear() {} } else { // Disable PLL rcu.ctl.modify(|_, w| w.pllen().clear_bit()); } // Switch to the configured clock source rcu.cfg0.modify(|_, w| unsafe { w.scs().bits(scs_bits) }); let usbclk_valid; if use_pll { let pllclk = target_sysclk; let (valid, pr) = match pllclk { 48_000_000 => (true, 0b01), // pllclk / 1 72_000_000 => (true, 0b00), // pllclk / 1.5 96_000_000 => (true, 0b11), // pllclk / 2 _ => (false, 0), }; usbclk_valid = valid; // Configure USB prescaler rcu.cfg0.modify(|_, w| unsafe { w.usbfspsc().bits(pr) }); } else { usbclk_valid = false; } let clocks = Clocks { sysclk: Hertz(target_sysclk), apb1_psc, apb2_psc, usbclk_valid }; Rcu { clocks, regs: self.regs } } } #[derive(Copy, Clone)] pub struct Clocks { sysclk: Hertz, apb1_psc: u8, apb2_psc: u8, usbclk_valid: bool, } impl Clocks { /// Returns the system (core) frequency pub const fn sysclk(&self) -> Hertz { self.sysclk } /// Returns the frequency of the AHB pub const fn hclk(&self) -> Hertz { self.sysclk } /// Returns the frequency of the APB1 pub const fn pclk1(&self) -> Hertz { Hertz(self.sysclk.0 / self.apb1_psc as u32) } /// Returns the frequency of the APB2 pub const fn pclk2(&self) -> Hertz { Hertz(self.sysclk.0 / self.apb2_psc as u32) } /// Returns the frequency of the SysTick timer pub const fn systick(&self) -> Hertz { Hertz(self.sysclk.0 / 4) } /// Returns the frequency of the TIMER0 base clock pub fn timer0(&self) -> Hertz { let pclk2 = self.pclk2(); if self.apb2_psc == 1 { pclk2 } else { Hertz(pclk2.0 * 2) } } /// Returns the frequency of the TIMER1..6 base clock pub fn timerx(&self) -> Hertz { let pclk1 = self.pclk1(); if self.apb1_psc == 1 { pclk1 } else { Hertz(pclk1.0 * 2) } } /// Returns whether the USBCLK clock frequency is valid for the USB peripheral pub const fn usbclk_valid(&self) -> bool { self.usbclk_valid } } macro_rules! base_freq { ($($PER:ident => $func:ident,)+) => { $( impl BaseFrequency for crate::pac::$PER { #[inline(always)] fn base_frequency(rcu: &Rcu) -> Hertz { rcu.clocks.$func() } } )+ } } base_freq! { ADC0 => pclk2, ADC1 => pclk2, I2C0 => pclk1, I2C1 => pclk1, SPI0 => pclk2, SPI1 => pclk1, SPI2 => pclk1, TIMER0 => timer0, TIMER1 => timerx, TIMER2 => timerx, TIMER3 => timerx, TIMER4 => timerx, TIMER5 => timerx, TIMER6 => timerx, UART3 => pclk1, UART4 => pclk1, USART0 => pclk2, USART1 => pclk1, USART2 => pclk1, } pub(crate) mod closed_traits { use super::Rcu; use crate::time::Hertz; /// Enable/disable peripheral
{ let pllmf; if let Some(hxtal_freq) = self.hxtal { // Use external clock + divider pllsel_bit = true; let calculate_pll = |source: u32, target: u32| -> Option<(u8, u8)> { const PLL_IN_MIN: u32 = 600_000; let div_max = cmp::min(16, source / PLL_IN_MIN); for d in 1..=div_max { let pllsource = source / d; let pllm = target / pllsource; if pllm < 2 || pllm == 15 || pllm > 32{ continue; } let actual_freq = pllsource * pllm; if actual_freq == target { return Some((d as u8, pllm as u8));
conditional_block
types.go
nil { reg.participantsByGaiaID = make(map[string]*ParticipantData) } reg.participantsByGaiaID[id] = data } if id := data.ID.ChatID; id != "" { if reg.participantsByChatID == nil { reg.participantsByChatID = make(map[string]*ParticipantData) } reg.participantsByChatID[id] = data } reg.allParticipants = append(reg.allParticipants, data) } func (reg *ParticipantRegistry) AllParticipants() []*ParticipantData { return reg.allParticipants } func (reg *ParticipantRegistry) ForID(pid *ParticipantID) *ParticipantData { if pid.GaiaID != "" { if v := reg.participantsByGaiaID[pid.GaiaID]; v != nil { return v } } if pid.ChatID != "" { if v := reg.participantsByChatID[pid.ChatID]; v != nil { return v } } return nil } type Conversation struct { Conversation *ConversationEntry `json:"conversation"` EventsMessage json.RawMessage `json:"events"` initialized bool reg ParticipantRegistry events []json.RawMessage decodedEvents []*Event } func (ce *Conversation) initialize() error { if ce.initialized { return nil } if err := json.Unmarshal([]byte(ce.EventsMessage), &ce.events); err != nil { return err } ce.decodedEvents = make([]*Event, len(ce.events)) for _, pd := range ce.Conversation.ConversationInfo.ParticipantData { ce.reg.Register(pd) } ce.initialized = true return nil } func (ce *Conversation) ParticipantRegistry() *ParticipantRegistry { return &ce.reg } func (ce *Conversation) EventsSize() int { return len(ce.events) } func (ce *Conversation) ResolveAll() error { for i := 0; i < len(ce.events); i++ { _, err := ce.Event(i) if err != nil { return err } } return nil } func (ce *Conversation) Event(i int) (*Event, error) { if i < 0 || i >= len(ce.events) { return nil, errors.New("Index out of bounds") } if ce.decodedEvents[i] == nil { var event Event if err := json.Unmarshal(ce.events[i], &event); err != nil { return nil, err } ce.decodedEvents[i] = &event } return ce.decodedEvents[i], nil } type ConversationEntry struct { ConversationInfo *ConversationInfo `json:"conversation"` } type SingleID struct { ID string `json:"id"` } type ConversationInfo struct { ID *SingleID `json:"id"` Type string `json:"type"` Name string `json:"name"` CurrentParticipant []*ParticipantID `json:"current_participant"` ParticipantData []*ParticipantData `json:"participant_data"` } type ParticipantID struct { GaiaID string `json:"gaia_id"` ChatID string `json:"chat_id"` } func (pid *ParticipantID) String() string
// Matches returns true if pid's Gaia or Chat ID are both populated and match // the equivalent values in other. func (pid *ParticipantID) Matches(other *ParticipantID) bool { if pid.GaiaID != "" && pid.GaiaID == other.GaiaID { return true } if pid.ChatID != "" && pid.ChatID == other.ChatID { return true } return false } type ParticipantData struct { ID ParticipantID `json:"id"` FallbackName string `json:"fallback_name"` ParticipantType string `json:"participant_type"` DomainID string `json:"domain_id"` } func (pd *ParticipantData) DisplayName() string { return pd.FallbackName } type MessageContentSegment struct { Type string `json:"type"` Text string `json:"text"` Formatting struct { Bold bool `json:"bold"` Italics bool `json:"italics"` Strikethrough bool `json:"strikethrough"` Underline bool `json:"underline"` } `json:"formatting"` LinkData *struct { LinkTarget string `json:"link_target"` } `json:"link_data"` } type Thumbnail struct { URL string `json:"url"` ImageURL string `json:"image_url"` WidthPx int64 `json:"width_px"` HeightPx int64 `json:"height_px"` } type PlusPhoto struct { Thumbnail *Thumbnail `json:"thumbnail"` OwnerObfuscatedID string `json:"owner_obfuscated_id"` AlbumID string `json:"album_id"` PhotoID string `json:"photo_id"` URL string `json:"url"` OriginalContentURL string `json:"original_content_url"` MediaType string `json:"media_type"` } type GeoCoordinatesV2 struct { Latitude float64 `json:"latitude"` Longitude float64 `json:"longitude"` } type PostalAddressV2 struct { StreetAddress string `json:"street_address"` Name string `json:"name"` AddressCountry string `json:"address_country"` AddressLocality string `json:"address_locality"` AddressRegion string `json:"address_region"` PostalCode string `json:"postal_code"` } type ImageObjectV2 struct { URL string `json:"url"` } type PlaceV2 struct { URL string `json:"url"` Name string `json:"name"` Address *struct { PostalAddressV2 PostalAddressV2 `json:"postal_address_v2"` } `json:"address"` Geo *struct { GeoCoordinatesV2 GeoCoordinatesV2 `json:"geo_coordinates_v2"` } `json:"geo"` ID string `json:"id"` ImageObjectV2 *ImageObjectV2 `json:"image_object_v2"` } type RepresentativeImage struct { Type []EmbedItemType `json:"type"` ID string `json:"id"` ImageObjectV2 *ImageObjectV2 `json:"image_object_v2"` } type ThingV2 struct { URL string `json:"url"` Name string `json:"name"` RepresentativeImage *RepresentativeImage `json:"representative_image"` } type EmbedItem struct { Type []EmbedItemType `json:"type"` ID string `json:"id"` PlusPhoto *PlusPhoto `json:"plus_photo"` PlaceV2 *PlaceV2 `json:"place_v2"` ThingV2 *ThingV2 `json:"thing_v2"` ImageObjectV2 *ImageObjectV2 `json:"image_object_v2"` } func (ei *EmbedItem) Key() string { if pp := ei.PlusPhoto; pp != nil { return fmt.Sprintf("%s:%s", pp.AlbumID, pp.PhotoID) } if p := ei.ThingV2; p != nil { // Use a hash of the Thing's URL. return util.HashForKey(p.URL) } return ei.ID } type MessageContentAttachment struct { EmbedItem *EmbedItem `json:"embed_item"` ID string `json:"id"` } type MessageContent struct { Segment []*MessageContentSegment `json:"segment"` Attachment []*MessageContentAttachment `json:"attachment"` } type ChatMessage struct { MessageContent *MessageContent `json:"message_content"` } type ConversationRename struct { NewName string `json:"new_name"` OldName string `json:"old_name"` } type MembershipChange struct { Type string `json:"type"` ParticipantID []*ParticipantID `json:"participant_id"` LeaveReason string `json:"leave_reason"` } type Event struct { ConversationID *SingleID `json:"conversation_id"` SenderID *ParticipantID `json:"sender_id"` Timestamp string `json:"timestamp"` ConversationRename *ConversationRename `json:"conversation_rename"` ChatMessage *ChatMessage `json:"chat_message"` MembershipChange *MembershipChange `json:"membership_change"` EventID string `json:"event_id"` EventType EventType `json:"event_type"` } func (e *Event) Time() (time.Time, error) { // Timestamp is in microseconds from epoch. micros, err := strconv.ParseInt(e.Timestamp, 10, 64) if err != nil { return time.Time{}, err } return time.Unix(0, micros*1000), nil } func (e *Event) Description(reg *ParticipantRegistry) (string, error) { var parts []string // Time switch t, err := e.Time(); err { case nil: parts = append(parts, t.In(time.Local).Format(time.RFC3339Nano)) default: parts = append(parts, fmt.Sprintf("Timestmap Error (%s)", e.Timestamp)) } if sid := e.SenderID; sid != nil { var pd *ParticipantData if reg != nil { pd = reg.ForID(sid)
{ return fmt.Sprintf("gaia:%s/chat:%s", pid.GaiaID, pid.ChatID) }
identifier_body
types.go
nil { reg.participantsByGaiaID = make(map[string]*ParticipantData) } reg.participantsByGaiaID[id] = data } if id := data.ID.ChatID; id != "" { if reg.participantsByChatID == nil { reg.participantsByChatID = make(map[string]*ParticipantData) } reg.participantsByChatID[id] = data } reg.allParticipants = append(reg.allParticipants, data) } func (reg *ParticipantRegistry) AllParticipants() []*ParticipantData { return reg.allParticipants } func (reg *ParticipantRegistry) ForID(pid *ParticipantID) *ParticipantData { if pid.GaiaID != "" { if v := reg.participantsByGaiaID[pid.GaiaID]; v != nil { return v } } if pid.ChatID != "" { if v := reg.participantsByChatID[pid.ChatID]; v != nil { return v } } return nil } type Conversation struct { Conversation *ConversationEntry `json:"conversation"` EventsMessage json.RawMessage `json:"events"` initialized bool reg ParticipantRegistry events []json.RawMessage decodedEvents []*Event } func (ce *Conversation) initialize() error { if ce.initialized { return nil } if err := json.Unmarshal([]byte(ce.EventsMessage), &ce.events); err != nil { return err } ce.decodedEvents = make([]*Event, len(ce.events)) for _, pd := range ce.Conversation.ConversationInfo.ParticipantData { ce.reg.Register(pd) } ce.initialized = true return nil } func (ce *Conversation) ParticipantRegistry() *ParticipantRegistry { return &ce.reg } func (ce *Conversation) EventsSize() int { return len(ce.events) } func (ce *Conversation) ResolveAll() error { for i := 0; i < len(ce.events); i++ { _, err := ce.Event(i) if err != nil { return err } } return nil } func (ce *Conversation) Event(i int) (*Event, error) { if i < 0 || i >= len(ce.events) { return nil, errors.New("Index out of bounds") } if ce.decodedEvents[i] == nil { var event Event if err := json.Unmarshal(ce.events[i], &event); err != nil { return nil, err } ce.decodedEvents[i] = &event } return ce.decodedEvents[i], nil } type ConversationEntry struct { ConversationInfo *ConversationInfo `json:"conversation"` } type SingleID struct { ID string `json:"id"` } type ConversationInfo struct { ID *SingleID `json:"id"` Type string `json:"type"` Name string `json:"name"` CurrentParticipant []*ParticipantID `json:"current_participant"` ParticipantData []*ParticipantData `json:"participant_data"` } type ParticipantID struct { GaiaID string `json:"gaia_id"` ChatID string `json:"chat_id"` } func (pid *ParticipantID) String() string { return fmt.Sprintf("gaia:%s/chat:%s", pid.GaiaID, pid.ChatID) } // Matches returns true if pid's Gaia or Chat ID are both populated and match // the equivalent values in other. func (pid *ParticipantID) Matches(other *ParticipantID) bool { if pid.GaiaID != "" && pid.GaiaID == other.GaiaID { return true } if pid.ChatID != "" && pid.ChatID == other.ChatID { return true } return false } type ParticipantData struct { ID ParticipantID `json:"id"` FallbackName string `json:"fallback_name"` ParticipantType string `json:"participant_type"` DomainID string `json:"domain_id"` } func (pd *ParticipantData) DisplayName() string { return pd.FallbackName } type MessageContentSegment struct { Type string `json:"type"` Text string `json:"text"` Formatting struct { Bold bool `json:"bold"` Italics bool `json:"italics"` Strikethrough bool `json:"strikethrough"` Underline bool `json:"underline"` } `json:"formatting"` LinkData *struct { LinkTarget string `json:"link_target"` } `json:"link_data"` } type Thumbnail struct { URL string `json:"url"` ImageURL string `json:"image_url"` WidthPx int64 `json:"width_px"` HeightPx int64 `json:"height_px"` } type PlusPhoto struct { Thumbnail *Thumbnail `json:"thumbnail"` OwnerObfuscatedID string `json:"owner_obfuscated_id"` AlbumID string `json:"album_id"` PhotoID string `json:"photo_id"` URL string `json:"url"` OriginalContentURL string `json:"original_content_url"` MediaType string `json:"media_type"` } type GeoCoordinatesV2 struct { Latitude float64 `json:"latitude"` Longitude float64 `json:"longitude"` } type PostalAddressV2 struct { StreetAddress string `json:"street_address"` Name string `json:"name"` AddressCountry string `json:"address_country"` AddressLocality string `json:"address_locality"` AddressRegion string `json:"address_region"` PostalCode string `json:"postal_code"` } type ImageObjectV2 struct { URL string `json:"url"` } type PlaceV2 struct { URL string `json:"url"` Name string `json:"name"` Address *struct { PostalAddressV2 PostalAddressV2 `json:"postal_address_v2"` } `json:"address"` Geo *struct { GeoCoordinatesV2 GeoCoordinatesV2 `json:"geo_coordinates_v2"` } `json:"geo"` ID string `json:"id"` ImageObjectV2 *ImageObjectV2 `json:"image_object_v2"` } type RepresentativeImage struct { Type []EmbedItemType `json:"type"` ID string `json:"id"` ImageObjectV2 *ImageObjectV2 `json:"image_object_v2"` } type ThingV2 struct { URL string `json:"url"` Name string `json:"name"` RepresentativeImage *RepresentativeImage `json:"representative_image"` } type EmbedItem struct { Type []EmbedItemType `json:"type"` ID string `json:"id"` PlusPhoto *PlusPhoto `json:"plus_photo"` PlaceV2 *PlaceV2 `json:"place_v2"` ThingV2 *ThingV2 `json:"thing_v2"` ImageObjectV2 *ImageObjectV2 `json:"image_object_v2"` } func (ei *EmbedItem)
() string { if pp := ei.PlusPhoto; pp != nil { return fmt.Sprintf("%s:%s", pp.AlbumID, pp.PhotoID) } if p := ei.ThingV2; p != nil { // Use a hash of the Thing's URL. return util.HashForKey(p.URL) } return ei.ID } type MessageContentAttachment struct { EmbedItem *EmbedItem `json:"embed_item"` ID string `json:"id"` } type MessageContent struct { Segment []*MessageContentSegment `json:"segment"` Attachment []*MessageContentAttachment `json:"attachment"` } type ChatMessage struct { MessageContent *MessageContent `json:"message_content"` } type ConversationRename struct { NewName string `json:"new_name"` OldName string `json:"old_name"` } type MembershipChange struct { Type string `json:"type"` ParticipantID []*ParticipantID `json:"participant_id"` LeaveReason string `json:"leave_reason"` } type Event struct { ConversationID *SingleID `json:"conversation_id"` SenderID *ParticipantID `json:"sender_id"` Timestamp string `json:"timestamp"` ConversationRename *ConversationRename `json:"conversation_rename"` ChatMessage *ChatMessage `json:"chat_message"` MembershipChange *MembershipChange `json:"membership_change"` EventID string `json:"event_id"` EventType EventType `json:"event_type"` } func (e *Event) Time() (time.Time, error) { // Timestamp is in microseconds from epoch. micros, err := strconv.ParseInt(e.Timestamp, 10, 64) if err != nil { return time.Time{}, err } return time.Unix(0, micros*1000), nil } func (e *Event) Description(reg *ParticipantRegistry) (string, error) { var parts []string // Time switch t, err := e.Time(); err { case nil: parts = append(parts, t.In(time.Local).Format(time.RFC3339Nano)) default: parts = append(parts, fmt.Sprintf("Timestmap Error (%s)", e.Timestamp)) } if sid := e.SenderID; sid != nil { var pd *ParticipantData if reg != nil { pd = reg.ForID(sid)
Key
identifier_name
types.go
nil { reg.participantsByGaiaID = make(map[string]*ParticipantData) } reg.participantsByGaiaID[id] = data } if id := data.ID.ChatID; id != "" { if reg.participantsByChatID == nil { reg.participantsByChatID = make(map[string]*ParticipantData) } reg.participantsByChatID[id] = data } reg.allParticipants = append(reg.allParticipants, data) } func (reg *ParticipantRegistry) AllParticipants() []*ParticipantData { return reg.allParticipants } func (reg *ParticipantRegistry) ForID(pid *ParticipantID) *ParticipantData { if pid.GaiaID != "" { if v := reg.participantsByGaiaID[pid.GaiaID]; v != nil { return v } } if pid.ChatID != "" { if v := reg.participantsByChatID[pid.ChatID]; v != nil { return v } } return nil } type Conversation struct { Conversation *ConversationEntry `json:"conversation"` EventsMessage json.RawMessage `json:"events"` initialized bool reg ParticipantRegistry events []json.RawMessage decodedEvents []*Event } func (ce *Conversation) initialize() error { if ce.initialized { return nil } if err := json.Unmarshal([]byte(ce.EventsMessage), &ce.events); err != nil { return err } ce.decodedEvents = make([]*Event, len(ce.events)) for _, pd := range ce.Conversation.ConversationInfo.ParticipantData { ce.reg.Register(pd) } ce.initialized = true return nil } func (ce *Conversation) ParticipantRegistry() *ParticipantRegistry { return &ce.reg } func (ce *Conversation) EventsSize() int { return len(ce.events) } func (ce *Conversation) ResolveAll() error { for i := 0; i < len(ce.events); i++ { _, err := ce.Event(i) if err != nil { return err } } return nil } func (ce *Conversation) Event(i int) (*Event, error) { if i < 0 || i >= len(ce.events) { return nil, errors.New("Index out of bounds") } if ce.decodedEvents[i] == nil
return ce.decodedEvents[i], nil } type ConversationEntry struct { ConversationInfo *ConversationInfo `json:"conversation"` } type SingleID struct { ID string `json:"id"` } type ConversationInfo struct { ID *SingleID `json:"id"` Type string `json:"type"` Name string `json:"name"` CurrentParticipant []*ParticipantID `json:"current_participant"` ParticipantData []*ParticipantData `json:"participant_data"` } type ParticipantID struct { GaiaID string `json:"gaia_id"` ChatID string `json:"chat_id"` } func (pid *ParticipantID) String() string { return fmt.Sprintf("gaia:%s/chat:%s", pid.GaiaID, pid.ChatID) } // Matches returns true if pid's Gaia or Chat ID are both populated and match // the equivalent values in other. func (pid *ParticipantID) Matches(other *ParticipantID) bool { if pid.GaiaID != "" && pid.GaiaID == other.GaiaID { return true } if pid.ChatID != "" && pid.ChatID == other.ChatID { return true } return false } type ParticipantData struct { ID ParticipantID `json:"id"` FallbackName string `json:"fallback_name"` ParticipantType string `json:"participant_type"` DomainID string `json:"domain_id"` } func (pd *ParticipantData) DisplayName() string { return pd.FallbackName } type MessageContentSegment struct { Type string `json:"type"` Text string `json:"text"` Formatting struct { Bold bool `json:"bold"` Italics bool `json:"italics"` Strikethrough bool `json:"strikethrough"` Underline bool `json:"underline"` } `json:"formatting"` LinkData *struct { LinkTarget string `json:"link_target"` } `json:"link_data"` } type Thumbnail struct { URL string `json:"url"` ImageURL string `json:"image_url"` WidthPx int64 `json:"width_px"` HeightPx int64 `json:"height_px"` } type PlusPhoto struct { Thumbnail *Thumbnail `json:"thumbnail"` OwnerObfuscatedID string `json:"owner_obfuscated_id"` AlbumID string `json:"album_id"` PhotoID string `json:"photo_id"` URL string `json:"url"` OriginalContentURL string `json:"original_content_url"` MediaType string `json:"media_type"` } type GeoCoordinatesV2 struct { Latitude float64 `json:"latitude"` Longitude float64 `json:"longitude"` } type PostalAddressV2 struct { StreetAddress string `json:"street_address"` Name string `json:"name"` AddressCountry string `json:"address_country"` AddressLocality string `json:"address_locality"` AddressRegion string `json:"address_region"` PostalCode string `json:"postal_code"` } type ImageObjectV2 struct { URL string `json:"url"` } type PlaceV2 struct { URL string `json:"url"` Name string `json:"name"` Address *struct { PostalAddressV2 PostalAddressV2 `json:"postal_address_v2"` } `json:"address"` Geo *struct { GeoCoordinatesV2 GeoCoordinatesV2 `json:"geo_coordinates_v2"` } `json:"geo"` ID string `json:"id"` ImageObjectV2 *ImageObjectV2 `json:"image_object_v2"` } type RepresentativeImage struct { Type []EmbedItemType `json:"type"` ID string `json:"id"` ImageObjectV2 *ImageObjectV2 `json:"image_object_v2"` } type ThingV2 struct { URL string `json:"url"` Name string `json:"name"` RepresentativeImage *RepresentativeImage `json:"representative_image"` } type EmbedItem struct { Type []EmbedItemType `json:"type"` ID string `json:"id"` PlusPhoto *PlusPhoto `json:"plus_photo"` PlaceV2 *PlaceV2 `json:"place_v2"` ThingV2 *ThingV2 `json:"thing_v2"` ImageObjectV2 *ImageObjectV2 `json:"image_object_v2"` } func (ei *EmbedItem) Key() string { if pp := ei.PlusPhoto; pp != nil { return fmt.Sprintf("%s:%s", pp.AlbumID, pp.PhotoID) } if p := ei.ThingV2; p != nil { // Use a hash of the Thing's URL. return util.HashForKey(p.URL) } return ei.ID } type MessageContentAttachment struct { EmbedItem *EmbedItem `json:"embed_item"` ID string `json:"id"` } type MessageContent struct { Segment []*MessageContentSegment `json:"segment"` Attachment []*MessageContentAttachment `json:"attachment"` } type ChatMessage struct { MessageContent *MessageContent `json:"message_content"` } type ConversationRename struct { NewName string `json:"new_name"` OldName string `json:"old_name"` } type MembershipChange struct { Type string `json:"type"` ParticipantID []*ParticipantID `json:"participant_id"` LeaveReason string `json:"leave_reason"` } type Event struct { ConversationID *SingleID `json:"conversation_id"` SenderID *ParticipantID `json:"sender_id"` Timestamp string `json:"timestamp"` ConversationRename *ConversationRename `json:"conversation_rename"` ChatMessage *ChatMessage `json:"chat_message"` MembershipChange *MembershipChange `json:"membership_change"` EventID string `json:"event_id"` EventType EventType `json:"event_type"` } func (e *Event) Time() (time.Time, error) { // Timestamp is in microseconds from epoch. micros, err := strconv.ParseInt(e.Timestamp, 10, 64) if err != nil { return time.Time{}, err } return time.Unix(0, micros*1000), nil } func (e *Event) Description(reg *ParticipantRegistry) (string, error) { var parts []string // Time switch t, err := e.Time(); err { case nil: parts = append(parts, t.In(time.Local).Format(time.RFC3339Nano)) default: parts = append(parts, fmt.Sprintf("Timestmap Error (%s)", e.Timestamp)) } if sid := e.SenderID; sid != nil { var pd *ParticipantData if reg != nil { pd = reg.ForID(sid)
{ var event Event if err := json.Unmarshal(ce.events[i], &event); err != nil { return nil, err } ce.decodedEvents[i] = &event }
conditional_block
types.go
nil { reg.participantsByGaiaID = make(map[string]*ParticipantData) } reg.participantsByGaiaID[id] = data } if id := data.ID.ChatID; id != "" { if reg.participantsByChatID == nil { reg.participantsByChatID = make(map[string]*ParticipantData) } reg.participantsByChatID[id] = data } reg.allParticipants = append(reg.allParticipants, data) } func (reg *ParticipantRegistry) AllParticipants() []*ParticipantData { return reg.allParticipants } func (reg *ParticipantRegistry) ForID(pid *ParticipantID) *ParticipantData { if pid.GaiaID != "" { if v := reg.participantsByGaiaID[pid.GaiaID]; v != nil { return v } } if pid.ChatID != "" { if v := reg.participantsByChatID[pid.ChatID]; v != nil { return v } } return nil } type Conversation struct { Conversation *ConversationEntry `json:"conversation"` EventsMessage json.RawMessage `json:"events"` initialized bool reg ParticipantRegistry events []json.RawMessage decodedEvents []*Event } func (ce *Conversation) initialize() error { if ce.initialized { return nil } if err := json.Unmarshal([]byte(ce.EventsMessage), &ce.events); err != nil { return err } ce.decodedEvents = make([]*Event, len(ce.events)) for _, pd := range ce.Conversation.ConversationInfo.ParticipantData { ce.reg.Register(pd) } ce.initialized = true return nil } func (ce *Conversation) ParticipantRegistry() *ParticipantRegistry { return &ce.reg } func (ce *Conversation) EventsSize() int { return len(ce.events) } func (ce *Conversation) ResolveAll() error { for i := 0; i < len(ce.events); i++ { _, err := ce.Event(i) if err != nil { return err } } return nil } func (ce *Conversation) Event(i int) (*Event, error) { if i < 0 || i >= len(ce.events) { return nil, errors.New("Index out of bounds") } if ce.decodedEvents[i] == nil { var event Event if err := json.Unmarshal(ce.events[i], &event); err != nil { return nil, err } ce.decodedEvents[i] = &event } return ce.decodedEvents[i], nil } type ConversationEntry struct { ConversationInfo *ConversationInfo `json:"conversation"` } type SingleID struct { ID string `json:"id"` } type ConversationInfo struct { ID *SingleID `json:"id"` Type string `json:"type"` Name string `json:"name"` CurrentParticipant []*ParticipantID `json:"current_participant"` ParticipantData []*ParticipantData `json:"participant_data"` } type ParticipantID struct { GaiaID string `json:"gaia_id"` ChatID string `json:"chat_id"` } func (pid *ParticipantID) String() string { return fmt.Sprintf("gaia:%s/chat:%s", pid.GaiaID, pid.ChatID) } // Matches returns true if pid's Gaia or Chat ID are both populated and match // the equivalent values in other. func (pid *ParticipantID) Matches(other *ParticipantID) bool { if pid.GaiaID != "" && pid.GaiaID == other.GaiaID { return true } if pid.ChatID != "" && pid.ChatID == other.ChatID { return true } return false } type ParticipantData struct { ID ParticipantID `json:"id"` FallbackName string `json:"fallback_name"` ParticipantType string `json:"participant_type"` DomainID string `json:"domain_id"` } func (pd *ParticipantData) DisplayName() string { return pd.FallbackName } type MessageContentSegment struct { Type string `json:"type"` Text string `json:"text"` Formatting struct { Bold bool `json:"bold"` Italics bool `json:"italics"` Strikethrough bool `json:"strikethrough"` Underline bool `json:"underline"` } `json:"formatting"` LinkData *struct { LinkTarget string `json:"link_target"` } `json:"link_data"` } type Thumbnail struct { URL string `json:"url"` ImageURL string `json:"image_url"` WidthPx int64 `json:"width_px"` HeightPx int64 `json:"height_px"` } type PlusPhoto struct { Thumbnail *Thumbnail `json:"thumbnail"` OwnerObfuscatedID string `json:"owner_obfuscated_id"` AlbumID string `json:"album_id"` PhotoID string `json:"photo_id"` URL string `json:"url"` OriginalContentURL string `json:"original_content_url"` MediaType string `json:"media_type"` } type GeoCoordinatesV2 struct { Latitude float64 `json:"latitude"` Longitude float64 `json:"longitude"` } type PostalAddressV2 struct { StreetAddress string `json:"street_address"` Name string `json:"name"` AddressCountry string `json:"address_country"` AddressLocality string `json:"address_locality"` AddressRegion string `json:"address_region"` PostalCode string `json:"postal_code"` } type ImageObjectV2 struct { URL string `json:"url"` } type PlaceV2 struct { URL string `json:"url"` Name string `json:"name"` Address *struct { PostalAddressV2 PostalAddressV2 `json:"postal_address_v2"` } `json:"address"` Geo *struct { GeoCoordinatesV2 GeoCoordinatesV2 `json:"geo_coordinates_v2"` } `json:"geo"` ID string `json:"id"` ImageObjectV2 *ImageObjectV2 `json:"image_object_v2"` } type RepresentativeImage struct { Type []EmbedItemType `json:"type"` ID string `json:"id"` ImageObjectV2 *ImageObjectV2 `json:"image_object_v2"` } type ThingV2 struct { URL string `json:"url"` Name string `json:"name"` RepresentativeImage *RepresentativeImage `json:"representative_image"` } type EmbedItem struct { Type []EmbedItemType `json:"type"` ID string `json:"id"` PlusPhoto *PlusPhoto `json:"plus_photo"` PlaceV2 *PlaceV2 `json:"place_v2"` ThingV2 *ThingV2 `json:"thing_v2"` ImageObjectV2 *ImageObjectV2 `json:"image_object_v2"` } func (ei *EmbedItem) Key() string { if pp := ei.PlusPhoto; pp != nil { return fmt.Sprintf("%s:%s", pp.AlbumID, pp.PhotoID) } if p := ei.ThingV2; p != nil { // Use a hash of the Thing's URL. return util.HashForKey(p.URL) } return ei.ID } type MessageContentAttachment struct { EmbedItem *EmbedItem `json:"embed_item"`
type MessageContent struct { Segment []*MessageContentSegment `json:"segment"` Attachment []*MessageContentAttachment `json:"attachment"` } type ChatMessage struct { MessageContent *MessageContent `json:"message_content"` } type ConversationRename struct { NewName string `json:"new_name"` OldName string `json:"old_name"` } type MembershipChange struct { Type string `json:"type"` ParticipantID []*ParticipantID `json:"participant_id"` LeaveReason string `json:"leave_reason"` } type Event struct { ConversationID *SingleID `json:"conversation_id"` SenderID *ParticipantID `json:"sender_id"` Timestamp string `json:"timestamp"` ConversationRename *ConversationRename `json:"conversation_rename"` ChatMessage *ChatMessage `json:"chat_message"` MembershipChange *MembershipChange `json:"membership_change"` EventID string `json:"event_id"` EventType EventType `json:"event_type"` } func (e *Event) Time() (time.Time, error) { // Timestamp is in microseconds from epoch. micros, err := strconv.ParseInt(e.Timestamp, 10, 64) if err != nil { return time.Time{}, err } return time.Unix(0, micros*1000), nil } func (e *Event) Description(reg *ParticipantRegistry) (string, error) { var parts []string // Time switch t, err := e.Time(); err { case nil: parts = append(parts, t.In(time.Local).Format(time.RFC3339Nano)) default: parts = append(parts, fmt.Sprintf("Timestmap Error (%s)", e.Timestamp)) } if sid := e.SenderID; sid != nil { var pd *ParticipantData if reg != nil { pd = reg.ForID(sid)
ID string `json:"id"` }
random_line_split
lib.rs
/// assert_eq!( select1(0, 0x0000_0000_0000_0002), Some(1) ); /// assert_eq!( select1(0, 0x0000_0000_0000_0004), Some(2) ); /// assert_eq!( select1(2, 0x0000_0000_0000_0004), None ); /// assert_eq!( select1(2, 0x0000_1010_1010_0114), Some(8) ); /// assert_eq!( select1(3, 0x0000_1010_1010_0114), Some(20) ); /// assert_eq!( select1(4, 0x0000_1010_1010_0114), Some(28) ); /// ``` #[inline] pub fn select1(r: usize, x: u64) -> Option<usize> { let result = select1_raw(r, x); if result == 72 {None} else {Some(result)} } /// Finds the index of the `r`th one bit in `x`, returning 72 when not found. /// /// Branchless. Uses the broadword algorithm from Vigna. /// Note that bits are numbered from least-significant to most. #[inline] #[allow(clippy::many_single_char_names)] pub fn select1_raw(r: usize, x: u64) -> usize { let r = r as u64; let mut s = x - ((x & 0xAAAA_AAAA_AAAA_AAAA) >> 1); s = (s & 0x3333_3333_3333_3333) + ((s >> 2) & 0x3333_3333_3333_3333); s = ((s + (s >> 4)) & 0x0F0F_0F0F_0F0F_0F0F).wrapping_mul(L8); let b = (i_le8(s, r.wrapping_mul(L8)) >> 7).wrapping_mul(L8)>> 53 & !7; let l = r - ((s << 8).wrapping_shr(b as u32) & 0xFF); s = (u_nz8((x.wrapping_shr(b as u32) & 0xFF) .wrapping_mul(L8) & 0x8040_2010_0804_0201) >> 7) .wrapping_mul(L8); (b + ((i_le8(s, l.wrapping_mul(L8)) >> 7).wrapping_mul(L8) >> 56)) as usize } /// Parallel ≤, treating a `u64` as a vector of 8 `u8`s. /// /// Branchless. /// /// # Examples /// /// ``` /// use broadword::u_le8; /// /// assert_eq!( u_le8(0x03_03_04_17_92_A0_A0_A1, /// 0x04_03_03_92_17_A0_A0_A0), /// 0x80_80_00_80_00_80_80_00 ); /// ``` #[inline] pub fn u_le8(x: u64, y: u64) -> u64 { ((((y | H8) - (x & !H8)) | (x ^ y)) ^ (x & !y)) & H8 } /// Parallel ≤, treating a `u64` as a vector of 8 `i8`s. /// /// Branchless. /// /// # Examples /// /// ``` /// use broadword::i_le8; /// /// assert_eq!( i_le8(0x03_03_04_00_FF_A0_A0_A1, /// 0x04_03_03_FF_00_A0_A0_A0), /// 0x80_80_00_00_80_80_80_00 ); /// ``` #[inline] pub fn i_le8(x: u64, y: u64) -> u64 { (((y | H8) - (x & !H8)) ^ x ^ y) & H8 } /// Parallel >0, treating a `u64` as a vector of 8 `u8`s. /// /// Branchless. /// /// # Examples /// /// ``` /// use broadword::u_nz8; /// /// assert_eq!( u_nz8(0x00_01_A9_40_20_17_00_06), /// 0x00_80_80_80_80_80_00_80 ); #[inline] pub fn u_nz8(x: u64) -> u64 { (((x | H8) - L8) | x) & H8 } #[cfg(test)] #[allow(clippy::many_single_char_names)] mod test { use std::hash::{Hash, Hasher}; use std::collections::hash_map::DefaultHasher; use quickcheck::TestResult; use super::*; #[test] fn count_ones_0() { assert_eq!(0, count_ones(0)); } #[test] fn count_ones_1() { assert_eq!(1, count_ones(1)); } #[test] fn count_ones_0000_0000_0000_0010() { assert_eq!(1, count_ones(0x0000_0000_0000_0010)); } #[test] fn count_ones_1000_0000_0000_0000() { assert_eq!(1, count_ones(0x1000_0000_0000_0000)); } #[test] fn count_ones_ffff_ffff_ffff_ffff() { assert_eq!(64, count_ones(0xFFFF_FFFF_FFFF_FFFF)); } fn count_ones_prop_base(word: u64) -> bool { count_ones(word) == word.count_ones() as usize } quickcheck! { fn count_ones_prop(word: u64) -> bool { count_ones_prop_base(word) } fn count_ones_prop_hash(word: u64) -> bool { count_ones_prop_base(hash(&word)) } } #[test] fn select1_0_0() { assert_eq!(None, select1(0, 0)); } #[test] fn select1_0_1() { assert_eq!(Some(0), select1(0, 1)); } #[test] fn select1_0_2() { assert_eq!(Some(1), select1(0, 2)); } #[test] fn select1_0_3() { assert_eq!(Some(0), select1(0, 3)); } #[test] fn select1_1_2() { assert_eq!(None, select1(1, 2)); } #[test] fn select1_1_3() { assert_eq!(Some(1), select1(1, 3)); } #[test] fn select1_3_13() { assert_eq!(None, select1(3, 0b1101)); } fn select1_slow(r: usize, x: u64) -> Option<usize> { let mut count = 0; for index in 0 .. 64 { if (x >> index) & 1 == 1 { count += 1; } if count == r + 1 { return Some(index); } } None } fn select1_prop_base(r: u8, x: u64) -> TestResult { if r > 64 { return TestResult::discard(); } TestResult::from_bool( select1(r as usize, x) == select1_slow(r as usize, x)) } quickcheck! { fn select1_prop(r: u8, x: u64) -> TestResult { select1_prop_base(r, x) } fn select1_prop_hash(r: u8, x: u64) -> TestResult { select1_prop_base(r, hash(&x)) } } fn get_bits(x: u64, i: u8, n: u8) -> u64 { let mask = if n == 64 {!0} else {(1 << n) - 1}; (x >> i) & mask
random_line_split
lib.rs
: u64) -> TestResult { if r > 64 { return TestResult::discard(); } TestResult::from_bool( select1(r as usize, x) == select1_slow(r as usize, x)) } quickcheck! { fn select1_prop(r: u8, x: u64) -> TestResult { select1_prop_base(r, x) } fn select1_prop_hash(r: u8, x: u64) -> TestResult { select1_prop_base(r, hash(&x)) } } fn get_bits(x: u64, i: u8, n: u8) -> u64 { let mask = if n == 64 {!0} else {(1 << n) - 1}; (x >> i) & mask } quickcheck! { fn u_nz8_prop(argument: (u64, u64, u64, u64)) -> bool { let n = hash(&argument); let r = u_nz8(n); for i in 0..8 { let ni = get_bits(n, 8 * i, 8); let ri = get_bits(r, 8 * i, 8); if (ni != 0) != (ri == 0x80) { return false; } } true } } #[test] fn u_nz8_works() { assert_eq!(b(0, 0, 0, 0, 0, 0, 0, 0), u_nz8(u(0, 0, 0, 0, 0, 0, 0, 0))); assert_eq!(b( 1, 1, 0, 1, 0, 1, 1, 1), u_nz8(u(45, 12, 0, 129, 0, 3, 80, 1))); assert_eq!(b(1, 1, 1, 1, 1, 1, 1, 1), u_nz8(u(1, 2, 3, 4, 5, 6, 7, 8))); assert_eq!(b( 1, 1, 1, 1, 0, 1, 1, 1), u_nz8(0xFF_FF_FF_FF_00_FF_FF_FF)); } fn u_le8_prop_base(n: u64, m: u64) -> bool { let r = u_le8(n, m); for i in 0..8 { let ni = get_bits(n, 8 * i, 8); let mi = get_bits(m, 8 * i, 8); let ri = get_bits(r, 8 * i, 8); if (ni <= mi) != (ri == 0x80) { return false; } } true } quickcheck! { fn u_le8_prop(n: u64, m: u64) -> bool { u_le8_prop_base(n, m) } fn u_le8_prop_hashed(n: (u64, u64, u64, u64), m: (u64, u64, u64, u64)) -> bool { let n = hash(&n); let m = hash(&m); u_le8_prop_base(n, m) } } #[test] fn le8_works() { assert_eq!(b( 1, 1, 1, 1, 0, 0, 0, 0), i_le8(i(0, 0, 0, 0, 0, 0, 0, 0), i( 3, 2, 1, 0, -1, -2, -3, -4))); assert_eq!(b( 0, 0, 0, 1, 1, 1, 1, 1), i_le8(i(3, 2, 1, 0, -1, -2, -3, -4), i( 0, 0, 0, 0, 0, 0, 0, 0))); assert_eq!(b( 0, 0, 1, 1, 1, 1, 1, 1), i_le8(i(19, 18, 17, 16, 15, 0, -1, -2), i(17, 17, 17, 17, 17, 17, 17, 17))); assert_eq!(b( 1, 1, 0, 0, 0, 0, 0, 0), i_le8(i(-9, -8, -7, 0, 1, 2, 3, 4), i(-8, -8, -8, -8, -8, -8, -8, -8))); assert_eq!(b( 0, 1, 0, 1, 1, 0, 1, 0), i_le8(i(8, 3, 46, 0, 0, 0, -6, -1), i( 7, 3, 24, 1, 0, -9, 5, -2))); } #[test] fn u_le8_works() { assert_eq!(b( 1, 1, 1, 1, 1, 1, 1, 1), u_le8(u( 0, 0, 0, 0, 0, 0, 0, 0), u( 7, 6, 5, 4, 3, 2, 1, 0))); assert_eq!(b( 1, 0, 0, 0, 0, 0, 0, 0), u_le8(u( 0, 1, 2, 3, 4, 5, 6, 7), u( 0, 0, 0, 0, 0, 0, 0, 0))); assert_eq!(b( 0, 0, 1, 1, 1, 1, 1, 1), u_le8(u(19, 18, 17, 16, 15, 14, 13, 12), u(17, 17, 17, 17, 17, 17, 17, 17))); assert_eq!(b( 0, 1, 0, 1, 1, 0, 1, 0), u_le8(u( 8, 3, 46, 0, 0, 9, 3, 2), u( 7, 3, 24, 1, 0, 0, 5, 1))); } /// Helpers for creating u64s. fn b(a: u64, b: u64, c: u64, d: u64, e: u64, f: u64, g: u64, h: u64) -> u64 { (a << 63) | (b << 55) | (c << 47) | (d << 39) | (e << 31) | (f << 23) | (g << 15) | (h << 7) } fn u(a: u8, b: u8, c: u8, d: u8, e: u8, f: u8, g: u8, h: u8) -> u64 { ((a
as u64) << 56) | ((b as u64) << 48) | ((c as u64) << 40) | ((d as u64) << 32) | ((e as u64) << 24) | ((f as u64) << 16) | ((g as u64) << 8) | (h as u64) } fn i(a:
identifier_body
lib.rs
/// Uses the broadword algorithm from Vigna. /// Note that bits are numbered from least-significant to most. /// /// # Examples /// /// ``` /// use broadword::select1; /// /// assert_eq!( select1(0, 0x0000_0000_0000_0000), None ); /// assert_eq!( select1(0, 0x0000_0000_0000_0001), Some(0) ); /// assert_eq!( select1(0, 0x0000_0000_0000_0002), Some(1) ); /// assert_eq!( select1(0, 0x0000_0000_0000_0004), Some(2) ); /// assert_eq!( select1(2, 0x0000_0000_0000_0004), None ); /// assert_eq!( select1(2, 0x0000_1010_1010_0114), Some(8) ); /// assert_eq!( select1(3, 0x0000_1010_1010_0114), Some(20) ); /// assert_eq!( select1(4, 0x0000_1010_1010_0114), Some(28) ); /// ``` #[inline] pub fn select1(r: usize, x: u64) -> Option<usize> { let result = select1_raw(r, x); if result == 72 {None} else {Some(result)} } /// Finds the index of the `r`th one bit in `x`, returning 72 when not found. /// /// Branchless. Uses the broadword algorithm from Vigna. /// Note that bits are numbered from least-significant to most. #[inline] #[allow(clippy::many_single_char_names)] pub fn select1_raw(r: usize, x: u64) -> usize { let r = r as u64; let mut s = x - ((x & 0xAAAA_AAAA_AAAA_AAAA) >> 1); s = (s & 0x3333_3333_3333_3333) + ((s >> 2) & 0x3333_3333_3333_3333); s = ((s + (s >> 4)) & 0x0F0F_0F0F_0F0F_0F0F).wrapping_mul(L8); let b = (i_le8(s, r.wrapping_mul(L8)) >> 7).wrapping_mul(L8)>> 53 & !7; let l = r - ((s << 8).wrapping_shr(b as u32) & 0xFF); s = (u_nz8((x.wrapping_shr(b as u32) & 0xFF) .wrapping_mul(L8) & 0x8040_2010_0804_0201) >> 7) .wrapping_mul(L8); (b + ((i_le8(s, l.wrapping_mul(L8)) >> 7).wrapping_mul(L8) >> 56)) as usize } /// Parallel ≤, treating a `u64` as a vector of 8 `u8`s. /// /// Branchless. /// /// # Examples /// /// ``` /// use broadword::u_le8; /// /// assert_eq!( u_le8(0x03_03_04_17_92_A0_A0_A1, /// 0x04_03_03_92_17_A0_A0_A0), /// 0x80_80_00_80_00_80_80_00 ); /// ``` #[inline] pub fn u_le8(x: u64, y: u64) -> u64 { ((((y | H8) - (x & !H8)) | (x ^ y)) ^ (x & !y)) & H8 } /// Parallel ≤, treating a `u64` as a vector of 8 `i8`s. /// /// Branchless. /// /// # Examples /// /// ``` /// use broadword::i_le8; /// /// assert_eq!( i_le8(0x03_03_04_00_FF_A0_A0_A1, /// 0x04_03_03_FF_00_A0_A0_A0), /// 0x80_80_00_00_80_80_80_00 ); /// ``` #[inline] pub fn i_le8(x: u64, y: u64) -> u64 { (((y | H8) - (x & !H8)) ^ x ^ y) & H8 } /// Parallel >0, treating a `u64` as a vector of 8 `u8`s. /// /// Branchless. /// /// # Examples /// /// ``` /// use broadword::u_nz8; /// /// assert_eq!( u_nz8(0x00_01_A9_40_20_17_00_06), /// 0x00_80_80_80_80_80_00_80 ); #[inline] pub fn u_nz8(x: u64) -> u64 { (((x | H8) - L8) | x) & H8 } #[cfg(test)] #[allow(clippy::many_single_char_names)] mod test { use std::hash::{Hash, Hasher}; use std::collections::hash_map::DefaultHasher; use quickcheck::TestResult; use super::*; #[test] fn count_ones_0() { assert_eq!(0, count_ones(0)); } #[test] fn count_ones_1() { assert_eq!(1, count_ones(1)); } #[test] fn count_ones_0000_0000_0000_0010() { assert_eq!(1, count_ones(0x0000_0000_0000_0010)); } #[test] fn count_ones_1000_0000_0000_0000() { assert_eq!(1, count_ones(0x1000_0000_0000_0000)); } #[test] fn count_ones_fff
ssert_eq!(64, count_ones(0xFFFF_FFFF_FFFF_FFFF)); } fn count_ones_prop_base(word: u64) -> bool { count_ones(word) == word.count_ones() as usize } quickcheck! { fn count_ones_prop(word: u64) -> bool { count_ones_prop_base(word) } fn count_ones_prop_hash(word: u64) -> bool { count_ones_prop_base(hash(&word)) } } #[test] fn select1_0_0() { assert_eq!(None, select1(0, 0)); } #[test] fn select1_0_1() { assert_eq!(Some(0), select1(0, 1)); } #[test] fn select1_0_2() { assert_eq!(Some(1), select1(0, 2)); } #[test] fn select1_0_3() { assert_eq!(Some(0), select1(0, 3)); } #[test] fn select1_1_2() { assert_eq!(None, select1(1, 2)); } #[test] fn select1_1_3() { assert_eq!(Some(1), select1(1, 3)); } #[test] fn select1_3_13() { assert_eq!(None, select1(3, 0b1101)); } fn select1_slow(r: usize, x: u64) -> Option<usize> { let mut count = 0; for index in 0 .. 64 { if (x >> index) & 1 == 1 { count += 1; } if count == r + 1 { return Some(index); } } None } fn select1_prop_base(r: u8, x: u64) -> TestResult { if r > 64 { return TestResult::discard(); } TestResult::from_bool( select1(r as usize, x) == select1_slow(r as usize, x)) } quickcheck! { fn select1_prop(r: u
f_ffff_ffff_ffff() { a
identifier_name
lib.rs
/// Uses the broadword algorithm from Vigna. /// Note that bits are numbered from least-significant to most. /// /// # Examples /// /// ``` /// use broadword::select1; /// /// assert_eq!( select1(0, 0x0000_0000_0000_0000), None ); /// assert_eq!( select1(0, 0x0000_0000_0000_0001), Some(0) ); /// assert_eq!( select1(0, 0x0000_0000_0000_0002), Some(1) ); /// assert_eq!( select1(0, 0x0000_0000_0000_0004), Some(2) ); /// assert_eq!( select1(2, 0x0000_0000_0000_0004), None ); /// assert_eq!( select1(2, 0x0000_1010_1010_0114), Some(8) ); /// assert_eq!( select1(3, 0x0000_1010_1010_0114), Some(20) ); /// assert_eq!( select1(4, 0x0000_1010_1010_0114), Some(28) ); /// ``` #[inline] pub fn select1(r: usize, x: u64) -> Option<usize> { let result = select1_raw(r, x); if result == 72 {None} else {Some(resu
nds the index of the `r`th one bit in `x`, returning 72 when not found. /// /// Branchless. Uses the broadword algorithm from Vigna. /// Note that bits are numbered from least-significant to most. #[inline] #[allow(clippy::many_single_char_names)] pub fn select1_raw(r: usize, x: u64) -> usize { let r = r as u64; let mut s = x - ((x & 0xAAAA_AAAA_AAAA_AAAA) >> 1); s = (s & 0x3333_3333_3333_3333) + ((s >> 2) & 0x3333_3333_3333_3333); s = ((s + (s >> 4)) & 0x0F0F_0F0F_0F0F_0F0F).wrapping_mul(L8); let b = (i_le8(s, r.wrapping_mul(L8)) >> 7).wrapping_mul(L8)>> 53 & !7; let l = r - ((s << 8).wrapping_shr(b as u32) & 0xFF); s = (u_nz8((x.wrapping_shr(b as u32) & 0xFF) .wrapping_mul(L8) & 0x8040_2010_0804_0201) >> 7) .wrapping_mul(L8); (b + ((i_le8(s, l.wrapping_mul(L8)) >> 7).wrapping_mul(L8) >> 56)) as usize } /// Parallel ≤, treating a `u64` as a vector of 8 `u8`s. /// /// Branchless. /// /// # Examples /// /// ``` /// use broadword::u_le8; /// /// assert_eq!( u_le8(0x03_03_04_17_92_A0_A0_A1, /// 0x04_03_03_92_17_A0_A0_A0), /// 0x80_80_00_80_00_80_80_00 ); /// ``` #[inline] pub fn u_le8(x: u64, y: u64) -> u64 { ((((y | H8) - (x & !H8)) | (x ^ y)) ^ (x & !y)) & H8 } /// Parallel ≤, treating a `u64` as a vector of 8 `i8`s. /// /// Branchless. /// /// # Examples /// /// ``` /// use broadword::i_le8; /// /// assert_eq!( i_le8(0x03_03_04_00_FF_A0_A0_A1, /// 0x04_03_03_FF_00_A0_A0_A0), /// 0x80_80_00_00_80_80_80_00 ); /// ``` #[inline] pub fn i_le8(x: u64, y: u64) -> u64 { (((y | H8) - (x & !H8)) ^ x ^ y) & H8 } /// Parallel >0, treating a `u64` as a vector of 8 `u8`s. /// /// Branchless. /// /// # Examples /// /// ``` /// use broadword::u_nz8; /// /// assert_eq!( u_nz8(0x00_01_A9_40_20_17_00_06), /// 0x00_80_80_80_80_80_00_80 ); #[inline] pub fn u_nz8(x: u64) -> u64 { (((x | H8) - L8) | x) & H8 } #[cfg(test)] #[allow(clippy::many_single_char_names)] mod test { use std::hash::{Hash, Hasher}; use std::collections::hash_map::DefaultHasher; use quickcheck::TestResult; use super::*; #[test] fn count_ones_0() { assert_eq!(0, count_ones(0)); } #[test] fn count_ones_1() { assert_eq!(1, count_ones(1)); } #[test] fn count_ones_0000_0000_0000_0010() { assert_eq!(1, count_ones(0x0000_0000_0000_0010)); } #[test] fn count_ones_1000_0000_0000_0000() { assert_eq!(1, count_ones(0x1000_0000_0000_0000)); } #[test] fn count_ones_ffff_ffff_ffff_ffff() { assert_eq!(64, count_ones(0xFFFF_FFFF_FFFF_FFFF)); } fn count_ones_prop_base(word: u64) -> bool { count_ones(word) == word.count_ones() as usize } quickcheck! { fn count_ones_prop(word: u64) -> bool { count_ones_prop_base(word) } fn count_ones_prop_hash(word: u64) -> bool { count_ones_prop_base(hash(&word)) } } #[test] fn select1_0_0() { assert_eq!(None, select1(0, 0)); } #[test] fn select1_0_1() { assert_eq!(Some(0), select1(0, 1)); } #[test] fn select1_0_2() { assert_eq!(Some(1), select1(0, 2)); } #[test] fn select1_0_3() { assert_eq!(Some(0), select1(0, 3)); } #[test] fn select1_1_2() { assert_eq!(None, select1(1, 2)); } #[test] fn select1_1_3() { assert_eq!(Some(1), select1(1, 3)); } #[test] fn select1_3_13() { assert_eq!(None, select1(3, 0b1101)); } fn select1_slow(r: usize, x: u64) -> Option<usize> { let mut count = 0; for index in 0 .. 64 { if (x >> index) & 1 == 1 { count += 1; } if count == r + 1 { return Some(index); } } None } fn select1_prop_base(r: u8, x: u64) -> TestResult { if r > 64 { return TestResult::discard(); } TestResult::from_bool( select1(r as usize, x) == select1_slow(r as usize, x)) } quickcheck! { fn select1_prop(r: u
lt)} } /// Fi
conditional_block
uart.rs
_rxto: ReadWrite<u32, Event::Register>, // 0x144-0x148 _reserved7: [u32; 1], // 0x148-0x14C pub event_rxstarted: ReadWrite<u32, Event::Register>, // 0x14C-0x150 pub event_txstarted: ReadWrite<u32, Event::Register>, // 0x150-0x154 _reserved8: [u32; 1], // 0x154-0x158 pub event_txstopped: ReadWrite<u32, Event::Register>, // 0x158-0x15c _reserved9: [u32; 41], // 0x15c-0x200 pub shorts: ReadWrite<u32, Shorts::Register>, // 0x200-0x204 _reserved10: [u32; 64], // 0x204-0x304 pub intenset: ReadWrite<u32, Interrupt::Register>, // 0x304-0x308 pub intenclr: ReadWrite<u32, Interrupt::Register>, // 0x308-0x30C _reserved11: [u32; 93], // 0x30C-0x480 pub errorsrc: ReadWrite<u32, ErrorSrc::Register>, // 0x480-0x484 _reserved12: [u32; 31], // 0x484-0x500 pub enable: ReadWrite<u32, Uart::Register>, // 0x500-0x504 _reserved13: [u32; 1], // 0x504-0x508 pub pselrts: ReadWrite<u32, Psel::Register>, // 0x508-0x50c pub pseltxd: ReadWrite<u32, Psel::Register>, // 0x50c-0x510 pub pselcts: ReadWrite<u32, Psel::Register>, // 0x510-0x514 pub pselrxd: ReadWrite<u32, Psel::Register>, // 0x514-0x518 _reserved14: [u32; 3], // 0x518-0x524 pub baudrate: ReadWrite<u32, Baudrate::Register>, // 0x524-0x528 _reserved15: [u32; 3], // 0x528-0x534 pub rxd_ptr: ReadWrite<u32, Pointer::Register>, // 0x534-0x538 pub rxd_maxcnt: ReadWrite<u32, Counter::Register>, // 0x538-0x53c pub rxd_amount: ReadOnly<u32, Counter::Register>, // 0x53c-0x540 _reserved16: [u32; 1], // 0x540-0x544 pub txd_ptr: ReadWrite<u32, Pointer::Register>, // 0x544-0x548 pub txd_maxcnt: ReadWrite<u32, Counter::Register>, // 0x548-0x54c pub txd_amount: ReadOnly<u32, Counter::Register>, // 0x54c-0x550 _reserved17: [u32; 7], // 0x550-0x56C pub config: ReadWrite<u32, Config::Register>, // 0x56C-0x570 } #[cfg_attr(rustfmt, rustfmt_skip)] register_bitfields! [u32, /// Start task Task [ ENABLE OFFSET(0) NUMBITS(1) ], /// Read event Event [ READY OFFSET(0) NUMBITS(1) ], /// Shortcuts Shorts [ // Shortcut between ENDRX and STARTRX ENDRX_STARTRX OFFSET(5) NUMBITS(1), // Shortcut between ENDRX and STOPRX ENDRX_STOPRX OFFSET(6) NUMBITS(1) ], /// UART Interrupts Interrupt [ CTS OFFSET(0) NUMBITS(1), NCTS OFFSET(1) NUMBITS(1), ENDRX OFFSET(4) NUMBITS(1), ENDTX OFFSET(8) NUMBITS(1), ERROR OFFSET(9) NUMBITS(1), RXTO OFFSET(17) NUMBITS(1), RXSTARTED OFFSET(19) NUMBITS(1), TXSTARTED OFFSET(20) NUMBITS(1), TXSTOPPED OFFSET(22) NUMBITS(1) ], /// UART Errors ErrorSrc [ OVERRUN OFFSET(0) NUMBITS(1), PARITY OFFSET(1) NUMBITS(1), FRAMING OFFSET(2) NUMBITS(1), BREAK OFFSET(3) NUMBITS(1) ], /// Enable UART Uart [ ENABLE OFFSET(0) NUMBITS(4) [ ON = 8, OFF = 0 ] ], /// Pin select Psel [ // Pin number PIN OFFSET(0) NUMBITS(5), // Connect/Disconnect CONNECT OFFSET(31) NUMBITS(1) ], /// Baudrate Baudrate [ BAUDRAUTE OFFSET(0) NUMBITS(32) ], /// DMA pointer Pointer [ POINTER OFFSET(0) NUMBITS(32) ], /// Counter value Counter [ COUNTER OFFSET(0) NUMBITS(8) ], /// Configuration of parity and flow control Config [ HWFC OFFSET(0) NUMBITS(1), PARITY OFFSET(1) NUMBITS(3) ] ]; /// UARTE // It should never be instanced outside this module but because a static mutable reference to it // is exported outside this module it must be `pub` pub struct
{ regs: *const UarteRegisters, client: Cell<Option<&'static kernel::hil::uart::Client>>, buffer: kernel::common::take_cell::TakeCell<'static, [u8]>, remaining_bytes: Cell<usize>, offset: Cell<usize>, } #[derive(Copy, Clone)] pub struct UARTParams { pub baud_rate: u32, } /// UARTE0 handle // This should only be accessed by the reset_handler on startup pub static mut UARTE0: Uarte = Uarte::new(); impl Uarte { /// Constructor pub const fn new() -> Uarte { Uarte { regs: UARTE_BASE as *const UarteRegisters, client: Cell::new(None), buffer: kernel::common::take_cell::TakeCell::empty(), remaining_bytes: Cell::new(0), offset: Cell::new(0), } } /// Configure which pins the UART should use for txd, rxd, cts and rts pub fn configure( &self, txd: pinmux::Pinmux, rxd: pinmux::Pinmux, cts: pinmux::Pinmux, rts: pinmux::Pinmux, ) { let regs = unsafe { &*self.regs }; regs.pseltxd.write(Psel::PIN.val(txd.into())); regs.pselrxd.write(Psel::PIN.val(rxd.into())); regs.pselcts.write(Psel::PIN.val(cts.into())); regs.pselrts.write(Psel::PIN.val(rts.into())); } fn set_baud_rate(&self, baud_rate: u32) { let regs = unsafe { &*self.regs }; match baud_rate { 1200 => regs.baudrate.set(0x0004F000), 2400 => regs.baudrate.set(0x0009D000), 4800 => regs.baudrate.set(0x0013B000), 9600 => regs.baudrate.set(0x00275000), 14400 => regs.baudrate.set(0x003AF000), 19200 => regs.baudrate.set(0x004EA000), 28800 =>
Uarte
identifier_name
uart.rs
event_rxto: ReadWrite<u32, Event::Register>, // 0x144-0x148 _reserved7: [u32; 1], // 0x148-0x14C pub event_rxstarted: ReadWrite<u32, Event::Register>, // 0x14C-0x150 pub event_txstarted: ReadWrite<u32, Event::Register>, // 0x150-0x154 _reserved8: [u32; 1], // 0x154-0x158 pub event_txstopped: ReadWrite<u32, Event::Register>, // 0x158-0x15c _reserved9: [u32; 41], // 0x15c-0x200 pub shorts: ReadWrite<u32, Shorts::Register>, // 0x200-0x204 _reserved10: [u32; 64], // 0x204-0x304 pub intenset: ReadWrite<u32, Interrupt::Register>, // 0x304-0x308 pub intenclr: ReadWrite<u32, Interrupt::Register>, // 0x308-0x30C _reserved11: [u32; 93], // 0x30C-0x480
pub pselrts: ReadWrite<u32, Psel::Register>, // 0x508-0x50c pub pseltxd: ReadWrite<u32, Psel::Register>, // 0x50c-0x510 pub pselcts: ReadWrite<u32, Psel::Register>, // 0x510-0x514 pub pselrxd: ReadWrite<u32, Psel::Register>, // 0x514-0x518 _reserved14: [u32; 3], // 0x518-0x524 pub baudrate: ReadWrite<u32, Baudrate::Register>, // 0x524-0x528 _reserved15: [u32; 3], // 0x528-0x534 pub rxd_ptr: ReadWrite<u32, Pointer::Register>, // 0x534-0x538 pub rxd_maxcnt: ReadWrite<u32, Counter::Register>, // 0x538-0x53c pub rxd_amount: ReadOnly<u32, Counter::Register>, // 0x53c-0x540 _reserved16: [u32; 1], // 0x540-0x544 pub txd_ptr: ReadWrite<u32, Pointer::Register>, // 0x544-0x548 pub txd_maxcnt: ReadWrite<u32, Counter::Register>, // 0x548-0x54c pub txd_amount: ReadOnly<u32, Counter::Register>, // 0x54c-0x550 _reserved17: [u32; 7], // 0x550-0x56C pub config: ReadWrite<u32, Config::Register>, // 0x56C-0x570 } #[cfg_attr(rustfmt, rustfmt_skip)] register_bitfields! [u32, /// Start task Task [ ENABLE OFFSET(0) NUMBITS(1) ], /// Read event Event [ READY OFFSET(0) NUMBITS(1) ], /// Shortcuts Shorts [ // Shortcut between ENDRX and STARTRX ENDRX_STARTRX OFFSET(5) NUMBITS(1), // Shortcut between ENDRX and STOPRX ENDRX_STOPRX OFFSET(6) NUMBITS(1) ], /// UART Interrupts Interrupt [ CTS OFFSET(0) NUMBITS(1), NCTS OFFSET(1) NUMBITS(1), ENDRX OFFSET(4) NUMBITS(1), ENDTX OFFSET(8) NUMBITS(1), ERROR OFFSET(9) NUMBITS(1), RXTO OFFSET(17) NUMBITS(1), RXSTARTED OFFSET(19) NUMBITS(1), TXSTARTED OFFSET(20) NUMBITS(1), TXSTOPPED OFFSET(22) NUMBITS(1) ], /// UART Errors ErrorSrc [ OVERRUN OFFSET(0) NUMBITS(1), PARITY OFFSET(1) NUMBITS(1), FRAMING OFFSET(2) NUMBITS(1), BREAK OFFSET(3) NUMBITS(1) ], /// Enable UART Uart [ ENABLE OFFSET(0) NUMBITS(4) [ ON = 8, OFF = 0 ] ], /// Pin select Psel [ // Pin number PIN OFFSET(0) NUMBITS(5), // Connect/Disconnect CONNECT OFFSET(31) NUMBITS(1) ], /// Baudrate Baudrate [ BAUDRAUTE OFFSET(0) NUMBITS(32) ], /// DMA pointer Pointer [ POINTER OFFSET(0) NUMBITS(32) ], /// Counter value Counter [ COUNTER OFFSET(0) NUMBITS(8) ], /// Configuration of parity and flow control Config [ HWFC OFFSET(0) NUMBITS(1), PARITY OFFSET(1) NUMBITS(3) ] ]; /// UARTE // It should never be instanced outside this module but because a static mutable reference to it // is exported outside this module it must be `pub` pub struct Uarte { regs: *const UarteRegisters, client: Cell<Option<&'static kernel::hil::uart::Client>>, buffer: kernel::common::take_cell::TakeCell<'static, [u8]>, remaining_bytes: Cell<usize>, offset: Cell<usize>, } #[derive(Copy, Clone)] pub struct UARTParams { pub baud_rate: u32, } /// UARTE0 handle // This should only be accessed by the reset_handler on startup pub static mut UARTE0: Uarte = Uarte::new(); impl Uarte { /// Constructor pub const fn new() -> Uarte { Uarte { regs: UARTE_BASE as *const UarteRegisters, client: Cell::new(None), buffer: kernel::common::take_cell::TakeCell::empty(), remaining_bytes: Cell::new(0), offset: Cell::new(0), } } /// Configure which pins the UART should use for txd, rxd, cts and rts pub fn configure( &self, txd: pinmux::Pinmux, rxd: pinmux::Pinmux, cts: pinmux::Pinmux, rts: pinmux::Pinmux, ) { let regs = unsafe { &*self.regs }; regs.pseltxd.write(Psel::PIN.val(txd.into())); regs.pselrxd.write(Psel::PIN.val(rxd.into())); regs.pselcts.write(Psel::PIN.val(cts.into())); regs.pselrts.write(Psel::PIN.val(rts.into())); } fn set_baud_rate(&self, baud_rate: u32) { let regs = unsafe { &*self.regs }; match baud_rate { 1200 => regs.baudrate.set(0x0004F000), 2400 => regs.baudrate.set(0x0009D000), 4800 => regs.baudrate.set(0x0013B000), 9600 => regs.baudrate.set(0x00275000), 14400 => regs.baudrate.set(0x003AF000), 19200 => regs.baudrate.set(0x004EA000), 28800 => regs.b
pub errorsrc: ReadWrite<u32, ErrorSrc::Register>, // 0x480-0x484 _reserved12: [u32; 31], // 0x484-0x500 pub enable: ReadWrite<u32, Uart::Register>, // 0x500-0x504 _reserved13: [u32; 1], // 0x504-0x508
random_line_split
uart.rs
38-0x53c pub rxd_amount: ReadOnly<u32, Counter::Register>, // 0x53c-0x540 _reserved16: [u32; 1], // 0x540-0x544 pub txd_ptr: ReadWrite<u32, Pointer::Register>, // 0x544-0x548 pub txd_maxcnt: ReadWrite<u32, Counter::Register>, // 0x548-0x54c pub txd_amount: ReadOnly<u32, Counter::Register>, // 0x54c-0x550 _reserved17: [u32; 7], // 0x550-0x56C pub config: ReadWrite<u32, Config::Register>, // 0x56C-0x570 } #[cfg_attr(rustfmt, rustfmt_skip)] register_bitfields! [u32, /// Start task Task [ ENABLE OFFSET(0) NUMBITS(1) ], /// Read event Event [ READY OFFSET(0) NUMBITS(1) ], /// Shortcuts Shorts [ // Shortcut between ENDRX and STARTRX ENDRX_STARTRX OFFSET(5) NUMBITS(1), // Shortcut between ENDRX and STOPRX ENDRX_STOPRX OFFSET(6) NUMBITS(1) ], /// UART Interrupts Interrupt [ CTS OFFSET(0) NUMBITS(1), NCTS OFFSET(1) NUMBITS(1), ENDRX OFFSET(4) NUMBITS(1), ENDTX OFFSET(8) NUMBITS(1), ERROR OFFSET(9) NUMBITS(1), RXTO OFFSET(17) NUMBITS(1), RXSTARTED OFFSET(19) NUMBITS(1), TXSTARTED OFFSET(20) NUMBITS(1), TXSTOPPED OFFSET(22) NUMBITS(1) ], /// UART Errors ErrorSrc [ OVERRUN OFFSET(0) NUMBITS(1), PARITY OFFSET(1) NUMBITS(1), FRAMING OFFSET(2) NUMBITS(1), BREAK OFFSET(3) NUMBITS(1) ], /// Enable UART Uart [ ENABLE OFFSET(0) NUMBITS(4) [ ON = 8, OFF = 0 ] ], /// Pin select Psel [ // Pin number PIN OFFSET(0) NUMBITS(5), // Connect/Disconnect CONNECT OFFSET(31) NUMBITS(1) ], /// Baudrate Baudrate [ BAUDRAUTE OFFSET(0) NUMBITS(32) ], /// DMA pointer Pointer [ POINTER OFFSET(0) NUMBITS(32) ], /// Counter value Counter [ COUNTER OFFSET(0) NUMBITS(8) ], /// Configuration of parity and flow control Config [ HWFC OFFSET(0) NUMBITS(1), PARITY OFFSET(1) NUMBITS(3) ] ]; /// UARTE // It should never be instanced outside this module but because a static mutable reference to it // is exported outside this module it must be `pub` pub struct Uarte { regs: *const UarteRegisters, client: Cell<Option<&'static kernel::hil::uart::Client>>, buffer: kernel::common::take_cell::TakeCell<'static, [u8]>, remaining_bytes: Cell<usize>, offset: Cell<usize>, } #[derive(Copy, Clone)] pub struct UARTParams { pub baud_rate: u32, } /// UARTE0 handle // This should only be accessed by the reset_handler on startup pub static mut UARTE0: Uarte = Uarte::new(); impl Uarte { /// Constructor pub const fn new() -> Uarte { Uarte { regs: UARTE_BASE as *const UarteRegisters, client: Cell::new(None), buffer: kernel::common::take_cell::TakeCell::empty(), remaining_bytes: Cell::new(0), offset: Cell::new(0), } } /// Configure which pins the UART should use for txd, rxd, cts and rts pub fn configure( &self, txd: pinmux::Pinmux, rxd: pinmux::Pinmux, cts: pinmux::Pinmux, rts: pinmux::Pinmux, ) { let regs = unsafe { &*self.regs }; regs.pseltxd.write(Psel::PIN.val(txd.into())); regs.pselrxd.write(Psel::PIN.val(rxd.into())); regs.pselcts.write(Psel::PIN.val(cts.into())); regs.pselrts.write(Psel::PIN.val(rts.into())); } fn set_baud_rate(&self, baud_rate: u32) { let regs = unsafe { &*self.regs }; match baud_rate { 1200 => regs.baudrate.set(0x0004F000), 2400 => regs.baudrate.set(0x0009D000), 4800 => regs.baudrate.set(0x0013B000), 9600 => regs.baudrate.set(0x00275000), 14400 => regs.baudrate.set(0x003AF000), 19200 => regs.baudrate.set(0x004EA000), 28800 => regs.baudrate.set(0x0075C000), 38400 => regs.baudrate.set(0x009D0000), 57600 => regs.baudrate.set(0x00EB0000), 76800 => regs.baudrate.set(0x013A9000), 115200 => regs.baudrate.set(0x01D60000), 230400 => regs.baudrate.set(0x03B00000), 250000 => regs.baudrate.set(0x04000000), 460800 => regs.baudrate.set(0x07400000), 921600 => regs.baudrate.set(0x0F000000), 1000000 => regs.baudrate.set(0x10000000), _ => regs.baudrate.set(0x01D60000), //setting default to 115200 } } // Enable UART peripheral, this need to disabled for low power applications fn enable_uart(&self) { let regs = unsafe { &*self.regs }; regs.enable.write(Uart::ENABLE::ON); } #[allow(dead_code)] fn disable_uart(&self) { let regs = unsafe { &*self.regs }; regs.enable.write(Uart::ENABLE::OFF); } #[allow(dead_code)] fn enable_rx_interrupts(&self) { let regs = unsafe { &*self.regs }; regs.intenset.write(Interrupt::ENDRX::SET); } fn enable_tx_interrupts(&self) { let regs = unsafe { &*self.regs }; regs.intenset.write(Interrupt::ENDTX::SET); } #[allow(dead_code)] fn disable_rx_interrupts(&self) { let regs = unsafe { &*self.regs }; regs.intenclr.write(Interrupt::ENDRX::SET); } fn disable_tx_interrupts(&self) { let regs = unsafe { &*self.regs }; regs.intenclr.write(Interrupt::ENDTX::SET); } /// UART interrupt handler that only listens to `tx_end` events #[inline(never)] pub fn handle_interrupt(&mut self) { // disable interrupts self.disable_tx_interrupts(); let regs = unsafe { &*self.regs }; if self.tx_ready()
{ regs.event_endtx.write(Event::READY::CLEAR); let tx_bytes = regs.txd_amount.get() as usize; let rem = self.remaining_bytes.get(); // More bytes transmitted than requested `return silently` // Cause probably a hardware fault // FIXME: Progate error to the capsule if tx_bytes > rem { debug!("error more bytes than requested\r\n"); return; } self.remaining_bytes.set(rem - tx_bytes); self.offset.set(tx_bytes); if self.remaining_bytes.get() == 0 { // Signal client write done self.client.get().map(|client| { self.buffer.take().map(|buffer| {
conditional_block
plonk_util.rs
(x)` for a given `x`, where `n` is the injective function related to the Halo /// endomorphism. pub fn halo_n<C: HaloCurve>(s_bits: &[bool]) -> C::ScalarField { // This is based on Algorithm 2 of the Halo paper, except that we start with (a, b) = (0, 0). debug_assert_eq!(s_bits.len() % 2, 0, "Number of scalar bits must be even"); let zero = C::ScalarField::ZERO; let mut a = zero; let mut b = zero; for s_bits_chunk in s_bits.chunks(2) { let bit_lo = s_bits_chunk[0]; let bit_hi = s_bits_chunk[1]; let sign = if bit_lo { C::ScalarField::ONE } else { C::ScalarField::NEG_ONE }; let (c, d) = if bit_hi { (sign, zero) } else { (zero, sign) }; a = a.double() + c; b = b.double() + d; } a * C::ZETA_SCALAR + b } /// Compute `[n(s)].P` for a given `s`, where `n` is the injective function related to the Halo /// endomorphism. pub fn halo_n_mul<C: HaloCurve>(s_bits: &[bool], p: AffinePoint<C>) -> AffinePoint<C> { // This is based on Algorithm 1 of the Halo paper, except that we start with Acc = O. debug_assert_eq!(s_bits.len() % 2, 0, "Number of scalar bits must be even"); let p_p = p.to_projective(); let p_n = -p_p; let endo_p_p = p.endomorphism().to_projective(); let endo_p_n = -endo_p_p; let mut acc = ProjectivePoint::<C>::ZERO; for s_bits_chunk in s_bits.chunks(2) { let bit_lo = s_bits_chunk[0]; let bit_hi = s_bits_chunk[1]; let s = if bit_hi { if bit_lo { endo_p_p } else { endo_p_n } } else if bit_lo { p_p } else { p_n }; acc = acc.double() + s; } acc.to_affine() } pub fn
<F: Field>(coeffs: &[F], x: F) -> F { let mut ans = F::ZERO; let mut x_pow = F::ONE; for &c in coeffs { ans = ans + (c * x_pow); x_pow = x_pow * x; } ans } /// Compute `[x^0, x^1, ..., x^(n - 1)]`. pub fn powers<F: Field>(x: F, n: usize) -> Vec<F> { let mut powers = Vec::new(); let mut current = F::ONE; for i in 0..n { if i != 0 { current = current * x; } powers.push(current); } powers } /// Compute `[x^0, x^1, ..., x^(n - 1)]`. pub(crate) fn powers_recursive<C: HaloCurve>( builder: &mut CircuitBuilder<C>, x: Target<C::ScalarField>, n: usize, ) -> Vec<Target<C::ScalarField>> { let mut powers = Vec::new(); let mut current = builder.one_wire(); for i in 0..n { if i != 0 { current = builder.mul(current, x); } powers.push(current); } powers } /// Returns the evaluation of a list of polynomials at a point. pub(crate) fn eval_polys<F: Field>(polys: &[Polynomial<F>], powers: &[F]) -> Vec<F> { polys.iter().map(|p| p.eval_from_power(powers)).collect() } /// Zero-pad a list of `n` polynomial coefficients to a length of `8n`, which is the degree at /// which we do most polynomial arithmetic. pub(crate) fn pad_to_8n<F: Field>(coeffs: &[F]) -> Vec<F> { let n = coeffs.len(); let mut result = coeffs.to_vec(); while result.len() < 8 * n { result.push(F::ZERO); } result } pub(crate) fn values_to_polynomials<F: Field>( values_vec: &[Vec<F>], fft_precomputation: &FftPrecomputation<F>, ) -> Vec<Polynomial<F>> { values_vec .par_iter() .map(|values| Polynomial::from_evaluations(values, fft_precomputation)) .collect() } pub(crate) fn polynomials_to_values_padded<F: Field>( polys_vec: &[Polynomial<F>], fft_precomputation: &FftPrecomputation<F>, ) -> Vec<Vec<F>> { polys_vec .par_iter() .map(|poly| { let padded_poly = poly.padded(poly.len() * 8); padded_poly.eval_domain(fft_precomputation) }) .collect() } /// Like `pedersen_commit`, but with no blinding factor. pub fn pedersen_hash<C: Curve>( xs: &[C::ScalarField], pedersen_g_msm_precomputation: &MsmPrecomputation<C>, ) -> ProjectivePoint<C> { msm_execute_parallel(pedersen_g_msm_precomputation, xs) } #[allow(dead_code)] fn pedersen_commit<C: Curve>( xs: &[C::ScalarField], opening: C::ScalarField, h: AffinePoint<C>, pedersen_g_msm_precomputation: &MsmPrecomputation<C>, ) -> ProjectivePoint<C> { // TODO: Couldn't get this working with *. let h = h.to_projective(); let mul_precomputation = h.mul_precompute(); let blinding_term = h.mul_with_precomputation(opening, mul_precomputation); msm_execute_parallel(pedersen_g_msm_precomputation, xs) + blinding_term } pub fn commit_polynomials<C: Curve>( polynomials: &[Polynomial<C::ScalarField>], msm_precomputation: &MsmPrecomputation<C>, blinding_point: AffinePoint<C>, blinding: bool, ) -> Vec<PolynomialCommitment<C>> { PolynomialCommitment::coeffs_vec_to_commitments( polynomials .iter() .map(|p| p.coeffs()) .collect::<Vec<_>>() .as_slice(), msm_precomputation, blinding_point, blinding, ) } // Generate Z, which is used in Plonk's permutation argument. pub fn permutation_polynomial<F: Field>( degree: usize, subgroup: &[F], witness: &Witness<F>, sigma_values: &[Vec<F>], beta: F, gamma: F, ) -> Vec<F> { let mut plonk_z_points = vec![F::ONE]; let k_is = (0..NUM_ROUTED_WIRES) .map(get_subgroup_shift::<F>) .collect::<Vec<_>>(); for i in 1..degree { let x = subgroup[i - 1]; let mut numerator = F::ONE; let mut denominator = F::ONE; for j in 0..NUM_ROUTED_WIRES { let wire_value = witness.get_indices(i - 1, j); let k_i = k_is[j]; let s_id = k_i * x; let s_sigma = sigma_values[j][8 * (i - 1)]; numerator = numerator * (wire_value + beta * s_id + gamma); denominator = denominator * (wire_value + beta * s_sigma + gamma); } let last = *plonk_z_points.last().unwrap(); plonk_z_points.push(last * numerator / denominator); } plonk_z_points } pub fn sigma_polynomials<F: Field>( sigma: Vec<usize>, degree: usize, subgroup_generator: F, ) -> Vec<Vec<F>> { sigma .chunks(degree) .map(|chunk| { chunk .par_iter() .map(|&x| { get_subgroup_shift::<F>(x / degree) * subgroup_generator.exp_usize(x % degree) }) .collect::<Vec<_>>() }) .collect() } /// Given polynomials `[p_0,...,p_k]` of degree `degree` and `alpha \in F`, returns `\sum_{i=0}^k alpha^i p_i`. pub(crate) fn scale_polynomials<F: Field>( polynomials: Vec<Polynomial<F>>, alpha: F, degree: usize, ) -> Polynomial<F> { let alpha_powers = powers(alpha, polynomials.len()); Polynomial::from( (0..degree) .map(|i| { (0..polynomials.len()) .map(|j| polynomials[j][i] * alpha_powers[j]) .fold(F::ZERO, |acc, x| acc + x) }) .collect::<Vec<_>>(), ) } #[allow(dead_code
eval_poly
identifier_name
plonk_util.rs
ProjectivePoint<C> { msm_execute_parallel(pedersen_g_msm_precomputation, xs) } #[allow(dead_code)] fn pedersen_commit<C: Curve>( xs: &[C::ScalarField], opening: C::ScalarField, h: AffinePoint<C>, pedersen_g_msm_precomputation: &MsmPrecomputation<C>, ) -> ProjectivePoint<C> { // TODO: Couldn't get this working with *. let h = h.to_projective(); let mul_precomputation = h.mul_precompute(); let blinding_term = h.mul_with_precomputation(opening, mul_precomputation); msm_execute_parallel(pedersen_g_msm_precomputation, xs) + blinding_term } pub fn commit_polynomials<C: Curve>( polynomials: &[Polynomial<C::ScalarField>], msm_precomputation: &MsmPrecomputation<C>, blinding_point: AffinePoint<C>, blinding: bool, ) -> Vec<PolynomialCommitment<C>> { PolynomialCommitment::coeffs_vec_to_commitments( polynomials .iter() .map(|p| p.coeffs()) .collect::<Vec<_>>() .as_slice(), msm_precomputation, blinding_point, blinding, ) } // Generate Z, which is used in Plonk's permutation argument. pub fn permutation_polynomial<F: Field>( degree: usize, subgroup: &[F], witness: &Witness<F>, sigma_values: &[Vec<F>], beta: F, gamma: F, ) -> Vec<F> { let mut plonk_z_points = vec![F::ONE]; let k_is = (0..NUM_ROUTED_WIRES) .map(get_subgroup_shift::<F>) .collect::<Vec<_>>(); for i in 1..degree { let x = subgroup[i - 1]; let mut numerator = F::ONE; let mut denominator = F::ONE; for j in 0..NUM_ROUTED_WIRES { let wire_value = witness.get_indices(i - 1, j); let k_i = k_is[j]; let s_id = k_i * x; let s_sigma = sigma_values[j][8 * (i - 1)]; numerator = numerator * (wire_value + beta * s_id + gamma); denominator = denominator * (wire_value + beta * s_sigma + gamma); } let last = *plonk_z_points.last().unwrap(); plonk_z_points.push(last * numerator / denominator); } plonk_z_points } pub fn sigma_polynomials<F: Field>( sigma: Vec<usize>, degree: usize, subgroup_generator: F, ) -> Vec<Vec<F>> { sigma .chunks(degree) .map(|chunk| { chunk .par_iter() .map(|&x| { get_subgroup_shift::<F>(x / degree) * subgroup_generator.exp_usize(x % degree) }) .collect::<Vec<_>>() }) .collect() } /// Given polynomials `[p_0,...,p_k]` of degree `degree` and `alpha \in F`, returns `\sum_{i=0}^k alpha^i p_i`. pub(crate) fn scale_polynomials<F: Field>( polynomials: Vec<Polynomial<F>>, alpha: F, degree: usize, ) -> Polynomial<F> { let alpha_powers = powers(alpha, polynomials.len()); Polynomial::from( (0..degree) .map(|i| { (0..polynomials.len()) .map(|j| polynomials[j][i] * alpha_powers[j]) .fold(F::ZERO, |acc, x| acc + x) }) .collect::<Vec<_>>(), ) } #[allow(dead_code)] pub(crate) fn polynomial_degree_plus_1<F: Field>( points: &[F], fft_precomputation: &FftPrecomputation<F>, ) -> usize { let coeffs = ifft_with_precomputation_power_of_2(points, fft_precomputation); coeffs.iter().rev().skip_while(|c| c.is_zero()).count() } // TODO: Maybe a streaming version using an `Iterator` would be faster and wouldn't require as much memory for large circuits. // TODO: Optimize this. pub fn halo_s<F: Field>(us: &[F]) -> Vec<F> { let n = 1 << us.len(); let mut res = vec![F::ONE; n]; let us_inv = F::batch_multiplicative_inverse(us); for (j, (&u, &u_inv)) in us.iter().rev().zip(us_inv.iter().rev()).enumerate() { for (i, x) in res.iter_mut().enumerate() { if i & (1 << j) == 0 { *x = *x * u_inv; } else { *x = *x * u; } } } res } /// Evaluate `g(X, {u_i})` as defined in the Halo paper. pub fn halo_g<F: Field>(x: F, us: &[F]) -> F { let mut product = F::ONE; let mut x_power = x; for &u_i in us.iter().rev() { let u_i_inv = u_i.multiplicative_inverse_assuming_nonzero(); let term = u_i * x_power + u_i_inv; product = product * term; x_power = x_power.square(); } product } #[cfg(test)] mod test { use super::*; use crate::{CircuitBuilder, Curve, Field, PartialWitness, Tweedledee}; #[test] fn test_halo_n() { type C = Tweedledee; type SF = <Tweedledee as Curve>::ScalarField; let p = C::convert(SF::rand()) * C::GENERATOR_PROJECTIVE; let r = SF::rand(); let res = C::convert(halo_n::<C>(&r.to_canonical_bool_vec()[..128])) * p; let p = p.to_affine(); assert_eq!( res.to_affine(), halo_n_mul::<C>(&r.to_canonical_bool_vec()[..128], p) ) } #[test] fn test_permutation_polynomial() { let mut builder = CircuitBuilder::<Tweedledee>::new(128); let one = builder.one_wire(); let t = builder.add_virtual_target(); let t_sq = builder.square(t); let quad = builder.add_many(&[one, t, t_sq]); let seven = builder.constant_wire(<Tweedledee as Curve>::ScalarField::from_canonical_usize(7)); let res = builder.sub(quad, seven); builder.assert_zero(res); let mut partial_witness = PartialWitness::new(); partial_witness.set_target(t, <Tweedledee as Curve>::ScalarField::TWO); let circuit = builder.build(); let witness = circuit.generate_witness(partial_witness); let beta = <Tweedledee as Curve>::ScalarField::rand(); let gamma = <Tweedledee as Curve>::ScalarField::rand(); let plonk_z_points_n = permutation_polynomial( circuit.degree(), &circuit.subgroup_n, &witness, &circuit.s_sigma_values_8n, beta, gamma, ); // Verify that the permutation polynomial is well-formed. let k_is = (0..NUM_ROUTED_WIRES) .map(get_subgroup_shift::<<Tweedledee as Curve>::ScalarField>) .collect::<Vec<_>>(); let wire_values = &witness.transpose(); for (i, &x) in circuit.subgroup_n.iter().enumerate() { let (z_x, z_gz) = ( plonk_z_points_n[i], plonk_z_points_n[(i + 1) % circuit.degree()], ); let mut f_prime = <Tweedledee as Curve>::ScalarField::ONE; let mut g_prime = <Tweedledee as Curve>::ScalarField::ONE; for j in 0..NUM_ROUTED_WIRES { let wire_value = wire_values[j][i]; let k_i = k_is[j]; let s_id = k_i * x; let s_sigma = circuit.s_sigma_values_8n[j][8 * i]; f_prime = f_prime * (wire_value + beta * s_id + gamma); g_prime = g_prime * (wire_value + beta * s_sigma + gamma); } let vanishing_v_shift_term = f_prime * z_x - g_prime * z_gz; assert_eq!( vanishing_v_shift_term, <Tweedledee as Curve>::ScalarField::ZERO ); } } #[test] fn test_s_vector_g_function()
{ type F = <Tweedledee as Curve>::ScalarField; let us = (0..10).map(|_| F::rand()).collect::<Vec<_>>(); let x = F::rand(); assert_eq!( F::inner_product(&halo_s(&us), &powers(x, 1 << 10)), halo_g(x, &us) ); }
identifier_body
plonk_util.rs
n(x)` for a given `x`, where `n` is the injective function related to the Halo /// endomorphism. pub fn halo_n<C: HaloCurve>(s_bits: &[bool]) -> C::ScalarField { // This is based on Algorithm 2 of the Halo paper, except that we start with (a, b) = (0, 0). debug_assert_eq!(s_bits.len() % 2, 0, "Number of scalar bits must be even"); let zero = C::ScalarField::ZERO; let mut a = zero; let mut b = zero; for s_bits_chunk in s_bits.chunks(2) { let bit_lo = s_bits_chunk[0]; let bit_hi = s_bits_chunk[1]; let sign = if bit_lo { C::ScalarField::ONE } else { C::ScalarField::NEG_ONE }; let (c, d) = if bit_hi { (sign, zero) } else { (zero, sign) }; a = a.double() + c; b = b.double() + d; } a * C::ZETA_SCALAR + b } /// Compute `[n(s)].P` for a given `s`, where `n` is the injective function related to the Halo /// endomorphism. pub fn halo_n_mul<C: HaloCurve>(s_bits: &[bool], p: AffinePoint<C>) -> AffinePoint<C> { // This is based on Algorithm 1 of the Halo paper, except that we start with Acc = O. debug_assert_eq!(s_bits.len() % 2, 0, "Number of scalar bits must be even"); let p_p = p.to_projective(); let p_n = -p_p; let endo_p_p = p.endomorphism().to_projective(); let endo_p_n = -endo_p_p; let mut acc = ProjectivePoint::<C>::ZERO; for s_bits_chunk in s_bits.chunks(2) { let bit_lo = s_bits_chunk[0]; let bit_hi = s_bits_chunk[1];
if bit_lo { endo_p_p } else { endo_p_n } } else if bit_lo { p_p } else { p_n }; acc = acc.double() + s; } acc.to_affine() } pub fn eval_poly<F: Field>(coeffs: &[F], x: F) -> F { let mut ans = F::ZERO; let mut x_pow = F::ONE; for &c in coeffs { ans = ans + (c * x_pow); x_pow = x_pow * x; } ans } /// Compute `[x^0, x^1, ..., x^(n - 1)]`. pub fn powers<F: Field>(x: F, n: usize) -> Vec<F> { let mut powers = Vec::new(); let mut current = F::ONE; for i in 0..n { if i != 0 { current = current * x; } powers.push(current); } powers } /// Compute `[x^0, x^1, ..., x^(n - 1)]`. pub(crate) fn powers_recursive<C: HaloCurve>( builder: &mut CircuitBuilder<C>, x: Target<C::ScalarField>, n: usize, ) -> Vec<Target<C::ScalarField>> { let mut powers = Vec::new(); let mut current = builder.one_wire(); for i in 0..n { if i != 0 { current = builder.mul(current, x); } powers.push(current); } powers } /// Returns the evaluation of a list of polynomials at a point. pub(crate) fn eval_polys<F: Field>(polys: &[Polynomial<F>], powers: &[F]) -> Vec<F> { polys.iter().map(|p| p.eval_from_power(powers)).collect() } /// Zero-pad a list of `n` polynomial coefficients to a length of `8n`, which is the degree at /// which we do most polynomial arithmetic. pub(crate) fn pad_to_8n<F: Field>(coeffs: &[F]) -> Vec<F> { let n = coeffs.len(); let mut result = coeffs.to_vec(); while result.len() < 8 * n { result.push(F::ZERO); } result } pub(crate) fn values_to_polynomials<F: Field>( values_vec: &[Vec<F>], fft_precomputation: &FftPrecomputation<F>, ) -> Vec<Polynomial<F>> { values_vec .par_iter() .map(|values| Polynomial::from_evaluations(values, fft_precomputation)) .collect() } pub(crate) fn polynomials_to_values_padded<F: Field>( polys_vec: &[Polynomial<F>], fft_precomputation: &FftPrecomputation<F>, ) -> Vec<Vec<F>> { polys_vec .par_iter() .map(|poly| { let padded_poly = poly.padded(poly.len() * 8); padded_poly.eval_domain(fft_precomputation) }) .collect() } /// Like `pedersen_commit`, but with no blinding factor. pub fn pedersen_hash<C: Curve>( xs: &[C::ScalarField], pedersen_g_msm_precomputation: &MsmPrecomputation<C>, ) -> ProjectivePoint<C> { msm_execute_parallel(pedersen_g_msm_precomputation, xs) } #[allow(dead_code)] fn pedersen_commit<C: Curve>( xs: &[C::ScalarField], opening: C::ScalarField, h: AffinePoint<C>, pedersen_g_msm_precomputation: &MsmPrecomputation<C>, ) -> ProjectivePoint<C> { // TODO: Couldn't get this working with *. let h = h.to_projective(); let mul_precomputation = h.mul_precompute(); let blinding_term = h.mul_with_precomputation(opening, mul_precomputation); msm_execute_parallel(pedersen_g_msm_precomputation, xs) + blinding_term } pub fn commit_polynomials<C: Curve>( polynomials: &[Polynomial<C::ScalarField>], msm_precomputation: &MsmPrecomputation<C>, blinding_point: AffinePoint<C>, blinding: bool, ) -> Vec<PolynomialCommitment<C>> { PolynomialCommitment::coeffs_vec_to_commitments( polynomials .iter() .map(|p| p.coeffs()) .collect::<Vec<_>>() .as_slice(), msm_precomputation, blinding_point, blinding, ) } // Generate Z, which is used in Plonk's permutation argument. pub fn permutation_polynomial<F: Field>( degree: usize, subgroup: &[F], witness: &Witness<F>, sigma_values: &[Vec<F>], beta: F, gamma: F, ) -> Vec<F> { let mut plonk_z_points = vec![F::ONE]; let k_is = (0..NUM_ROUTED_WIRES) .map(get_subgroup_shift::<F>) .collect::<Vec<_>>(); for i in 1..degree { let x = subgroup[i - 1]; let mut numerator = F::ONE; let mut denominator = F::ONE; for j in 0..NUM_ROUTED_WIRES { let wire_value = witness.get_indices(i - 1, j); let k_i = k_is[j]; let s_id = k_i * x; let s_sigma = sigma_values[j][8 * (i - 1)]; numerator = numerator * (wire_value + beta * s_id + gamma); denominator = denominator * (wire_value + beta * s_sigma + gamma); } let last = *plonk_z_points.last().unwrap(); plonk_z_points.push(last * numerator / denominator); } plonk_z_points } pub fn sigma_polynomials<F: Field>( sigma: Vec<usize>, degree: usize, subgroup_generator: F, ) -> Vec<Vec<F>> { sigma .chunks(degree) .map(|chunk| { chunk .par_iter() .map(|&x| { get_subgroup_shift::<F>(x / degree) * subgroup_generator.exp_usize(x % degree) }) .collect::<Vec<_>>() }) .collect() } /// Given polynomials `[p_0,...,p_k]` of degree `degree` and `alpha \in F`, returns `\sum_{i=0}^k alpha^i p_i`. pub(crate) fn scale_polynomials<F: Field>( polynomials: Vec<Polynomial<F>>, alpha: F, degree: usize, ) -> Polynomial<F> { let alpha_powers = powers(alpha, polynomials.len()); Polynomial::from( (0..degree) .map(|i| { (0..polynomials.len()) .map(|j| polynomials[j][i] * alpha_powers[j]) .fold(F::ZERO, |acc, x| acc + x) }) .collect::<Vec<_>>(), ) } #[allow(dead_code
let s = if bit_hi {
random_line_split
plonk_util.rs
(x)` for a given `x`, where `n` is the injective function related to the Halo /// endomorphism. pub fn halo_n<C: HaloCurve>(s_bits: &[bool]) -> C::ScalarField { // This is based on Algorithm 2 of the Halo paper, except that we start with (a, b) = (0, 0). debug_assert_eq!(s_bits.len() % 2, 0, "Number of scalar bits must be even"); let zero = C::ScalarField::ZERO; let mut a = zero; let mut b = zero; for s_bits_chunk in s_bits.chunks(2) { let bit_lo = s_bits_chunk[0]; let bit_hi = s_bits_chunk[1]; let sign = if bit_lo { C::ScalarField::ONE } else
; let (c, d) = if bit_hi { (sign, zero) } else { (zero, sign) }; a = a.double() + c; b = b.double() + d; } a * C::ZETA_SCALAR + b } /// Compute `[n(s)].P` for a given `s`, where `n` is the injective function related to the Halo /// endomorphism. pub fn halo_n_mul<C: HaloCurve>(s_bits: &[bool], p: AffinePoint<C>) -> AffinePoint<C> { // This is based on Algorithm 1 of the Halo paper, except that we start with Acc = O. debug_assert_eq!(s_bits.len() % 2, 0, "Number of scalar bits must be even"); let p_p = p.to_projective(); let p_n = -p_p; let endo_p_p = p.endomorphism().to_projective(); let endo_p_n = -endo_p_p; let mut acc = ProjectivePoint::<C>::ZERO; for s_bits_chunk in s_bits.chunks(2) { let bit_lo = s_bits_chunk[0]; let bit_hi = s_bits_chunk[1]; let s = if bit_hi { if bit_lo { endo_p_p } else { endo_p_n } } else if bit_lo { p_p } else { p_n }; acc = acc.double() + s; } acc.to_affine() } pub fn eval_poly<F: Field>(coeffs: &[F], x: F) -> F { let mut ans = F::ZERO; let mut x_pow = F::ONE; for &c in coeffs { ans = ans + (c * x_pow); x_pow = x_pow * x; } ans } /// Compute `[x^0, x^1, ..., x^(n - 1)]`. pub fn powers<F: Field>(x: F, n: usize) -> Vec<F> { let mut powers = Vec::new(); let mut current = F::ONE; for i in 0..n { if i != 0 { current = current * x; } powers.push(current); } powers } /// Compute `[x^0, x^1, ..., x^(n - 1)]`. pub(crate) fn powers_recursive<C: HaloCurve>( builder: &mut CircuitBuilder<C>, x: Target<C::ScalarField>, n: usize, ) -> Vec<Target<C::ScalarField>> { let mut powers = Vec::new(); let mut current = builder.one_wire(); for i in 0..n { if i != 0 { current = builder.mul(current, x); } powers.push(current); } powers } /// Returns the evaluation of a list of polynomials at a point. pub(crate) fn eval_polys<F: Field>(polys: &[Polynomial<F>], powers: &[F]) -> Vec<F> { polys.iter().map(|p| p.eval_from_power(powers)).collect() } /// Zero-pad a list of `n` polynomial coefficients to a length of `8n`, which is the degree at /// which we do most polynomial arithmetic. pub(crate) fn pad_to_8n<F: Field>(coeffs: &[F]) -> Vec<F> { let n = coeffs.len(); let mut result = coeffs.to_vec(); while result.len() < 8 * n { result.push(F::ZERO); } result } pub(crate) fn values_to_polynomials<F: Field>( values_vec: &[Vec<F>], fft_precomputation: &FftPrecomputation<F>, ) -> Vec<Polynomial<F>> { values_vec .par_iter() .map(|values| Polynomial::from_evaluations(values, fft_precomputation)) .collect() } pub(crate) fn polynomials_to_values_padded<F: Field>( polys_vec: &[Polynomial<F>], fft_precomputation: &FftPrecomputation<F>, ) -> Vec<Vec<F>> { polys_vec .par_iter() .map(|poly| { let padded_poly = poly.padded(poly.len() * 8); padded_poly.eval_domain(fft_precomputation) }) .collect() } /// Like `pedersen_commit`, but with no blinding factor. pub fn pedersen_hash<C: Curve>( xs: &[C::ScalarField], pedersen_g_msm_precomputation: &MsmPrecomputation<C>, ) -> ProjectivePoint<C> { msm_execute_parallel(pedersen_g_msm_precomputation, xs) } #[allow(dead_code)] fn pedersen_commit<C: Curve>( xs: &[C::ScalarField], opening: C::ScalarField, h: AffinePoint<C>, pedersen_g_msm_precomputation: &MsmPrecomputation<C>, ) -> ProjectivePoint<C> { // TODO: Couldn't get this working with *. let h = h.to_projective(); let mul_precomputation = h.mul_precompute(); let blinding_term = h.mul_with_precomputation(opening, mul_precomputation); msm_execute_parallel(pedersen_g_msm_precomputation, xs) + blinding_term } pub fn commit_polynomials<C: Curve>( polynomials: &[Polynomial<C::ScalarField>], msm_precomputation: &MsmPrecomputation<C>, blinding_point: AffinePoint<C>, blinding: bool, ) -> Vec<PolynomialCommitment<C>> { PolynomialCommitment::coeffs_vec_to_commitments( polynomials .iter() .map(|p| p.coeffs()) .collect::<Vec<_>>() .as_slice(), msm_precomputation, blinding_point, blinding, ) } // Generate Z, which is used in Plonk's permutation argument. pub fn permutation_polynomial<F: Field>( degree: usize, subgroup: &[F], witness: &Witness<F>, sigma_values: &[Vec<F>], beta: F, gamma: F, ) -> Vec<F> { let mut plonk_z_points = vec![F::ONE]; let k_is = (0..NUM_ROUTED_WIRES) .map(get_subgroup_shift::<F>) .collect::<Vec<_>>(); for i in 1..degree { let x = subgroup[i - 1]; let mut numerator = F::ONE; let mut denominator = F::ONE; for j in 0..NUM_ROUTED_WIRES { let wire_value = witness.get_indices(i - 1, j); let k_i = k_is[j]; let s_id = k_i * x; let s_sigma = sigma_values[j][8 * (i - 1)]; numerator = numerator * (wire_value + beta * s_id + gamma); denominator = denominator * (wire_value + beta * s_sigma + gamma); } let last = *plonk_z_points.last().unwrap(); plonk_z_points.push(last * numerator / denominator); } plonk_z_points } pub fn sigma_polynomials<F: Field>( sigma: Vec<usize>, degree: usize, subgroup_generator: F, ) -> Vec<Vec<F>> { sigma .chunks(degree) .map(|chunk| { chunk .par_iter() .map(|&x| { get_subgroup_shift::<F>(x / degree) * subgroup_generator.exp_usize(x % degree) }) .collect::<Vec<_>>() }) .collect() } /// Given polynomials `[p_0,...,p_k]` of degree `degree` and `alpha \in F`, returns `\sum_{i=0}^k alpha^i p_i`. pub(crate) fn scale_polynomials<F: Field>( polynomials: Vec<Polynomial<F>>, alpha: F, degree: usize, ) -> Polynomial<F> { let alpha_powers = powers(alpha, polynomials.len()); Polynomial::from( (0..degree) .map(|i| { (0..polynomials.len()) .map(|j| polynomials[j][i] * alpha_powers[j]) .fold(F::ZERO, |acc, x| acc + x) }) .collect::<Vec<_>>(), ) } #[allow(dead
{ C::ScalarField::NEG_ONE }
conditional_block
Zhihu__Crawler_JsonVersion.py
File) self.opener = request.build_opener(request.HTTPCookieProcessor(self.cj), RedirectHandler()) self.client_info = 'monsterzpc@gmail.com' self.passwd = 'Zpc920515' self.url = 'https://www.zhihu.com/login/email' self.target_page = '' self.position = 0 print(''' ############################################################ # # # Zhihu Auto_Login and Crawler by Pengcheng Zhou. # # # ############################################################ ''') def get_xsrf(self) : ''' Get a special dynamic string for login ''' login_target_page = request.urlopen(self.url) pattern = re.compile('<input type="hidden" name="_xsrf" value="(.*)"/>') _xsrf = re.findall(pattern, login_target_page.read().decode('utf-8'))[0] return _xsrf def get_captcha_url(self): url = 'https://www.zhihu.com' + '/captcha.gif?r=' + str(int(time.time())) + '&type=login' f = request.urlopen(url) with open('./cap.png', 'wb') as image: image.write(f.read()) image.close() def login(self): ''' Execution of login ''' if (self.client_info == '' or self.passwd == '') : self.client_info = input('请输入账号:') self.passwd = getpass('请输入密码:') self.get_captcha_url() captcha = input('请输入验证码:') if (self.client_info.find("@") != -1) : print('''正在使用邮箱登录...\n用户名:''' + self.client_info+ '\n' + '密码 : ' + len(self.passwd) * '*'+ '\n' ) else : self.url = '' print('正在使用手机登录...') form = {'_xsrf' : self.get_xsrf(), 'password' : self.passwd, 'email' : self.client_info, 'captcha': captcha } print(form) try: req = request.Request(self.url, parse.urlencode(form).encode('utf-8')) f = self.opener.open(req) self.cj.save() print(json.loads(f.read().decode('utf-8'))["msg"] + "!") print("=" * 100) except: print('Error!') def get_capthca(self) : ''' Interface for getting the captcha ''' pass def get_target_page(self): ''' Get main target_page content after logged in ''' try: self.cj.load() print('Cookie loaded....') self.target_page = self.opener.open('https://www.zhihu.com/people/edit') f = open('zhihu.html', 'wb') f.write(target_page_content.read()) except: self.login() self.get_target_page() def isLogged(self, user_client): ''' test if Logged ''' f = user_client.open('https://www.zhihu.com/settings/profile').geturl() if (f != 'https://www.zhihu.com/settings/profile'): return False return True def user_getter(sefl): return User_Getter('https://www.zhihu.com/people/xiao-guai-shou-2/activ
s() def profile_collector(self, text_path=None) : ''' main entry for collecting user's profile including id, gender, education, career ''' count = 0 self.cj.load() user_list = [] # check the source of the data if (text_path != None) : with open(text_path, 'r') as source_list : for line in source_list : user_list.append(line.split('\n')[0]) source_list.close() else : user_list = [] initial_time = time.time() while(len(user_list) > 0): # this try except block is for resuming from the server' shutdown try : for item in user_list : start_time = time.time() user_id = item print('=Writing information of [', user_id,']...') url = 'https://www.zhihu.com/api/v4/members/' + user_id + '?include=locations%2Cemployments%2Cgender%2Ceducations%2Cbusiness%2Cvoteup_count%2Cthanked_Count%2Cfollower_count%2Cfollowing_count%2Ccover_url%2Cfollowing_topic_count%2Cfollowing_question_count%2Cfollowing_favlists_count%2Cfollowing_columns_count%2Canswer_count%2Carticles_count%2Cpins_count%2Cquestion_count%2Ccommercial_question_count%2Cfavorite_count%2Cfavorited_count%2Clogs_count%2Cmarked_answers_count%2Cmarked_answers_text%2Cmessage_thread_token%2Caccount_status%2Cis_active%2Cis_force_renamed%2Cis_bind_sina%2Csina_weibo_url%2Csina_weibo_name%2Cshow_sina_weibo%2Cis_blocking%2Cis_blocked%2Cis_following%2Cis_followed%2Cmutual_followees_count%2Cvote_to_count%2Cvote_from_count%2Cthank_to_count%2Cthank_from_count%2Cthanked_count%2Cdescription%2Chosted_live_count%2Cparticipated_live_count%2Callow_message%2Cindustry_category%2Corg_name%2Corg_homepage%2Cbadge%5B%3F(type%3Dbest_answerer)%5D.topics' req = request.Request(url) raw_data = self.opener.open(req).read() json_data = json.loads(raw_data) # get key and value pic_url = json_data["avatar_url"].split('_')[0] + '_xll.jpg' number_id = json_data["id"] user_name = json_data["name"] # education if ("educations" in json_data) : if (len(json_data["educations"]) != 0) : if ("school" in json_data["educations"][0]) : university = json_data["educations"][0]["school"]["name"] else : university = 'None' if ("major" in json_data["educations"][0]) : major = json_data["educations"][0]["major"]["name"] else: major = 'None' else : university = 'None' major = 'None' else : university = 'None' major = 'None' # employments if ("employments" in json_data) : if (len(json_data["employments"]) != 0) : if ("company" in json_data["employments"][0]) : company = json_data["employments"][0]["company"]["name"] else : company = 'None' if ("occupation" in json_data["employments"][0]) : occupation = json_data["employments"][0]["job"]["name"] else : occupation = 'None' else : company = 'None' occupation = 'None' else : company = 'None' occupation = 'None' # location if ("locations" in json_data) : if (len(json_data["locations"]) != 0) : location = json_data["locations"][0]["name"] else : location = 'None' else : location = 'None' # business if ("business" in json_data ) : industry = json_data["business"]["name"] else : industry = 'None' intro = json_data["headline"] autobiography = json_data["description"] user_type = json_data["type"] follower_count = json_data["follower_count"] following_count = json_data["following_count"] answers_count = json_data["answer_count"] articles_count = json_data["articles_count"] if (json_data["gender"] == 1) : gender = 'male' else : gender = 'female' data = { 'id' : number_id, 'user_id' : user_id, 'name' : user_name, 'gender' : gender, 'university' : university, 'major' : major, 'industry' : industry, 'company' : company, 'occupation' : occupation, 'location' : location, 'intro' : intro, 'autobiography' : autobiography, 'user_type' : str(user_type), 'follower_count' : str(follower_count), 'following_count' : str(following_count), 'answer-count' : str(answers_count), 'articles_count' : str(articles_count) } # process folder if not (os.path.exists(os.path.join('./data/' ,user_name
ities').url
identifier_name
Zhihu__Crawler_JsonVersion.py
File) self.opener = request.build_opener(request.HTTPCookieProcessor(self.cj), RedirectHandler()) self.client_info = 'monsterzpc@gmail.com' self.passwd = 'Zpc920515' self.url = 'https://www.zhihu.com/login/email' self.target_page = '' self.position = 0 print(''' ############################################################ # # # Zhihu Auto_Login and Crawler by Pengcheng Zhou. # # # ############################################################ ''') def get_xsrf(self) : ''' Get a special dynamic string for login ''' login_target_page = request.urlopen(self.url) pattern = re.compile('<input type="hidden" name="_xsrf" value="(.*)"/>') _xsrf = re.findall(pattern, login_target_page.read().decode('utf-8'))[0] return _xsrf def get_captcha_url(self): url = 'https://www.zhihu.com' + '/captcha.gif?r=' + str(int(time.time())) + '&type=login' f = request.urlopen(url) with open('./cap.png', 'wb') as image: image.write(f.read()) image.close() def login(self): ''' Execution of login ''' if (self.client_info == '' or self.passwd == '') : self.client_info = input('请输入账号:') self.passwd = getpass('请输入密码:') self.get_captcha_url() captcha = input('请输入验证码:') if (self.client_info.find("@") != -1) : print('''正在使用邮箱登录...\n用户名:''' + self.client_info+ '\n' + '密码 : ' + len(self.passwd) * '*'+ '\n' ) else : self.url = '' print('正在使用手机登录...') form = {'_xsrf' : self.get_xsrf(), 'password' : self.passwd, 'email' : self.client_info, 'captcha': captcha } print(form) try: req = request.Request(self.url, parse.urlencode(form).encode('utf-8')) f = self.opener.open(req) self.cj.save() print(json.loads(f.read().decode('utf-8'))["msg"] + "!") print("=" * 100) except: print('Error!') def get_capthca(self) : ''' Interface for getting the captcha ''' pass def get_target_page(self): ''' Get main target_page content after logged in ''' try: self.cj.load() print('Cookie loaded....') self.target_page = self.opener.open('https://www.zhihu.com/people/edit') f = open('zhihu.html', 'wb') f.write(target_page_content.read()) except: self.login() self.get_target_page() def isLogged(self, user_client): ''' test if Logged ''' f = user_client.open('https://www.zhihu.com/settings/profile').geturl() if (f != 'https://www.zhihu.com/settings/profile'): return False return True
def user_getter(sefl): return User_Getter('https://www.zhihu.com/people/xiao-guai-shou-2/activities').urls() def profile_collector(self, text_path=None) : ''' main entry for collecting user's profile including id, gender, education, career ''' count = 0 self.cj.load() user_list = [] # check the source of the data if (text_path != None) : with open(text_path, 'r') as source_list : for line in source_list : user_list.append(line.split('\n')[0]) source_list.close() else : user_list = [] initial_time = time.time() while(len(user_list) > 0): # this try except block is for resuming from the server' shutdown try : for item in user_list : start_time = time.time() user_id = item print('=Writing information of [', user_id,']...') url = 'https://www.zhihu.com/api/v4/members/' + user_id + '?include=locations%2Cemployments%2Cgender%2Ceducations%2Cbusiness%2Cvoteup_count%2Cthanked_Count%2Cfollower_count%2Cfollowing_count%2Ccover_url%2Cfollowing_topic_count%2Cfollowing_question_count%2Cfollowing_favlists_count%2Cfollowing_columns_count%2Canswer_count%2Carticles_count%2Cpins_count%2Cquestion_count%2Ccommercial_question_count%2Cfavorite_count%2Cfavorited_count%2Clogs_count%2Cmarked_answers_count%2Cmarked_answers_text%2Cmessage_thread_token%2Caccount_status%2Cis_active%2Cis_force_renamed%2Cis_bind_sina%2Csina_weibo_url%2Csina_weibo_name%2Cshow_sina_weibo%2Cis_blocking%2Cis_blocked%2Cis_following%2Cis_followed%2Cmutual_followees_count%2Cvote_to_count%2Cvote_from_count%2Cthank_to_count%2Cthank_from_count%2Cthanked_count%2Cdescription%2Chosted_live_count%2Cparticipated_live_count%2Callow_message%2Cindustry_category%2Corg_name%2Corg_homepage%2Cbadge%5B%3F(type%3Dbest_answerer)%5D.topics' req = request.Request(url) raw_data = self.opener.open(req).read() json_data = json.loads(raw_data) # get key and value pic_url = json_data["avatar_url"].split('_')[0] + '_xll.jpg' number_id = json_data["id"] user_name = json_data["name"] # education if ("educations" in json_data) : if (len(json_data["educations"]) != 0) : if ("school" in json_data["educations"][0]) : university = json_data["educations"][0]["school"]["name"] else : university = 'None' if ("major" in json_data["educations"][0]) : major = json_data["educations"][0]["major"]["name"] else: major = 'None' else : university = 'None' major = 'None' else : university = 'None' major = 'None' # employments if ("employments" in json_data) : if (len(json_data["employments"]) != 0) : if ("company" in json_data["employments"][0]) : company = json_data["employments"][0]["company"]["name"] else : company = 'None' if ("occupation" in json_data["employments"][0]) : occupation = json_data["employments"][0]["job"]["name"] else : occupation = 'None' else : company = 'None' occupation = 'None' else : company = 'None' occupation = 'None' # location if ("locations" in json_data) : if (len(json_data["locations"]) != 0) : location = json_data["locations"][0]["name"] else : location = 'None' else : location = 'None' # business if ("business" in json_data ) : industry = json_data["business"]["name"] else : industry = 'None' intro = json_data["headline"] autobiography = json_data["description"] user_type = json_data["type"] follower_count = json_data["follower_count"] following_count = json_data["following_count"] answers_count = json_data["answer_count"] articles_count = json_data["articles_count"] if (json_data["gender"] == 1) : gender = 'male' else : gender = 'female' data = { 'id' : number_id, 'user_id' : user_id, 'name' : user_name, 'gender' : gender, 'university' : university, 'major' : major, 'industry' : industry, 'company' : company, 'occupation' : occupation, 'location' : location, 'intro' : intro, 'autobiography' : autobiography, 'user_type' : str(user_type), 'follower_count' : str(follower_count), 'following_count' : str(following_count), 'answer-count' : str(answers_count), 'articles_count' : str(articles_count) } # process folder if not (os.path.exists(os.path.join('./data/' ,user_name))
conditional_block
Zhihu__Crawler_JsonVersion.py
File) self.opener = request.build_opener(request.HTTPCookieProcessor(self.cj), RedirectHandler()) self.client_info = 'monsterzpc@gmail.com' self.passwd = 'Zpc920515' self.url = 'https://www.zhihu.com/login/email' self.target_page = '' self.position = 0 print(''' ############################################################ # # # Zhihu Auto_Login and Crawler by Pengcheng Zhou. # # # ############################################################ ''') def get_xsrf(self) : ''' Get a special dynamic string for login ''' login_target_page = request.urlopen(self.url) pattern = re.compile('<input type="hidden" name="_xsrf" value="(.*)"/>') _xsrf = re.findall(pattern, login_target_page.read().decode('utf-8'))[0] return _xsrf def get_captcha_url(self): url = 'https://
f): ''' Execution of login ''' if (self.client_info == '' or self.passwd == '') : self.client_info = input('请输入账号:') self.passwd = getpass('请输入密码:') self.get_captcha_url() captcha = input('请输入验证码:') if (self.client_info.find("@") != -1) : print('''正在使用邮箱登录...\n用户名:''' + self.client_info+ '\n' + '密码 : ' + len(self.passwd) * '*'+ '\n' ) else : self.url = '' print('正在使用手机登录...') form = {'_xsrf' : self.get_xsrf(), 'password' : self.passwd, 'email' : self.client_info, 'captcha': captcha } print(form) try: req = request.Request(self.url, parse.urlencode(form).encode('utf-8')) f = self.opener.open(req) self.cj.save() print(json.loads(f.read().decode('utf-8'))["msg"] + "!") print("=" * 100) except: print('Error!') def get_capthca(self) : ''' Interface for getting the captcha ''' pass def get_target_page(self): ''' Get main target_page content after logged in ''' try: self.cj.load() print('Cookie loaded....') self.target_page = self.opener.open('https://www.zhihu.com/people/edit') f = open('zhihu.html', 'wb') f.write(target_page_content.read()) except: self.login() self.get_target_page() def isLogged(self, user_client): ''' test if Logged ''' f = user_client.open('https://www.zhihu.com/settings/profile').geturl() if (f != 'https://www.zhihu.com/settings/profile'): return False return True def user_getter(sefl): return User_Getter('https://www.zhihu.com/people/xiao-guai-shou-2/activities').urls() def profile_collector(self, text_path=None) : ''' main entry for collecting user's profile including id, gender, education, career ''' count = 0 self.cj.load() user_list = [] # check the source of the data if (text_path != None) : with open(text_path, 'r') as source_list : for line in source_list : user_list.append(line.split('\n')[0]) source_list.close() else : user_list = [] initial_time = time.time() while(len(user_list) > 0): # this try except block is for resuming from the server' shutdown try : for item in user_list : start_time = time.time() user_id = item print('=Writing information of [', user_id,']...') url = 'https://www.zhihu.com/api/v4/members/' + user_id + '?include=locations%2Cemployments%2Cgender%2Ceducations%2Cbusiness%2Cvoteup_count%2Cthanked_Count%2Cfollower_count%2Cfollowing_count%2Ccover_url%2Cfollowing_topic_count%2Cfollowing_question_count%2Cfollowing_favlists_count%2Cfollowing_columns_count%2Canswer_count%2Carticles_count%2Cpins_count%2Cquestion_count%2Ccommercial_question_count%2Cfavorite_count%2Cfavorited_count%2Clogs_count%2Cmarked_answers_count%2Cmarked_answers_text%2Cmessage_thread_token%2Caccount_status%2Cis_active%2Cis_force_renamed%2Cis_bind_sina%2Csina_weibo_url%2Csina_weibo_name%2Cshow_sina_weibo%2Cis_blocking%2Cis_blocked%2Cis_following%2Cis_followed%2Cmutual_followees_count%2Cvote_to_count%2Cvote_from_count%2Cthank_to_count%2Cthank_from_count%2Cthanked_count%2Cdescription%2Chosted_live_count%2Cparticipated_live_count%2Callow_message%2Cindustry_category%2Corg_name%2Corg_homepage%2Cbadge%5B%3F(type%3Dbest_answerer)%5D.topics' req = request.Request(url) raw_data = self.opener.open(req).read() json_data = json.loads(raw_data) # get key and value pic_url = json_data["avatar_url"].split('_')[0] + '_xll.jpg' number_id = json_data["id"] user_name = json_data["name"] # education if ("educations" in json_data) : if (len(json_data["educations"]) != 0) : if ("school" in json_data["educations"][0]) : university = json_data["educations"][0]["school"]["name"] else : university = 'None' if ("major" in json_data["educations"][0]) : major = json_data["educations"][0]["major"]["name"] else: major = 'None' else : university = 'None' major = 'None' else : university = 'None' major = 'None' # employments if ("employments" in json_data) : if (len(json_data["employments"]) != 0) : if ("company" in json_data["employments"][0]) : company = json_data["employments"][0]["company"]["name"] else : company = 'None' if ("occupation" in json_data["employments"][0]) : occupation = json_data["employments"][0]["job"]["name"] else : occupation = 'None' else : company = 'None' occupation = 'None' else : company = 'None' occupation = 'None' # location if ("locations" in json_data) : if (len(json_data["locations"]) != 0) : location = json_data["locations"][0]["name"] else : location = 'None' else : location = 'None' # business if ("business" in json_data ) : industry = json_data["business"]["name"] else : industry = 'None' intro = json_data["headline"] autobiography = json_data["description"] user_type = json_data["type"] follower_count = json_data["follower_count"] following_count = json_data["following_count"] answers_count = json_data["answer_count"] articles_count = json_data["articles_count"] if (json_data["gender"] == 1) : gender = 'male' else : gender = 'female' data = { 'id' : number_id, 'user_id' : user_id, 'name' : user_name, 'gender' : gender, 'university' : university, 'major' : major, 'industry' : industry, 'company' : company, 'occupation' : occupation, 'location' : location, 'intro' : intro, 'autobiography' : autobiography, 'user_type' : str(user_type), 'follower_count' : str(follower_count), 'following_count' : str(following_count), 'answer-count' : str(answers_count), 'articles_count' : str(articles_count) } # process folder if not (os.path.exists(os.path.join('./data/' ,user_name))
www.zhihu.com' + '/captcha.gif?r=' + str(int(time.time())) + '&type=login' f = request.urlopen(url) with open('./cap.png', 'wb') as image: image.write(f.read()) image.close() def login(sel
identifier_body
Zhihu__Crawler_JsonVersion.py
# # # Created on : 03/07/17 # Last Modified : # # Author : Pengcheng Zhou(Kevin) # #################################################### import re from urllib import parse, request, error from multiprocessing import Pool import http.cookiejar as Cookie import time import json from getpass import getpass import ssl # Cancel the certification of target site ssl._create_default_https_context = ssl._create_unverified_context import os from User_Getter import User_Getter import random cookieFile = 'zhihu_cookie.txt' class RedirectHandler(request.HTTPRedirectHandler): def http_error_302(self, req, fp, code, msg, headers): print ("Cookie过期,重新登录中....") return http_error_301 = http_error_302 class Zhihu(object): def __init__(self) : ''' Initialize ''' self.pool = Pool(4) self.cj = Cookie.MozillaCookieJar(cookieFile) self.opener = request.build_opener(request.HTTPCookieProcessor(self.cj), RedirectHandler()) self.client_info = 'monsterzpc@gmail.com' self.passwd = 'Zpc920515' self.url = 'https://www.zhihu.com/login/email' self.target_page = '' self.position = 0 print(''' ############################################################ # # # Zhihu Auto_Login and Crawler by Pengcheng Zhou. # # # ############################################################ ''') def get_xsrf(self) : ''' Get a special dynamic string for login ''' login_target_page = request.urlopen(self.url) pattern = re.compile('<input type="hidden" name="_xsrf" value="(.*)"/>') _xsrf = re.findall(pattern, login_target_page.read().decode('utf-8'))[0] return _xsrf def get_captcha_url(self): url = 'https://www.zhihu.com' + '/captcha.gif?r=' + str(int(time.time())) + '&type=login' f = request.urlopen(url) with open('./cap.png', 'wb') as image: image.write(f.read()) image.close() def login(self): ''' Execution of login ''' if (self.client_info == '' or self.passwd == '') : self.client_info = input('请输入账号:') self.passwd = getpass('请输入密码:') self.get_captcha_url() captcha = input('请输入验证码:') if (self.client_info.find("@") != -1) : print('''正在使用邮箱登录...\n用户名:''' + self.client_info+ '\n' + '密码 : ' + len(self.passwd) * '*'+ '\n' ) else : self.url = '' print('正在使用手机登录...') form = {'_xsrf' : self.get_xsrf(), 'password' : self.passwd, 'email' : self.client_info, 'captcha': captcha } print(form) try: req = request.Request(self.url, parse.urlencode(form).encode('utf-8')) f = self.opener.open(req) self.cj.save() print(json.loads(f.read().decode('utf-8'))["msg"] + "!") print("=" * 100) except: print('Error!') def get_capthca(self) : ''' Interface for getting the captcha ''' pass def get_target_page(self): ''' Get main target_page content after logged in ''' try: self.cj.load() print('Cookie loaded....') self.target_page = self.opener.open('https://www.zhihu.com/people/edit') f = open('zhihu.html', 'wb') f.write(target_page_content.read()) except: self.login() self.get_target_page() def isLogged(self, user_client): ''' test if Logged ''' f = user_client.open('https://www.zhihu.com/settings/profile').geturl() if (f != 'https://www.zhihu.com/settings/profile'): return False return True def user_getter(sefl): return User_Getter('https://www.zhihu.com/people/xiao-guai-shou-2/activities').urls() def profile_collector(self, text_path=None) : ''' main entry for collecting user's profile including id, gender, education, career ''' count = 0 self.cj.load() user_list = [] # check the source of the data if (text_path != None) : with open(text_path, 'r') as source_list : for line in source_list : user_list.append(line.split('\n')[0]) source_list.close() else : user_list = [] initial_time = time.time() while(len(user_list) > 0): # this try except block is for resuming from the server' shutdown try : for item in user_list : start_time = time.time() user_id = item print('=Writing information of [', user_id,']...') url = 'https://www.zhihu.com/api/v4/members/' + user_id + '?include=locations%2Cemployments%2Cgender%2Ceducations%2Cbusiness%2Cvoteup_count%2Cthanked_Count%2Cfollower_count%2Cfollowing_count%2Ccover_url%2Cfollowing_topic_count%2Cfollowing_question_count%2Cfollowing_favlists_count%2Cfollowing_columns_count%2Canswer_count%2Carticles_count%2Cpins_count%2Cquestion_count%2Ccommercial_question_count%2Cfavorite_count%2Cfavorited_count%2Clogs_count%2Cmarked_answers_count%2Cmarked_answers_text%2Cmessage_thread_token%2Caccount_status%2Cis_active%2Cis_force_renamed%2Cis_bind_sina%2Csina_weibo_url%2Csina_weibo_name%2Cshow_sina_weibo%2Cis_blocking%2Cis_blocked%2Cis_following%2Cis_followed%2Cmutual_followees_count%2Cvote_to_count%2Cvote_from_count%2Cthank_to_count%2Cthank_from_count%2Cthanked_count%2Cdescription%2Chosted_live_count%2Cparticipated_live_count%2Callow_message%2Cindustry_category%2Corg_name%2Corg_homepage%2Cbadge%5B%3F(type%3Dbest_answerer)%5D.topics' req = request.Request(url) raw_data = self.opener.open(req).read() json_data = json.loads(raw_data) # get key and value pic_url = json_data["avatar_url"].split('_')[0] + '_xll.jpg' number_id = json_data["id"] user_name = json_data["name"] # education if ("educations" in json_data) : if (len(json_data["educations"]) != 0) : if ("school" in json_data["educations"][0]) : university = json_data["educations"][0]["school"]["name"] else : university = 'None' if ("major" in json_data["educations"][0]) : major = json_data["educations"][0]["major"]["name"] else: major = 'None' else : university = 'None' major = 'None' else : university = 'None' major = 'None' # employments if ("employments" in json_data) : if (len(json_data["employments"]) != 0) : if ("company" in json_data["employments"][0]) : company = json_data["employments"][0]["company"]["name"] else : company = 'None' if ("occupation" in json_data["employments"][0]) : occupation = json_data["employments"][0]["job"]["name"] else : occupation = 'None' else : company = 'None' occupation = 'None' else : company = 'None' occupation = 'None' # location if ("locations" in json_data) : if (len(json_data["locations"]) != 0) : location = json_data["locations"][0]["name"] else : location = 'None' else : location = 'None' # business if ("business" in json_data ) : industry = json_data["business"]["name"] else : industry = 'None' intro = json_data["headline"] autobiography = json_data["description"] user_type = json_data["type"] follower_count = json_data["follower_count"] following_count = json_data
#!/usr/bin/python # --*-- encoding:utf-8 --*-- #################################################### # Zhihu Auto-Aogin
random_line_split
main.rs
Cell<Option<TypeEncoder>>> = Mutex::new(RefCell::new(None)); type TypeMotor = Motor; static G_MOTOR: Mutex<RefCell<Option<TypeMotor>>> = Mutex::new(RefCell::new(None)); static G_TIM: Mutex<RefCell<Option<Timer<TIM2>>>> = Mutex::new(RefCell::new(None)); static mut CURRENT_A: f32 = 0.0; static mut CURRENT_B: f32 = 0.0; static mut CURRENT_C: f32 = 0.0; static mut MOT_ANGLE: u16 = 0; static mut MOT_ANGLE_OLD: u16 = 0; static mut MOT_VELOCITY: f32 = 0.0; static mut MOT_VELOCITY_OLD: f32 = 0.0; static mut ERR_VELOCITY: f32 = 0.0; static mut ERR_VELOCITY_INT: f32 = 0.0; static mut REF_CURR_D: f32 = 0.0; static mut REF_CURR_Q: f32 = 0.0; // System const TIM2_FREQ_KHZ: u32 = 10; // Motor const MOT_POLE_PAIRS: u16 = 12; // Encoder const ENC_RESOLUTION: u16 = 16384; #[interrupt] fn ADC() { // current sensing unsafe { let max_sample:u32 = (1 << 12) - 1; let device = pac::Peripherals::steal(); device.ADC1.sr.modify(|_, w| w.jeoc().clear_bit()); let jdr1_data = device.ADC1.jdr1.read().jdata().bits(); let jdr1_offset = 48u32; let so1 = ( ( (u32::from(jdr1_data) + jdr1_offset) * VDDA_CALIB ) / max_sample) as u16; let jdr2_data = device.ADC1.jdr2.read().jdata().bits(); let jdr2_offset = 118u32; let so2 = ( ( (u32::from(jdr2_data) + jdr2_offset) * VDDA_CALIB ) / max_sample) as u16; CURRENT_B = (so1 as f32 - 1650.0) / 200.0; CURRENT_C = (so2 as f32 - 1650.0) / 200.0; CURRENT_A = - CURRENT_B - CURRENT_C; } /* // LED Debug cortex_m::interrupt::free(|cs| { if let Some(ref mut led) = G_LED.borrow(cs).borrow_mut().as_mut() { led.toggle().unwrap(); } }); */ } #[interrupt] fn TIM2() { cortex_m::interrupt::free(|cs| { if let Some(ref mut tim) = G_TIM.borrow(cs).borrow_mut().as_mut()
}); static mut SPI3: Option<TypeSpi3> = None; static mut MOTOR: Option<TypeMotor> = None; unsafe{ let mut spi3 = SPI3.get_or_insert_with(|| { cortex_m::interrupt::free(|cs| { G_SPI3.borrow(cs).replace(None).unwrap() }) }); let device = pac::Peripherals::steal(); let gpioa = device.GPIOA.split(); let ncs = gpioa.pa3.into_push_pull_output(); let mut as5048: TypeEncoder = AS5048A::new(&mut spi3, ncs); // AS5048A let measured_angle = as5048.angle().unwrap(); let angle_offset = 650u16; MOT_ANGLE = (measured_angle - angle_offset) % ENC_RESOLUTION; let electric_angle = MOT_ANGLE % (ENC_RESOLUTION/MOT_POLE_PAIRS); let motor = MOTOR.get_or_insert_with(|| { cortex_m::interrupt::free(|cs| { G_MOTOR.borrow(cs).replace(None).unwrap() }) }); // Velocity control const REF_VELOCITY: f32 = - 100.0; const VELLOCITY_PGAIN: f32 = 0.1; const VELOCITY_IGAIN: f32 = 0.00001; let res_velocity = if (ENC_RESOLUTION-1000) < MOT_ANGLE_OLD && MOT_ANGLE < 1000 { ( MOT_ANGLE as f32 - MOT_ANGLE_OLD as f32 + ENC_RESOLUTION as f32 ) * TIM2_FREQ_KHZ as f32 } else if MOT_ANGLE_OLD < 1000 && MOT_ANGLE > (ENC_RESOLUTION-1000) { ( MOT_ANGLE as f32 - MOT_ANGLE_OLD as f32 - ENC_RESOLUTION as f32 ) * TIM2_FREQ_KHZ as f32 } else { ( MOT_ANGLE as f32 - MOT_ANGLE_OLD as f32 ) * TIM2_FREQ_KHZ as f32 }; let alpha = 0.1; MOT_VELOCITY = alpha * res_velocity + (1.0 - alpha) * MOT_VELOCITY_OLD; ERR_VELOCITY = MOT_VELOCITY - REF_VELOCITY; ERR_VELOCITY_INT += ERR_VELOCITY; REF_CURR_D = 0.0; REF_CURR_Q = VELLOCITY_PGAIN * ERR_VELOCITY + VELOCITY_IGAIN * ERR_VELOCITY_INT; REF_CURR_Q = -1.0 * REF_CURR_Q; MOT_ANGLE_OLD = MOT_ANGLE; MOT_VELOCITY_OLD = MOT_VELOCITY; // select control mode //motor.drive_profile().unwrap(); //motor.drive_sixstep().unwrap(); //motor.drive_anglebased_sixstep(electric_angle).unwrap(); motor.drive_foc(electric_angle, CURRENT_A, CURRENT_B, CURRENT_C, REF_CURR_D, REF_CURR_Q).unwrap(); } /* // LED Debug cortex_m::interrupt::free(|cs| { if let Some(ref mut led) = G_LED.borrow(cs).borrow_mut().as_mut() { led.toggle().unwrap(); } }); */ } #[entry] fn main() -> ! { let dp = pac::Peripherals::take().unwrap(); let cp = cortex_m::peripheral::Peripherals::take().unwrap(); let rcc = dp.RCC.constrain(); let clocks = rcc .cfgr .use_hse(8.mhz()) .sysclk(168.mhz()) .hclk(168.mhz()) .pclk1(42.mhz()) .pclk2(84.mhz()) .require_pll48clk() .freeze(); let mut delay = hal::delay::Delay::new(cp.SYST, clocks); let gpioa = dp.GPIOA.split(); let gpiob = dp.GPIOB.split(); let gpioc = dp.GPIOC.split(); // SPI3 let sck = gpioc.pc10.into_alternate_af6(); let miso = gpioc.pc11.into_alternate_af6(); let mosi = gpioc.pc12.into_alternate_af6(); let mut spi = Spi::spi3( dp.SPI3, (sck, miso, mosi), MODE_1, KiloHertz(2000).into(), clocks, ); // DRV8301 let ncs = gpioc.pc13.into_push_pull_output(); let en_gate = gpiob.pb12.into_push_pull_output(); let mut drv8301 = Drv8301::new(&mut spi, ncs, en_gate); drv8301.init().unwrap(); // Move the pin into our global storage cortex_m::interrupt::free(|cs| *G_SPI3.borrow(cs).borrow_mut() = Some(spi)); // PWM let channels = (gpioa.pa8.into_alternate_af1(), gpioa.pa9.into_alternate_af1(), gpioa.pa10.into_alternate_af1(), gpioa.pa11.into_alternate_af1()); let pwm = pwm::tim1(dp.TIM1, channels, clocks, 16u32.khz()); let (ch1, ch2, ch3, ch4) = pwm; let mut ch4 = ch4; { // Set complementary oututs mode as AF1 gpiob.pb13.into_alternate_af1(); gpiob.pb14.into_alternate_af1(); gpiob.pb15.into_alternate_af1(); unsafe { let tim1_regb = &(*(TIM1::ptr())); // Enable complementary outputs tim1_regb.ccer.modify(|_, w| w.cc1ne().set_bit()); tim1_regb.ccer.modify(|_, w| w.cc2ne().set_bit()); tim1_regb.ccer.modify(|_, w| w.cc3ne().set_bit()); // Set dead time tim1_regb.bdtr.modify(|_,
{ let _ = tim.wait(); }
conditional_block
main.rs
Cell<Option<TypeEncoder>>> = Mutex::new(RefCell::new(None)); type TypeMotor = Motor; static G_MOTOR: Mutex<RefCell<Option<TypeMotor>>> = Mutex::new(RefCell::new(None)); static G_TIM: Mutex<RefCell<Option<Timer<TIM2>>>> = Mutex::new(RefCell::new(None)); static mut CURRENT_A: f32 = 0.0; static mut CURRENT_B: f32 = 0.0; static mut CURRENT_C: f32 = 0.0; static mut MOT_ANGLE: u16 = 0; static mut MOT_ANGLE_OLD: u16 = 0; static mut MOT_VELOCITY: f32 = 0.0; static mut MOT_VELOCITY_OLD: f32 = 0.0; static mut ERR_VELOCITY: f32 = 0.0; static mut ERR_VELOCITY_INT: f32 = 0.0; static mut REF_CURR_D: f32 = 0.0; static mut REF_CURR_Q: f32 = 0.0; // System const TIM2_FREQ_KHZ: u32 = 10; // Motor const MOT_POLE_PAIRS: u16 = 12; // Encoder const ENC_RESOLUTION: u16 = 16384; #[interrupt] fn ADC()
/* // LED Debug cortex_m::interrupt::free(|cs| { if let Some(ref mut led) = G_LED.borrow(cs).borrow_mut().as_mut() { led.toggle().unwrap(); } }); */ } #[interrupt] fn TIM2() { cortex_m::interrupt::free(|cs| { if let Some(ref mut tim) = G_TIM.borrow(cs).borrow_mut().as_mut() { let _ = tim.wait(); } }); static mut SPI3: Option<TypeSpi3> = None; static mut MOTOR: Option<TypeMotor> = None; unsafe{ let mut spi3 = SPI3.get_or_insert_with(|| { cortex_m::interrupt::free(|cs| { G_SPI3.borrow(cs).replace(None).unwrap() }) }); let device = pac::Peripherals::steal(); let gpioa = device.GPIOA.split(); let ncs = gpioa.pa3.into_push_pull_output(); let mut as5048: TypeEncoder = AS5048A::new(&mut spi3, ncs); // AS5048A let measured_angle = as5048.angle().unwrap(); let angle_offset = 650u16; MOT_ANGLE = (measured_angle - angle_offset) % ENC_RESOLUTION; let electric_angle = MOT_ANGLE % (ENC_RESOLUTION/MOT_POLE_PAIRS); let motor = MOTOR.get_or_insert_with(|| { cortex_m::interrupt::free(|cs| { G_MOTOR.borrow(cs).replace(None).unwrap() }) }); // Velocity control const REF_VELOCITY: f32 = - 100.0; const VELLOCITY_PGAIN: f32 = 0.1; const VELOCITY_IGAIN: f32 = 0.00001; let res_velocity = if (ENC_RESOLUTION-1000) < MOT_ANGLE_OLD && MOT_ANGLE < 1000 { ( MOT_ANGLE as f32 - MOT_ANGLE_OLD as f32 + ENC_RESOLUTION as f32 ) * TIM2_FREQ_KHZ as f32 } else if MOT_ANGLE_OLD < 1000 && MOT_ANGLE > (ENC_RESOLUTION-1000) { ( MOT_ANGLE as f32 - MOT_ANGLE_OLD as f32 - ENC_RESOLUTION as f32 ) * TIM2_FREQ_KHZ as f32 } else { ( MOT_ANGLE as f32 - MOT_ANGLE_OLD as f32 ) * TIM2_FREQ_KHZ as f32 }; let alpha = 0.1; MOT_VELOCITY = alpha * res_velocity + (1.0 - alpha) * MOT_VELOCITY_OLD; ERR_VELOCITY = MOT_VELOCITY - REF_VELOCITY; ERR_VELOCITY_INT += ERR_VELOCITY; REF_CURR_D = 0.0; REF_CURR_Q = VELLOCITY_PGAIN * ERR_VELOCITY + VELOCITY_IGAIN * ERR_VELOCITY_INT; REF_CURR_Q = -1.0 * REF_CURR_Q; MOT_ANGLE_OLD = MOT_ANGLE; MOT_VELOCITY_OLD = MOT_VELOCITY; // select control mode //motor.drive_profile().unwrap(); //motor.drive_sixstep().unwrap(); //motor.drive_anglebased_sixstep(electric_angle).unwrap(); motor.drive_foc(electric_angle, CURRENT_A, CURRENT_B, CURRENT_C, REF_CURR_D, REF_CURR_Q).unwrap(); } /* // LED Debug cortex_m::interrupt::free(|cs| { if let Some(ref mut led) = G_LED.borrow(cs).borrow_mut().as_mut() { led.toggle().unwrap(); } }); */ } #[entry] fn main() -> ! { let dp = pac::Peripherals::take().unwrap(); let cp = cortex_m::peripheral::Peripherals::take().unwrap(); let rcc = dp.RCC.constrain(); let clocks = rcc .cfgr .use_hse(8.mhz()) .sysclk(168.mhz()) .hclk(168.mhz()) .pclk1(42.mhz()) .pclk2(84.mhz()) .require_pll48clk() .freeze(); let mut delay = hal::delay::Delay::new(cp.SYST, clocks); let gpioa = dp.GPIOA.split(); let gpiob = dp.GPIOB.split(); let gpioc = dp.GPIOC.split(); // SPI3 let sck = gpioc.pc10.into_alternate_af6(); let miso = gpioc.pc11.into_alternate_af6(); let mosi = gpioc.pc12.into_alternate_af6(); let mut spi = Spi::spi3( dp.SPI3, (sck, miso, mosi), MODE_1, KiloHertz(2000).into(), clocks, ); // DRV8301 let ncs = gpioc.pc13.into_push_pull_output(); let en_gate = gpiob.pb12.into_push_pull_output(); let mut drv8301 = Drv8301::new(&mut spi, ncs, en_gate); drv8301.init().unwrap(); // Move the pin into our global storage cortex_m::interrupt::free(|cs| *G_SPI3.borrow(cs).borrow_mut() = Some(spi)); // PWM let channels = (gpioa.pa8.into_alternate_af1(), gpioa.pa9.into_alternate_af1(), gpioa.pa10.into_alternate_af1(), gpioa.pa11.into_alternate_af1()); let pwm = pwm::tim1(dp.TIM1, channels, clocks, 16u32.khz()); let (ch1, ch2, ch3, ch4) = pwm; let mut ch4 = ch4; { // Set complementary oututs mode as AF1 gpiob.pb13.into_alternate_af1(); gpiob.pb14.into_alternate_af1(); gpiob.pb15.into_alternate_af1(); unsafe { let tim1_regb = &(*(TIM1::ptr())); // Enable complementary outputs tim1_regb.ccer.modify(|_, w| w.cc1ne().set_bit()); tim1_regb.ccer.modify(|_, w| w.cc2ne().set_bit()); tim1_regb.ccer.modify(|_, w| w.cc3ne().set_bit()); // Set dead time tim1_regb.bdtr.modify(|_,
{ // current sensing unsafe { let max_sample:u32 = (1 << 12) - 1; let device = pac::Peripherals::steal(); device.ADC1.sr.modify(|_, w| w.jeoc().clear_bit()); let jdr1_data = device.ADC1.jdr1.read().jdata().bits(); let jdr1_offset = 48u32; let so1 = ( ( (u32::from(jdr1_data) + jdr1_offset) * VDDA_CALIB ) / max_sample) as u16; let jdr2_data = device.ADC1.jdr2.read().jdata().bits(); let jdr2_offset = 118u32; let so2 = ( ( (u32::from(jdr2_data) + jdr2_offset) * VDDA_CALIB ) / max_sample) as u16; CURRENT_B = (so1 as f32 - 1650.0) / 200.0; CURRENT_C = (so2 as f32 - 1650.0) / 200.0; CURRENT_A = - CURRENT_B - CURRENT_C; }
identifier_body
main.rs
Cell<Option<TypeEncoder>>> = Mutex::new(RefCell::new(None)); type TypeMotor = Motor; static G_MOTOR: Mutex<RefCell<Option<TypeMotor>>> = Mutex::new(RefCell::new(None)); static G_TIM: Mutex<RefCell<Option<Timer<TIM2>>>> = Mutex::new(RefCell::new(None)); static mut CURRENT_A: f32 = 0.0; static mut CURRENT_B: f32 = 0.0; static mut CURRENT_C: f32 = 0.0; static mut MOT_ANGLE: u16 = 0; static mut MOT_ANGLE_OLD: u16 = 0; static mut MOT_VELOCITY: f32 = 0.0; static mut MOT_VELOCITY_OLD: f32 = 0.0; static mut ERR_VELOCITY: f32 = 0.0; static mut ERR_VELOCITY_INT: f32 = 0.0; static mut REF_CURR_D: f32 = 0.0; static mut REF_CURR_Q: f32 = 0.0; // System const TIM2_FREQ_KHZ: u32 = 10; // Motor const MOT_POLE_PAIRS: u16 = 12; // Encoder const ENC_RESOLUTION: u16 = 16384; #[interrupt] fn
() { // current sensing unsafe { let max_sample:u32 = (1 << 12) - 1; let device = pac::Peripherals::steal(); device.ADC1.sr.modify(|_, w| w.jeoc().clear_bit()); let jdr1_data = device.ADC1.jdr1.read().jdata().bits(); let jdr1_offset = 48u32; let so1 = ( ( (u32::from(jdr1_data) + jdr1_offset) * VDDA_CALIB ) / max_sample) as u16; let jdr2_data = device.ADC1.jdr2.read().jdata().bits(); let jdr2_offset = 118u32; let so2 = ( ( (u32::from(jdr2_data) + jdr2_offset) * VDDA_CALIB ) / max_sample) as u16; CURRENT_B = (so1 as f32 - 1650.0) / 200.0; CURRENT_C = (so2 as f32 - 1650.0) / 200.0; CURRENT_A = - CURRENT_B - CURRENT_C; } /* // LED Debug cortex_m::interrupt::free(|cs| { if let Some(ref mut led) = G_LED.borrow(cs).borrow_mut().as_mut() { led.toggle().unwrap(); } }); */ } #[interrupt] fn TIM2() { cortex_m::interrupt::free(|cs| { if let Some(ref mut tim) = G_TIM.borrow(cs).borrow_mut().as_mut() { let _ = tim.wait(); } }); static mut SPI3: Option<TypeSpi3> = None; static mut MOTOR: Option<TypeMotor> = None; unsafe{ let mut spi3 = SPI3.get_or_insert_with(|| { cortex_m::interrupt::free(|cs| { G_SPI3.borrow(cs).replace(None).unwrap() }) }); let device = pac::Peripherals::steal(); let gpioa = device.GPIOA.split(); let ncs = gpioa.pa3.into_push_pull_output(); let mut as5048: TypeEncoder = AS5048A::new(&mut spi3, ncs); // AS5048A let measured_angle = as5048.angle().unwrap(); let angle_offset = 650u16; MOT_ANGLE = (measured_angle - angle_offset) % ENC_RESOLUTION; let electric_angle = MOT_ANGLE % (ENC_RESOLUTION/MOT_POLE_PAIRS); let motor = MOTOR.get_or_insert_with(|| { cortex_m::interrupt::free(|cs| { G_MOTOR.borrow(cs).replace(None).unwrap() }) }); // Velocity control const REF_VELOCITY: f32 = - 100.0; const VELLOCITY_PGAIN: f32 = 0.1; const VELOCITY_IGAIN: f32 = 0.00001; let res_velocity = if (ENC_RESOLUTION-1000) < MOT_ANGLE_OLD && MOT_ANGLE < 1000 { ( MOT_ANGLE as f32 - MOT_ANGLE_OLD as f32 + ENC_RESOLUTION as f32 ) * TIM2_FREQ_KHZ as f32 } else if MOT_ANGLE_OLD < 1000 && MOT_ANGLE > (ENC_RESOLUTION-1000) { ( MOT_ANGLE as f32 - MOT_ANGLE_OLD as f32 - ENC_RESOLUTION as f32 ) * TIM2_FREQ_KHZ as f32 } else { ( MOT_ANGLE as f32 - MOT_ANGLE_OLD as f32 ) * TIM2_FREQ_KHZ as f32 }; let alpha = 0.1; MOT_VELOCITY = alpha * res_velocity + (1.0 - alpha) * MOT_VELOCITY_OLD; ERR_VELOCITY = MOT_VELOCITY - REF_VELOCITY; ERR_VELOCITY_INT += ERR_VELOCITY; REF_CURR_D = 0.0; REF_CURR_Q = VELLOCITY_PGAIN * ERR_VELOCITY + VELOCITY_IGAIN * ERR_VELOCITY_INT; REF_CURR_Q = -1.0 * REF_CURR_Q; MOT_ANGLE_OLD = MOT_ANGLE; MOT_VELOCITY_OLD = MOT_VELOCITY; // select control mode //motor.drive_profile().unwrap(); //motor.drive_sixstep().unwrap(); //motor.drive_anglebased_sixstep(electric_angle).unwrap(); motor.drive_foc(electric_angle, CURRENT_A, CURRENT_B, CURRENT_C, REF_CURR_D, REF_CURR_Q).unwrap(); } /* // LED Debug cortex_m::interrupt::free(|cs| { if let Some(ref mut led) = G_LED.borrow(cs).borrow_mut().as_mut() { led.toggle().unwrap(); } }); */ } #[entry] fn main() -> ! { let dp = pac::Peripherals::take().unwrap(); let cp = cortex_m::peripheral::Peripherals::take().unwrap(); let rcc = dp.RCC.constrain(); let clocks = rcc .cfgr .use_hse(8.mhz()) .sysclk(168.mhz()) .hclk(168.mhz()) .pclk1(42.mhz()) .pclk2(84.mhz()) .require_pll48clk() .freeze(); let mut delay = hal::delay::Delay::new(cp.SYST, clocks); let gpioa = dp.GPIOA.split(); let gpiob = dp.GPIOB.split(); let gpioc = dp.GPIOC.split(); // SPI3 let sck = gpioc.pc10.into_alternate_af6(); let miso = gpioc.pc11.into_alternate_af6(); let mosi = gpioc.pc12.into_alternate_af6(); let mut spi = Spi::spi3( dp.SPI3, (sck, miso, mosi), MODE_1, KiloHertz(2000).into(), clocks, ); // DRV8301 let ncs = gpioc.pc13.into_push_pull_output(); let en_gate = gpiob.pb12.into_push_pull_output(); let mut drv8301 = Drv8301::new(&mut spi, ncs, en_gate); drv8301.init().unwrap(); // Move the pin into our global storage cortex_m::interrupt::free(|cs| *G_SPI3.borrow(cs).borrow_mut() = Some(spi)); // PWM let channels = (gpioa.pa8.into_alternate_af1(), gpioa.pa9.into_alternate_af1(), gpioa.pa10.into_alternate_af1(), gpioa.pa11.into_alternate_af1()); let pwm = pwm::tim1(dp.TIM1, channels, clocks, 16u32.khz()); let (ch1, ch2, ch3, ch4) = pwm; let mut ch4 = ch4; { // Set complementary oututs mode as AF1 gpiob.pb13.into_alternate_af1(); gpiob.pb14.into_alternate_af1(); gpiob.pb15.into_alternate_af1(); unsafe { let tim1_regb = &(*(TIM1::ptr())); // Enable complementary outputs tim1_regb.ccer.modify(|_, w| w.cc1ne().set_bit()); tim1_regb.ccer.modify(|_, w| w.cc2ne().set_bit()); tim1_regb.ccer.modify(|_, w| w.cc3ne().set_bit()); // Set dead time tim1_regb.bdtr.modify(|_, w
ADC
identifier_name
main.rs
static G_MOTOR: Mutex<RefCell<Option<TypeMotor>>> = Mutex::new(RefCell::new(None)); static G_TIM: Mutex<RefCell<Option<Timer<TIM2>>>> = Mutex::new(RefCell::new(None)); static mut CURRENT_A: f32 = 0.0; static mut CURRENT_B: f32 = 0.0; static mut CURRENT_C: f32 = 0.0; static mut MOT_ANGLE: u16 = 0; static mut MOT_ANGLE_OLD: u16 = 0; static mut MOT_VELOCITY: f32 = 0.0; static mut MOT_VELOCITY_OLD: f32 = 0.0; static mut ERR_VELOCITY: f32 = 0.0; static mut ERR_VELOCITY_INT: f32 = 0.0; static mut REF_CURR_D: f32 = 0.0; static mut REF_CURR_Q: f32 = 0.0; // System const TIM2_FREQ_KHZ: u32 = 10; // Motor const MOT_POLE_PAIRS: u16 = 12; // Encoder const ENC_RESOLUTION: u16 = 16384; #[interrupt] fn ADC() { // current sensing unsafe { let max_sample:u32 = (1 << 12) - 1; let device = pac::Peripherals::steal(); device.ADC1.sr.modify(|_, w| w.jeoc().clear_bit()); let jdr1_data = device.ADC1.jdr1.read().jdata().bits(); let jdr1_offset = 48u32; let so1 = ( ( (u32::from(jdr1_data) + jdr1_offset) * VDDA_CALIB ) / max_sample) as u16; let jdr2_data = device.ADC1.jdr2.read().jdata().bits(); let jdr2_offset = 118u32; let so2 = ( ( (u32::from(jdr2_data) + jdr2_offset) * VDDA_CALIB ) / max_sample) as u16; CURRENT_B = (so1 as f32 - 1650.0) / 200.0; CURRENT_C = (so2 as f32 - 1650.0) / 200.0; CURRENT_A = - CURRENT_B - CURRENT_C; } /* // LED Debug cortex_m::interrupt::free(|cs| { if let Some(ref mut led) = G_LED.borrow(cs).borrow_mut().as_mut() { led.toggle().unwrap(); } }); */ } #[interrupt] fn TIM2() { cortex_m::interrupt::free(|cs| { if let Some(ref mut tim) = G_TIM.borrow(cs).borrow_mut().as_mut() { let _ = tim.wait(); } }); static mut SPI3: Option<TypeSpi3> = None; static mut MOTOR: Option<TypeMotor> = None; unsafe{ let mut spi3 = SPI3.get_or_insert_with(|| { cortex_m::interrupt::free(|cs| { G_SPI3.borrow(cs).replace(None).unwrap() }) }); let device = pac::Peripherals::steal(); let gpioa = device.GPIOA.split(); let ncs = gpioa.pa3.into_push_pull_output(); let mut as5048: TypeEncoder = AS5048A::new(&mut spi3, ncs); // AS5048A let measured_angle = as5048.angle().unwrap(); let angle_offset = 650u16; MOT_ANGLE = (measured_angle - angle_offset) % ENC_RESOLUTION; let electric_angle = MOT_ANGLE % (ENC_RESOLUTION/MOT_POLE_PAIRS); let motor = MOTOR.get_or_insert_with(|| { cortex_m::interrupt::free(|cs| { G_MOTOR.borrow(cs).replace(None).unwrap() }) }); // Velocity control const REF_VELOCITY: f32 = - 100.0; const VELLOCITY_PGAIN: f32 = 0.1; const VELOCITY_IGAIN: f32 = 0.00001; let res_velocity = if (ENC_RESOLUTION-1000) < MOT_ANGLE_OLD && MOT_ANGLE < 1000 { ( MOT_ANGLE as f32 - MOT_ANGLE_OLD as f32 + ENC_RESOLUTION as f32 ) * TIM2_FREQ_KHZ as f32 } else if MOT_ANGLE_OLD < 1000 && MOT_ANGLE > (ENC_RESOLUTION-1000) { ( MOT_ANGLE as f32 - MOT_ANGLE_OLD as f32 - ENC_RESOLUTION as f32 ) * TIM2_FREQ_KHZ as f32 } else { ( MOT_ANGLE as f32 - MOT_ANGLE_OLD as f32 ) * TIM2_FREQ_KHZ as f32 }; let alpha = 0.1; MOT_VELOCITY = alpha * res_velocity + (1.0 - alpha) * MOT_VELOCITY_OLD; ERR_VELOCITY = MOT_VELOCITY - REF_VELOCITY; ERR_VELOCITY_INT += ERR_VELOCITY; REF_CURR_D = 0.0; REF_CURR_Q = VELLOCITY_PGAIN * ERR_VELOCITY + VELOCITY_IGAIN * ERR_VELOCITY_INT; REF_CURR_Q = -1.0 * REF_CURR_Q; MOT_ANGLE_OLD = MOT_ANGLE; MOT_VELOCITY_OLD = MOT_VELOCITY; // select control mode //motor.drive_profile().unwrap(); //motor.drive_sixstep().unwrap(); //motor.drive_anglebased_sixstep(electric_angle).unwrap(); motor.drive_foc(electric_angle, CURRENT_A, CURRENT_B, CURRENT_C, REF_CURR_D, REF_CURR_Q).unwrap(); } /* // LED Debug cortex_m::interrupt::free(|cs| { if let Some(ref mut led) = G_LED.borrow(cs).borrow_mut().as_mut() { led.toggle().unwrap(); } }); */ } #[entry] fn main() -> ! { let dp = pac::Peripherals::take().unwrap(); let cp = cortex_m::peripheral::Peripherals::take().unwrap(); let rcc = dp.RCC.constrain(); let clocks = rcc .cfgr .use_hse(8.mhz()) .sysclk(168.mhz()) .hclk(168.mhz()) .pclk1(42.mhz()) .pclk2(84.mhz()) .require_pll48clk() .freeze(); let mut delay = hal::delay::Delay::new(cp.SYST, clocks); let gpioa = dp.GPIOA.split(); let gpiob = dp.GPIOB.split(); let gpioc = dp.GPIOC.split(); // SPI3 let sck = gpioc.pc10.into_alternate_af6(); let miso = gpioc.pc11.into_alternate_af6(); let mosi = gpioc.pc12.into_alternate_af6(); let mut spi = Spi::spi3( dp.SPI3, (sck, miso, mosi), MODE_1, KiloHertz(2000).into(), clocks, ); // DRV8301 let ncs = gpioc.pc13.into_push_pull_output(); let en_gate = gpiob.pb12.into_push_pull_output(); let mut drv8301 = Drv8301::new(&mut spi, ncs, en_gate); drv8301.init().unwrap(); // Move the pin into our global storage cortex_m::interrupt::free(|cs| *G_SPI3.borrow(cs).borrow_mut() = Some(spi)); // PWM let channels = (gpioa.pa8.into_alternate_af1(), gpioa.pa9.into_alternate_af1(), gpioa.pa10.into_alternate_af1(), gpioa.pa11.into_alternate_af1()); let pwm = pwm::tim1(dp.TIM1, channels, clocks, 16u32.khz()); let (ch1, ch2, ch3, ch4) = pwm; let mut ch4 = ch4; { // Set complementary oututs mode as AF1 gpiob.pb13.into_alternate_af1(); gpiob.pb14.into_alternate_af1(); gpiob.pb15.into_alternate_af1(); unsafe { let tim1_regb = &(*(TIM1::ptr())); // Enable complementary outputs tim1_regb.ccer.modify(|_, w| w.cc1ne().set_bit()); tim1_regb.ccer.modify(|_, w| w
type TypeEncoder<'a> = AS5048A<'a, TypeSpi3, gpioa::PA3<Output<PushPull>>>; //static G_AS5048A: Mutex<RefCell<Option<TypeEncoder>>> = Mutex::new(RefCell::new(None)); type TypeMotor = Motor;
random_line_split
MouvementBrownien.py
= -math.floor(GRID_HEIGHT/2); # Coordonnees dans referentiel enonce GRID_Y_MAX = math.floor(GRID_HEIGHT/2); # Coordonnees dans referentiel enonce # CANVAS_WIDTH = GRID_WIDTH*GRID_STEP; # Taille resultante du canvas (zone dessin) CANVAS_HEIGHT = GRID_HEIGHT*GRID_STEP; # Taille resultante du canvas (zone dessin) CANVAS_BG_COLOR = "white"; # Couleur de fond du canvas (zone dessin) # NB_PARTICLES = 500; # Nombre de particles a simuler PARTICLE_SIZE = GRID_STEP*2/3; # Taille des cercles pour representer chq particule PARTICLE_COLOR = "red"; # Couleur des particules # NB_SIMU_STEPS = 100; # Nombre de pas de temps dans la simulation SIMU_INVSPEED = 100; # Nombre de millisecondes entre chaque pas de temps ## The only shared var paused=False; #### Fonction des particules # Initialise les positions des particules def initParticles(N): particles = []; for p in xrange(N): initialPos = { 'x': 0, 'y': 0 }; particles.append(initialPos); return particles; # Deplace aleatoirement une unique particule, en respectant la "condition de bordure" def moveSingleParticle(oldPos): aleat = random.randint(0, 3); if (aleat==0): newPos = { 'x':oldPos['x']+1, 'y': oldPos['y']}; elif (aleat==1): newPos = { 'x': oldPos['x']-1, 'y': oldPos['y']}; elif (aleat==2): newPos = { 'x': oldPos['x'], 'y': oldPos['y']+1}; elif (aleat==3): newPos = { 'x': oldPos['x'], 'y': oldPos['y']-1}; else: print("Should not have reached here! "+str(aleat)+" is not an accepted output"); ## "Lorsqu'une particule rencontre une paroi, elle ne bouge pas si le mouvement ## determine aleatoirement la fait traverser cette paroi." if (newPos['x']<GRID_X_MIN+1): newPos['x'] = GRID_X_MIN+1; elif (newPos['x']>GRID_X_MAX-1): newPos['x'] = GRID_X_MAX-1; elif (newPos['y']<GRID_Y_MIN+1): newPos['y'] = GRID_Y_MIN+1; elif (newPos['y']>GRID_Y_MAX-1): newPos['y'] = GRID_Y_MAX-1; return newPos; # Deplace toutes les particules d'un pas aleatoire def moveParticles(particles): newParticles = []; for p in xrange(len(particles)): newParticles.append(moveSingleParticle(particles[p])); return(newParticles); def applyGravity(particles): newParticles = []; for p in xrange(len(particles)): part = particles[p]; if (part['y']<GRID_Y_MAX-1): # Only change value if it particle does not exit screen part['y'] += 1; # Make particle go down newParticles.append(part); return(newParticles); #### Problem/Model to GUI/View functions # Convertit la liste des positions des particules en un tableau # (proche de la "grille graphique") def convertToMatrix(particles): matrix = [[0 for x in range(GRID_WIDTH)] for y in range(GRID_WIDTH)]; for p in xrange(len(particles)): currentParticle = particles[p]; tx = int(math.floor(currentParticle['x']-GRID_X_MIN)); ty = int(math.floor(currentParticle['y']-GRID_Y_MIN)); ##print("----------- Particle position: ("+str(x)+","+str(y)+")->("+str(tx)+","+str(ty)+")"); ## Debug ##print("----------- Grid: ["+str(GRID_X_MIN)+"->"+str(GRID_X_MAX)+" ; "+str(GRID_Y_MIN)+"->"+str(GRID_Y_MAX)+"] / ("+str(GRID_WIDTH)+", "+str(GRID_HEIGHT)+")"); matrix[tx][ty] += 1; ##print(matrix); # for debugging return matrix; #### Fonctions graphiques # Un/Pauses the simulation def pause(pauseButton): global paused; ## required to set global var paused = not paused; if (paused): pauseButton.config(text="Unpause"); else: pauseButton.config(text="Pause"); # Creation & Placement des elements graphiques def initGUI(rootWindow): ##global canvas; ## for debugging purpose canvas = tkinter.Canvas(rootWindow, width=CANVAS_WIDTH, height=CANVAS_HEIGHT); canvas.config(background=CANVAS_BG_COLOR); canvas.pack(); slider = tkinter.Scale(rootWindow, from_=0, to=10, orient=tkinter.HORIZONTAL); slider.set(0); #slider.config(height=GUI_ELT_HEIGHT, width=GUI_ELT_WIDTH); slider.pack(); label = tkinter.Label(rootWindow, text='t=0'); labelfont = ('times', 20, 'bold'); label.config(bg='black', fg='yellow'); label.config(font=labelfont); label.config(height=GUI_ELT_HEIGHT, width=GUI_ELT_WIDTH); label.pack(); # expand=YES, fill=BOTH startButton = tkinter.Button(rootWindow, text="Démarrer", command= lambda: startSimulationLoop(canvas,label,slider)); startButton.config(height=GUI_ELT_HEIGHT, width=GUI_ELT_WIDTH); startButton.pack(); pauseButton = tkinter.Button(rootWindow, text="Pause", command=lambda: pause(pauseButton)); pauseButton.config(height=GUI_ELT_HEIGHT, width=GUI_ELT_WIDTH); pauseButton.pack(); exitButton = tkinter.Button(rootWindow, text="Sortir", command=rootWindow.destroy); exitButton.config(height=GUI_ELT_HEIGHT, width=GUI_ELT_WIDTH); exitButton.pack(); def drawGrid(canvas): for x in xrange(0, CANVAS_WIDTH, GRID_STEP): l1 = canvas.create_line(x, 0, x, CANVAS_HEIGHT, fill=GRID_COLOR); canvas.itemconfig(l1, tags=("vl")); ## Useless for y in xrange(0, CANVAS_HEIGHT, GRID_STEP): l2 = canvas.create_line(0, y, CANVAS_WIDTH, y, fill=GRID_COLOR); canvas.itemconfig(l2, tags=("hl")); ## Useless ctr = canvas.create_oval(-GRID_X_MIN*GRID_STEP-PARTICLE_SIZE/2, -GRID_Y_MIN*GRID_STEP-PARTICLE_SIZE/2, -GRID_X_MIN*GRID_STEP+PARTICLE_SIZE/2, -GRID_Y_MIN*GRID_STEP+PARTICLE_SIZE/2, outline="grey", fill="grey"); # def drawParticlesFromPositions(canvas, particles): # for p in xrange(len(particles)): # currentParticle = particles[p]; # tx = currentParticle['x']-GRID_X_MIN; # ty = currentParticle['y']-GRID_Y_MIN; # txg = tx*GRID_STEP; # tyg = ty*GRID_STEP; # o = canvas.create_oval(txg-PARTICLE_SIZE, tyg-PARTICLE_SIZE, # txg+PARTICLE_SIZE, tyg+PARTICLE_SIZE, # outline=PARTICLE_COLOR, fill=PARTICLE_COLOR); # canvas.itemconfig(o, tags=("part"+str(p))); ## Useless # canvas.update_idletasks(); # THIS IS A DIRTY HACK!!! def drawParticlesFromGrid(canvas, matrix): for x in xrange(GRID_WIDTH): for y in xrange(GRID_HEIGHT): if (matrix[x][y]>0): xg = x*GRID_STEP; yg = y*GRID_STEP; o = canvas.create_oval(xg-PARTICLE_SIZE, yg-PARTICLE_SIZE, xg+PARTICLE_SIZE, yg+PARTICLE_SIZE, outline=PARTICLE_COLOR, fill=PARTICLE_COLOR); canvas.itemconfig(o, tags=("part("+str(x)+"/"+str(y)+")")); ## Useless ##print("----------- Particle position: "+str(xg)+"/"+str(yg)); ## Debug t = canvas.create_text((xg, yg), text=str(matrix[x][y]), font=FONT); canvas.itemconfig(t, tags=("partCount("+str(x)+"/"+str(y)+")")); ## Useless canvas.update_idletasks(); # THIS IS A DIRTY HACK!!! def drawTime(label, t): l
# Lance la simulation with loop ## Problem1: we do no give a chance to the GUI elements to update => we need to force them ## Problem2: since the event loop is broken, the "Sortir" button does not work. # def startSimulationLoop(canvas, label): # particles = initParticles(NB_PARTICLES); # for step in xrange(N
abel.configure(text="t="+str(t)); label.update_idletasks(); # THIS IS A DIRTY HACK!!!
identifier_body
MouvementBrownien.py
= -math.floor(GRID_HEIGHT/2); # Coordonnees dans referentiel enonce GRID_Y_MAX = math.floor(GRID_HEIGHT/2); # Coordonnees dans referentiel enonce # CANVAS_WIDTH = GRID_WIDTH*GRID_STEP; # Taille resultante du canvas (zone dessin) CANVAS_HEIGHT = GRID_HEIGHT*GRID_STEP; # Taille resultante du canvas (zone dessin) CANVAS_BG_COLOR = "white"; # Couleur de fond du canvas (zone dessin) # NB_PARTICLES = 500; # Nombre de particles a simuler PARTICLE_SIZE = GRID_STEP*2/3; # Taille des cercles pour representer chq particule PARTICLE_COLOR = "red"; # Couleur des particules # NB_SIMU_STEPS = 100; # Nombre de pas de temps dans la simulation SIMU_INVSPEED = 100; # Nombre de millisecondes entre chaque pas de temps ## The only shared var paused=False; #### Fonction des particules # Initialise les positions des particules def initParticles(N): particles = []; for p in xrange(N): initialPos = { 'x': 0, 'y': 0 }; particles.append(initialPos); return particles; # Deplace aleatoirement une unique particule, en respectant la "condition de bordure" def moveSingleParticle(oldPos): aleat = random.randint(0, 3); if (aleat==0): newPos = { 'x':oldPos['x']+1, 'y': oldPos['y']}; elif (aleat==1): newPos = { 'x': oldPos['x']-1, 'y': oldPos['y']}; elif (aleat==2): newPos = { 'x': oldPos['x'], 'y': oldPos['y']+1}; elif (aleat==3): newPos = { 'x': oldPos['x'], 'y': oldPos['y']-1}; else: print("Should not have reached here! "+str(aleat)+" is not an accepted output"); ## "Lorsqu'une particule rencontre une paroi, elle ne bouge pas si le mouvement ## determine aleatoirement la fait traverser cette paroi." if (newPos['x']<GRID_X_MIN+1): newPos['x'] = GRID_X_MIN+1; elif (newPos['x']>GRID_X_MAX-1): newPos['x'] = GRID_X_MAX-1; elif (newPos['y']<GRID_Y_MIN+1): newPos['y'] = GRID_Y_MIN+1; elif (newPos['y']>GRID_Y_MAX-1): newPos['y'] = GRID_Y_MAX-1; return newPos; # Deplace toutes les particules d'un pas aleatoire def moveParticles(particles): newParticles = []; for p in xrange(len(particles)): newParticles.append(moveSingleParticle(particles[p])); return(newParticles); def applyGravity(particles): newParticles = []; for p in xrange(len(particles)): part = particles[p]; if (part['y']<GRID_Y_MAX-1): # Only change value if it particle does not exit screen part['y'] += 1; # Make particle go down newParticles.append(part); return(newParticles); #### Problem/Model to GUI/View functions # Convertit la liste des positions des particules en un tableau # (proche de la "grille graphique") def convertToMatrix(particles): matrix = [[0 for x in range(GRID_WIDTH)] for y in range(GRID_WIDTH)]; for p in xrange(len(particles)): currentParticle = particles[p]; tx = int(math.floor(currentParticle['x']-GRID_X_MIN)); ty = int(math.floor(currentParticle['y']-GRID_Y_MIN)); ##print("----------- Particle position: ("+str(x)+","+str(y)+")->("+str(tx)+","+str(ty)+")"); ## Debug ##print("----------- Grid: ["+str(GRID_X_MIN)+"->"+str(GRID_X_MAX)+" ; "+str(GRID_Y_MIN)+"->"+str(GRID_Y_MAX)+"] / ("+str(GRID_WIDTH)+", "+str(GRID_HEIGHT)+")"); matrix[tx][ty] += 1; ##print(matrix); # for debugging return matrix; #### Fonctions graphiques # Un/Pauses the simulation def pause(pauseButton): global paused; ## required to set global var paused = not paused; if (paused): pauseButton.config(text="Unpause"); else: pauseButton.config(text="Pause"); # Creation & Placement des elements graphiques def initGUI(rootWindow): ##global canvas; ## for debugging purpose canvas = tkinter.Canvas(rootWindow, width=CANVAS_WIDTH, height=CANVAS_HEIGHT); canvas.config(background=CANVAS_BG_COLOR); canvas.pack(); slider = tkinter.Scale(rootWindow, from_=0, to=10, orient=tkinter.HORIZONTAL); slider.set(0); #slider.config(height=GUI_ELT_HEIGHT, width=GUI_ELT_WIDTH); slider.pack(); label = tkinter.Label(rootWindow, text='t=0'); labelfont = ('times', 20, 'bold'); label.config(bg='black', fg='yellow'); label.config(font=labelfont); label.config(height=GUI_ELT_HEIGHT, width=GUI_ELT_WIDTH); label.pack(); # expand=YES, fill=BOTH startButton = tkinter.Button(rootWindow, text="Démarrer", command= lambda: startSimulationLoop(canvas,label,slider)); startButton.config(height=GUI_ELT_HEIGHT, width=GUI_ELT_WIDTH); startButton.pack(); pauseButton = tkinter.Button(rootWindow, text="Pause", command=lambda: pause(pauseButton)); pauseButton.config(height=GUI_ELT_HEIGHT, width=GUI_ELT_WIDTH); pauseButton.pack(); exitButton = tkinter.Button(rootWindow, text="Sortir", command=rootWindow.destroy); exitButton.config(height=GUI_ELT_HEIGHT, width=GUI_ELT_WIDTH); exitButton.pack(); def drawGrid(canvas): for x in xrange(0, CANVAS_WIDTH, GRID_STEP): l1 = canvas.create_line(x, 0, x, CANVAS_HEIGHT, fill=GRID_COLOR); canvas.itemconfig(l1, tags=("vl")); ## Useless for y in xrange(0, CANVAS_HEIGHT, GRID_STEP): l2 = canvas.create_line(0, y, CANVAS_WIDTH, y, fill=GRID_COLOR); canvas.itemconfig(l2, tags=("hl")); ## Useless ctr = canvas.create_oval(-GRID_X_MIN*GRID_STEP-PARTICLE_SIZE/2, -GRID_Y_MIN*GRID_STEP-PARTICLE_SIZE/2, -GRID_X_MIN*GRID_STEP+PARTICLE_SIZE/2, -GRID_Y_MIN*GRID_STEP+PARTICLE_SIZE/2, outline="grey", fill="grey"); # def drawParticlesFromPositions(canvas, particles): # for p in xrange(len(particles)): # currentParticle = particles[p]; # tx = currentParticle['x']-GRID_X_MIN;
# ty = currentParticle['y']-GRID_Y_MIN; # txg = tx*GRID_STEP; # tyg = ty*GRID_STEP; # o = canvas.create_oval(txg-PARTICLE_SIZE, tyg-PARTICLE_SIZE, # txg+PARTICLE_SIZE, tyg+PARTICLE_SIZE, # outline=PARTICLE_COLOR, fill=PARTICLE_COLOR); # canvas.itemconfig(o, tags=("part"+str(p))); ## Useless # canvas.update_idletasks(); # THIS IS A DIRTY HACK!!! def drawParticlesFromGrid(canvas, matrix): for x in xrange(GRID_WIDTH): for y in xrange(GRID_HEIGHT): if (matrix[x][y]>0): xg = x*GRID_STEP; yg = y*GRID_STEP; o = canvas.create_oval(xg-PARTICLE_SIZE, yg-PARTICLE_SIZE, xg+PARTICLE_SIZE, yg+PARTICLE_SIZE, outline=PARTICLE_COLOR, fill=PARTICLE_COLOR); canvas.itemconfig(o, tags=("part("+str(x)+"/"+str(y)+")")); ## Useless ##print("----------- Particle position: "+str(xg)+"/"+str(yg)); ## Debug t = canvas.create_text((xg, yg), text=str(matrix[x][y]), font=FONT); canvas.itemconfig(t, tags=("partCount("+str(x)+"/"+str(y)+")")); ## Useless canvas.update_idletasks(); # THIS IS A DIRTY HACK!!! def drawTime(label, t): label.configure(text="t="+str(t)); label.update_idletasks(); # THIS IS A DIRTY HACK!!! # Lance la simulation with loop ## Problem1: we do no give a chance to the GUI elements to update => we need to force them ## Problem2: since the event loop is broken, the "Sortir" button does not work. # def startSimulationLoop(canvas, label): # particles = initParticles(NB_PARTICLES); # for step in xrange(NB_SIM
random_line_split
MouvementBrownien.py
= -math.floor(GRID_HEIGHT/2); # Coordonnees dans referentiel enonce GRID_Y_MAX = math.floor(GRID_HEIGHT/2); # Coordonnees dans referentiel enonce # CANVAS_WIDTH = GRID_WIDTH*GRID_STEP; # Taille resultante du canvas (zone dessin) CANVAS_HEIGHT = GRID_HEIGHT*GRID_STEP; # Taille resultante du canvas (zone dessin) CANVAS_BG_COLOR = "white"; # Couleur de fond du canvas (zone dessin) # NB_PARTICLES = 500; # Nombre de particles a simuler PARTICLE_SIZE = GRID_STEP*2/3; # Taille des cercles pour representer chq particule PARTICLE_COLOR = "red"; # Couleur des particules # NB_SIMU_STEPS = 100; # Nombre de pas de temps dans la simulation SIMU_INVSPEED = 100; # Nombre de millisecondes entre chaque pas de temps ## The only shared var paused=False; #### Fonction des particules # Initialise les positions des particules def initParticles(N): particles = []; for p in xrange(N): initialPos = { 'x': 0, 'y': 0 }; particles.append(initialPos); return particles; # Deplace aleatoirement une unique particule, en respectant la "condition de bordure" def moveSingleParticle(oldPos): aleat = random.randint(0, 3); if (aleat==0): newPos = { 'x':oldPos['x']+1, 'y': oldPos['y']}; elif (aleat==1): newPos = { 'x': oldPos['x']-1, 'y': oldPos['y']}; elif (aleat==2): newPos = { 'x': oldPos['x'], 'y': oldPos['y']+1}; elif (aleat==3): newPos = { 'x': oldPos['x'], 'y': oldPos['y']-1}; else: print("Should not have reached here! "+str(aleat)+" is not an accepted output"); ## "Lorsqu'une particule rencontre une paroi, elle ne bouge pas si le mouvement ## determine aleatoirement la fait traverser cette paroi." if (newPos['x']<GRID_X_MIN+1): newPos['x'] = GRID_X_MIN+1; elif (newPos['x']>GRID_X_MAX-1): newPos['x'] = GRID_X_MAX-1; elif (newPos['y']<GRID_Y_MIN+1): newPos['y'] = GRID_Y_MIN+1; elif (newPos['y']>GRID_Y_MAX-1): newPos['y'] = GRID_Y_MAX-1; return newPos; # Deplace toutes les particules d'un pas aleatoire def moveParticles(particles): newParticles = []; for p in xrange(len(particles)): newParticles.append(moveSingleParticle(particles[p])); return(newParticles); def applyGravity(particles): newParticles = []; for p in xrange(len(particles)): part = particles[p]; if (part['y']<GRID_Y_MAX-1): # Only change value if it particle does not exit screen part['y'] += 1; # Make particle go down newParticles.append(part); return(newParticles); #### Problem/Model to GUI/View functions # Convertit la liste des positions des particules en un tableau # (proche de la "grille graphique") def convertToMatrix(particles): matrix = [[0 for x in range(GRID_WIDTH)] for y in range(GRID_WIDTH)]; for p in xrange(len(particles)): currentParticle = particles[p]; tx = int(math.floor(currentParticle['x']-GRID_X_MIN)); ty = int(math.floor(currentParticle['y']-GRID_Y_MIN)); ##print("----------- Particle position: ("+str(x)+","+str(y)+")->("+str(tx)+","+str(ty)+")"); ## Debug ##print("----------- Grid: ["+str(GRID_X_MIN)+"->"+str(GRID_X_MAX)+" ; "+str(GRID_Y_MIN)+"->"+str(GRID_Y_MAX)+"] / ("+str(GRID_WIDTH)+", "+str(GRID_HEIGHT)+")"); matrix[tx][ty] += 1; ##print(matrix); # for debugging return matrix; #### Fonctions graphiques # Un/Pauses the simulation def pause(pauseButton): global paused; ## required to set global var paused = not paused; if (paused): pauseButton.config(text="Unpause"); else:
# Creation & Placement des elements graphiques def initGUI(rootWindow): ##global canvas; ## for debugging purpose canvas = tkinter.Canvas(rootWindow, width=CANVAS_WIDTH, height=CANVAS_HEIGHT); canvas.config(background=CANVAS_BG_COLOR); canvas.pack(); slider = tkinter.Scale(rootWindow, from_=0, to=10, orient=tkinter.HORIZONTAL); slider.set(0); #slider.config(height=GUI_ELT_HEIGHT, width=GUI_ELT_WIDTH); slider.pack(); label = tkinter.Label(rootWindow, text='t=0'); labelfont = ('times', 20, 'bold'); label.config(bg='black', fg='yellow'); label.config(font=labelfont); label.config(height=GUI_ELT_HEIGHT, width=GUI_ELT_WIDTH); label.pack(); # expand=YES, fill=BOTH startButton = tkinter.Button(rootWindow, text="Démarrer", command= lambda: startSimulationLoop(canvas,label,slider)); startButton.config(height=GUI_ELT_HEIGHT, width=GUI_ELT_WIDTH); startButton.pack(); pauseButton = tkinter.Button(rootWindow, text="Pause", command=lambda: pause(pauseButton)); pauseButton.config(height=GUI_ELT_HEIGHT, width=GUI_ELT_WIDTH); pauseButton.pack(); exitButton = tkinter.Button(rootWindow, text="Sortir", command=rootWindow.destroy); exitButton.config(height=GUI_ELT_HEIGHT, width=GUI_ELT_WIDTH); exitButton.pack(); def drawGrid(canvas): for x in xrange(0, CANVAS_WIDTH, GRID_STEP): l1 = canvas.create_line(x, 0, x, CANVAS_HEIGHT, fill=GRID_COLOR); canvas.itemconfig(l1, tags=("vl")); ## Useless for y in xrange(0, CANVAS_HEIGHT, GRID_STEP): l2 = canvas.create_line(0, y, CANVAS_WIDTH, y, fill=GRID_COLOR); canvas.itemconfig(l2, tags=("hl")); ## Useless ctr = canvas.create_oval(-GRID_X_MIN*GRID_STEP-PARTICLE_SIZE/2, -GRID_Y_MIN*GRID_STEP-PARTICLE_SIZE/2, -GRID_X_MIN*GRID_STEP+PARTICLE_SIZE/2, -GRID_Y_MIN*GRID_STEP+PARTICLE_SIZE/2, outline="grey", fill="grey"); # def drawParticlesFromPositions(canvas, particles): # for p in xrange(len(particles)): # currentParticle = particles[p]; # tx = currentParticle['x']-GRID_X_MIN; # ty = currentParticle['y']-GRID_Y_MIN; # txg = tx*GRID_STEP; # tyg = ty*GRID_STEP; # o = canvas.create_oval(txg-PARTICLE_SIZE, tyg-PARTICLE_SIZE, # txg+PARTICLE_SIZE, tyg+PARTICLE_SIZE, # outline=PARTICLE_COLOR, fill=PARTICLE_COLOR); # canvas.itemconfig(o, tags=("part"+str(p))); ## Useless # canvas.update_idletasks(); # THIS IS A DIRTY HACK!!! def drawParticlesFromGrid(canvas, matrix): for x in xrange(GRID_WIDTH): for y in xrange(GRID_HEIGHT): if (matrix[x][y]>0): xg = x*GRID_STEP; yg = y*GRID_STEP; o = canvas.create_oval(xg-PARTICLE_SIZE, yg-PARTICLE_SIZE, xg+PARTICLE_SIZE, yg+PARTICLE_SIZE, outline=PARTICLE_COLOR, fill=PARTICLE_COLOR); canvas.itemconfig(o, tags=("part("+str(x)+"/"+str(y)+")")); ## Useless ##print("----------- Particle position: "+str(xg)+"/"+str(yg)); ## Debug t = canvas.create_text((xg, yg), text=str(matrix[x][y]), font=FONT); canvas.itemconfig(t, tags=("partCount("+str(x)+"/"+str(y)+")")); ## Useless canvas.update_idletasks(); # THIS IS A DIRTY HACK!!! def drawTime(label, t): label.configure(text="t="+str(t)); label.update_idletasks(); # THIS IS A DIRTY HACK!!! # Lance la simulation with loop ## Problem1: we do no give a chance to the GUI elements to update => we need to force them ## Problem2: since the event loop is broken, the "Sortir" button does not work. # def startSimulationLoop(canvas, label): # particles = initParticles(NB_PARTICLES); # for step in xrange(NB
pauseButton.config(text="Pause");
conditional_block
MouvementBrownien.py
= -math.floor(GRID_HEIGHT/2); # Coordonnees dans referentiel enonce GRID_Y_MAX = math.floor(GRID_HEIGHT/2); # Coordonnees dans referentiel enonce # CANVAS_WIDTH = GRID_WIDTH*GRID_STEP; # Taille resultante du canvas (zone dessin) CANVAS_HEIGHT = GRID_HEIGHT*GRID_STEP; # Taille resultante du canvas (zone dessin) CANVAS_BG_COLOR = "white"; # Couleur de fond du canvas (zone dessin) # NB_PARTICLES = 500; # Nombre de particles a simuler PARTICLE_SIZE = GRID_STEP*2/3; # Taille des cercles pour representer chq particule PARTICLE_COLOR = "red"; # Couleur des particules # NB_SIMU_STEPS = 100; # Nombre de pas de temps dans la simulation SIMU_INVSPEED = 100; # Nombre de millisecondes entre chaque pas de temps ## The only shared var paused=False; #### Fonction des particules # Initialise les positions des particules def initParticles(N): particles = []; for p in xrange(N): initialPos = { 'x': 0, 'y': 0 }; particles.append(initialPos); return particles; # Deplace aleatoirement une unique particule, en respectant la "condition de bordure" def moveSingleParticle(oldPos): aleat = random.randint(0, 3); if (aleat==0): newPos = { 'x':oldPos['x']+1, 'y': oldPos['y']}; elif (aleat==1): newPos = { 'x': oldPos['x']-1, 'y': oldPos['y']}; elif (aleat==2): newPos = { 'x': oldPos['x'], 'y': oldPos['y']+1}; elif (aleat==3): newPos = { 'x': oldPos['x'], 'y': oldPos['y']-1}; else: print("Should not have reached here! "+str(aleat)+" is not an accepted output"); ## "Lorsqu'une particule rencontre une paroi, elle ne bouge pas si le mouvement ## determine aleatoirement la fait traverser cette paroi." if (newPos['x']<GRID_X_MIN+1): newPos['x'] = GRID_X_MIN+1; elif (newPos['x']>GRID_X_MAX-1): newPos['x'] = GRID_X_MAX-1; elif (newPos['y']<GRID_Y_MIN+1): newPos['y'] = GRID_Y_MIN+1; elif (newPos['y']>GRID_Y_MAX-1): newPos['y'] = GRID_Y_MAX-1; return newPos; # Deplace toutes les particules d'un pas aleatoire def moveParticles(particles): newParticles = []; for p in xrange(len(particles)): newParticles.append(moveSingleParticle(particles[p])); return(newParticles); def applyGravity(particles): newParticles = []; for p in xrange(len(particles)): part = particles[p]; if (part['y']<GRID_Y_MAX-1): # Only change value if it particle does not exit screen part['y'] += 1; # Make particle go down newParticles.append(part); return(newParticles); #### Problem/Model to GUI/View functions # Convertit la liste des positions des particules en un tableau # (proche de la "grille graphique") def convertToMatrix(particles): matrix = [[0 for x in range(GRID_WIDTH)] for y in range(GRID_WIDTH)]; for p in xrange(len(particles)): currentParticle = particles[p]; tx = int(math.floor(currentParticle['x']-GRID_X_MIN)); ty = int(math.floor(currentParticle['y']-GRID_Y_MIN)); ##print("----------- Particle position: ("+str(x)+","+str(y)+")->("+str(tx)+","+str(ty)+")"); ## Debug ##print("----------- Grid: ["+str(GRID_X_MIN)+"->"+str(GRID_X_MAX)+" ; "+str(GRID_Y_MIN)+"->"+str(GRID_Y_MAX)+"] / ("+str(GRID_WIDTH)+", "+str(GRID_HEIGHT)+")"); matrix[tx][ty] += 1; ##print(matrix); # for debugging return matrix; #### Fonctions graphiques # Un/Pauses the simulation def pause(pauseButton): global paused; ## required to set global var paused = not paused; if (paused): pauseButton.config(text="Unpause"); else: pauseButton.config(text="Pause"); # Creation & Placement des elements graphiques def initGUI(rootWindow): ##global canvas; ## for debugging purpose canvas = tkinter.Canvas(rootWindow, width=CANVAS_WIDTH, height=CANVAS_HEIGHT); canvas.config(background=CANVAS_BG_COLOR); canvas.pack(); slider = tkinter.Scale(rootWindow, from_=0, to=10, orient=tkinter.HORIZONTAL); slider.set(0); #slider.config(height=GUI_ELT_HEIGHT, width=GUI_ELT_WIDTH); slider.pack(); label = tkinter.Label(rootWindow, text='t=0'); labelfont = ('times', 20, 'bold'); label.config(bg='black', fg='yellow'); label.config(font=labelfont); label.config(height=GUI_ELT_HEIGHT, width=GUI_ELT_WIDTH); label.pack(); # expand=YES, fill=BOTH startButton = tkinter.Button(rootWindow, text="Démarrer", command= lambda: startSimulationLoop(canvas,label,slider)); startButton.config(height=GUI_ELT_HEIGHT, width=GUI_ELT_WIDTH); startButton.pack(); pauseButton = tkinter.Button(rootWindow, text="Pause", command=lambda: pause(pauseButton)); pauseButton.config(height=GUI_ELT_HEIGHT, width=GUI_ELT_WIDTH); pauseButton.pack(); exitButton = tkinter.Button(rootWindow, text="Sortir", command=rootWindow.destroy); exitButton.config(height=GUI_ELT_HEIGHT, width=GUI_ELT_WIDTH); exitButton.pack(); def d
canvas): for x in xrange(0, CANVAS_WIDTH, GRID_STEP): l1 = canvas.create_line(x, 0, x, CANVAS_HEIGHT, fill=GRID_COLOR); canvas.itemconfig(l1, tags=("vl")); ## Useless for y in xrange(0, CANVAS_HEIGHT, GRID_STEP): l2 = canvas.create_line(0, y, CANVAS_WIDTH, y, fill=GRID_COLOR); canvas.itemconfig(l2, tags=("hl")); ## Useless ctr = canvas.create_oval(-GRID_X_MIN*GRID_STEP-PARTICLE_SIZE/2, -GRID_Y_MIN*GRID_STEP-PARTICLE_SIZE/2, -GRID_X_MIN*GRID_STEP+PARTICLE_SIZE/2, -GRID_Y_MIN*GRID_STEP+PARTICLE_SIZE/2, outline="grey", fill="grey"); # def drawParticlesFromPositions(canvas, particles): # for p in xrange(len(particles)): # currentParticle = particles[p]; # tx = currentParticle['x']-GRID_X_MIN; # ty = currentParticle['y']-GRID_Y_MIN; # txg = tx*GRID_STEP; # tyg = ty*GRID_STEP; # o = canvas.create_oval(txg-PARTICLE_SIZE, tyg-PARTICLE_SIZE, # txg+PARTICLE_SIZE, tyg+PARTICLE_SIZE, # outline=PARTICLE_COLOR, fill=PARTICLE_COLOR); # canvas.itemconfig(o, tags=("part"+str(p))); ## Useless # canvas.update_idletasks(); # THIS IS A DIRTY HACK!!! def drawParticlesFromGrid(canvas, matrix): for x in xrange(GRID_WIDTH): for y in xrange(GRID_HEIGHT): if (matrix[x][y]>0): xg = x*GRID_STEP; yg = y*GRID_STEP; o = canvas.create_oval(xg-PARTICLE_SIZE, yg-PARTICLE_SIZE, xg+PARTICLE_SIZE, yg+PARTICLE_SIZE, outline=PARTICLE_COLOR, fill=PARTICLE_COLOR); canvas.itemconfig(o, tags=("part("+str(x)+"/"+str(y)+")")); ## Useless ##print("----------- Particle position: "+str(xg)+"/"+str(yg)); ## Debug t = canvas.create_text((xg, yg), text=str(matrix[x][y]), font=FONT); canvas.itemconfig(t, tags=("partCount("+str(x)+"/"+str(y)+")")); ## Useless canvas.update_idletasks(); # THIS IS A DIRTY HACK!!! def drawTime(label, t): label.configure(text="t="+str(t)); label.update_idletasks(); # THIS IS A DIRTY HACK!!! # Lance la simulation with loop ## Problem1: we do no give a chance to the GUI elements to update => we need to force them ## Problem2: since the event loop is broken, the "Sortir" button does not work. # def startSimulationLoop(canvas, label): # particles = initParticles(NB_PARTICLES); # for step in xrange(N
rawGrid(
identifier_name
roster.go
(m *gtk.TreeStore, iter *gtk.TreeIter, index int) string { val, _ := m.GetValue(iter, index) v, _ := val.GetString() return v } func (r *roster) getAccountAndJidFromEvent(bt *gdk.EventButton) (jid string, account *account, rowType string, ok bool) { x := bt.X() y := bt.Y() path := new(gtk.TreePath) found := r.view.GetPathAtPos(int(x), int(y), path, nil, nil, nil) if !found { return "", nil, "", false } iter, err := r.model.GetIter(path) if err != nil { return "", nil, "", false } jid = getFromModelIter(r.model, iter, indexJid) accountID := getFromModelIter(r.model, iter, indexAccountID) rowType = getFromModelIter(r.model, iter, indexRowType) account, ok = r.getAccount(accountID) return jid, account, rowType, ok } func (r *roster) createAccountPeerPopup(jid string, account *account, bt *gdk.EventButton) { builder := builderForDefinition("ContactPopupMenu") obj, _ := builder.GetObject("contactMenu") mn := obj.(*gtk.Menu) builder.ConnectSignals(map[string]interface{}{ "on_remove_contact": func() { account.session.RemoveContact(jid) r.ui.removePeer(account, jid) r.redraw() }, "on_allow_contact_to_see_status": func() { account.session.ApprovePresenceSubscription(jid, "" /* generate id */) }, "on_forbid_contact_to_see_status": func() { account.session.DenyPresenceSubscription(jid, "" /* generate id */) }, "on_ask_contact_to_see_status": func() { account.session.RequestPresenceSubscription(jid) }, "on_dump_info": func() { r.debugPrintRosterFor(account.session.GetConfig().Account) }, }) mn.ShowAll() mn.PopupAtMouseCursor(nil, nil, int(bt.Button()), bt.Time()) } func (r *roster) createAccountPopup(jid string, account *account, bt *gdk.EventButton) { builder := builderForDefinition("AccountPopupMenu") obj, _ := builder.GetObject("accountMenu") mn := obj.(*gtk.Menu) builder.ConnectSignals(map[string]interface{}{ "on_connect": func() { account.connect() }, "on_disconnect": func() { account.disconnect() }, "on_dump_info": func() { r.debugPrintRosterFor(account.session.GetConfig().Account) }, }) connx, _ := builder.GetObject("connectMenuItem") connect := connx.(*gtk.MenuItem) dconnx, _ := builder.GetObject("disconnectMenuItem") disconnect := dconnx.(*gtk.MenuItem) connect.SetSensitive(account.session.IsDisconnected()) disconnect.SetSensitive(account.session.IsConnected()) mn.ShowAll() mn.PopupAtMouseCursor(nil, nil, int(bt.Button()), bt.Time()) } func (r *roster) onButtonPress(view *gtk.TreeView, ev *gdk.Event) bool { bt := &gdk.EventButton{ev} if bt.Button() == 0x03 { jid, account, rowType, ok := r.getAccountAndJidFromEvent(bt) if ok { switch rowType { case "peer": r.createAccountPeerPopup(jid, account, bt) case "account": r.createAccountPopup(jid, account, bt) } } } return false } func (r *roster) onActivateBuddy(v *gtk.TreeView, path *gtk.TreePath) { selection, _ := v.GetSelection() defer selection.UnselectPath(path) iter, err := r.model.GetIter(path) if err != nil { return } jid := getFromModelIter(r.model, iter, indexJid) accountID := getFromModelIter(r.model, iter, indexAccountID) rowType := getFromModelIter(r.model, iter, indexRowType) if rowType != "peer" { r.isCollapsed[jid] = !r.isCollapsed[jid] r.redraw() return } account, ok := r.getAccount(accountID) if !ok { return } r.openConversationWindow(account, jid) } func (r *roster) openConversationWindow(account *account, to string) (*conversationWindow, error) { c, ok := account.getConversationWith(to) if !ok { textBuffer := r.ui.getTags().createTextBuffer() c = account.createConversationWindow(to, r.ui.displaySettings, textBuffer) r.ui.connectShortcutsChildWindow(c.win) r.ui.connectShortcutsConversationWindow(c) c.parentWin = r.ui.window } c.Show() return c, nil } func (r *roster) displayNameFor(account *account, from string) string { p, ok := r.ui.getPeer(account, from) if !ok { return from } return p.NameForPresentation() } func (r *roster) presenceUpdated(account *account, from, show, showStatus string, gone bool) { c, ok := account.getConversationWith(from) if !ok { return } doInUIThread(func() { c.appendStatus(r.displayNameFor(account, from), time.Now(), show, showStatus, gone) }) } func (r *roster) messageReceived(account *account, from string, timestamp time.Time, encrypted bool, message []byte) { doInUIThread(func() { conv, err := r.openConversationWindow(account, from) if err != nil { return } conv.appendMessage(r.displayNameFor(account, from), timestamp, encrypted, ui.StripHTML(message), false) }) } func (r *roster) update(account *account, entries *rosters.List) { r.ui.accountManager.Lock() defer r.ui.accountManager.Unlock() r.ui.accountManager.setContacts(account, entries) } func (r *roster) debugPrintRosterFor(nm string) { r.ui.accountManager.RLock() defer r.ui.accountManager.RUnlock() for account, rs := range r.ui.accountManager.getAllContacts() { if account.session.GetConfig().Is(nm) { rs.Iter(func(_ int, item *rosters.Peer) { fmt.Printf("-> %s\n", item.Dump()) }) } } fmt.Printf(" ************************************** \n") fmt.Println() } func isNominallyVisible(p *rosters.Peer) bool { return (p.Subscription != "none" && p.Subscription != "") || p.PendingSubscribeID != "" } func shouldDisplay(p *rosters.Peer, showOffline bool) bool { return isNominallyVisible(p) && (showOffline || p.Online) } func isAway(p *rosters.Peer) bool { switch p.Status { case "dnd", "xa", "away": return true } return false } func isOnline(p *rosters.Peer) bool { return p.PendingSubscribeID == "" && p.Online } func decideStatusFor(p *rosters.Peer) string { if p.PendingSubscribeID != "" { return "unknown" } if !p.Online { return "offline" } switch p.Status { case "dnd": return "busy" case "xa": return "extended-away" case "away": return "away" } return "available" } func decideColorFor(p *rosters.Peer) string { if !p.Online { return "#aaaaaa" } return "#000000" } func createGroupDisplayName(parentName string, counter *counter, isExpanded bool) string { name := parentName if !isExpanded { name = fmt.Sprintf("[%s]", name) } return fmt.Sprintf("%s (%d/%d)", name, counter.online, counter.total) } func createTooltipFor(item *rosters.Peer) string { pname := html.EscapeString(item.NameForPresentation()) jid := html.EscapeString(item.Jid) if pname != jid { return fmt.Sprintf("%s (%s)", pname, jid) } return jid } func (r *roster) addItem(item *rosters.Peer, parentIter *gtk.TreeIter, indent string) { iter := r.model.Append(parentIter) setAll(r.model, iter, item.Jid, fmt.Sprintf("%s %s", indent, item.NameForPresentation()), item.BelongsTo, decideColorFor(item), "#ffffff", nil, createTooltipFor(item), ) r.model.SetValue(iter, indexRowType, "peer") r.model.SetValue(iter, indexStatusIcon, statusIcons[decideStatusFor(item)].getPixbuf()) } func (r *roster) redrawMerged() { showOffline := !r.ui.config.Display.ShowOnlyOnline r.ui.accountManager.RLock() defer r.ui.accountManager.RUnlock() r.toCollapse = nil grp := rosters.TopLevelGroup() for account, contacts := range r.ui.accountManager.getAllContacts() { contacts.AddTo(grp, account.session.GroupDelimiter) } accountCounter := &counter{} r.displayGroup(grp, nil, accountCounter, showOffline, "") r.view.ExpandAll() for _, path := range r.toCollapse {
getFromModelIter
identifier_name
roster.go
builder.GetObject("accountMenu") mn := obj.(*gtk.Menu) builder.ConnectSignals(map[string]interface{}{ "on_connect": func() { account.connect() }, "on_disconnect": func() { account.disconnect() }, "on_dump_info": func() { r.debugPrintRosterFor(account.session.GetConfig().Account) }, }) connx, _ := builder.GetObject("connectMenuItem") connect := connx.(*gtk.MenuItem) dconnx, _ := builder.GetObject("disconnectMenuItem") disconnect := dconnx.(*gtk.MenuItem) connect.SetSensitive(account.session.IsDisconnected()) disconnect.SetSensitive(account.session.IsConnected()) mn.ShowAll() mn.PopupAtMouseCursor(nil, nil, int(bt.Button()), bt.Time()) } func (r *roster) onButtonPress(view *gtk.TreeView, ev *gdk.Event) bool { bt := &gdk.EventButton{ev} if bt.Button() == 0x03 { jid, account, rowType, ok := r.getAccountAndJidFromEvent(bt) if ok { switch rowType { case "peer": r.createAccountPeerPopup(jid, account, bt) case "account": r.createAccountPopup(jid, account, bt) } } } return false } func (r *roster) onActivateBuddy(v *gtk.TreeView, path *gtk.TreePath) { selection, _ := v.GetSelection() defer selection.UnselectPath(path) iter, err := r.model.GetIter(path) if err != nil { return } jid := getFromModelIter(r.model, iter, indexJid) accountID := getFromModelIter(r.model, iter, indexAccountID) rowType := getFromModelIter(r.model, iter, indexRowType) if rowType != "peer" { r.isCollapsed[jid] = !r.isCollapsed[jid] r.redraw() return } account, ok := r.getAccount(accountID) if !ok { return } r.openConversationWindow(account, jid) } func (r *roster) openConversationWindow(account *account, to string) (*conversationWindow, error) { c, ok := account.getConversationWith(to) if !ok { textBuffer := r.ui.getTags().createTextBuffer() c = account.createConversationWindow(to, r.ui.displaySettings, textBuffer) r.ui.connectShortcutsChildWindow(c.win) r.ui.connectShortcutsConversationWindow(c) c.parentWin = r.ui.window } c.Show() return c, nil } func (r *roster) displayNameFor(account *account, from string) string { p, ok := r.ui.getPeer(account, from) if !ok { return from } return p.NameForPresentation() } func (r *roster) presenceUpdated(account *account, from, show, showStatus string, gone bool) { c, ok := account.getConversationWith(from) if !ok { return } doInUIThread(func() { c.appendStatus(r.displayNameFor(account, from), time.Now(), show, showStatus, gone) }) } func (r *roster) messageReceived(account *account, from string, timestamp time.Time, encrypted bool, message []byte) { doInUIThread(func() { conv, err := r.openConversationWindow(account, from) if err != nil { return } conv.appendMessage(r.displayNameFor(account, from), timestamp, encrypted, ui.StripHTML(message), false) }) } func (r *roster) update(account *account, entries *rosters.List) { r.ui.accountManager.Lock() defer r.ui.accountManager.Unlock() r.ui.accountManager.setContacts(account, entries) } func (r *roster) debugPrintRosterFor(nm string) { r.ui.accountManager.RLock() defer r.ui.accountManager.RUnlock() for account, rs := range r.ui.accountManager.getAllContacts() { if account.session.GetConfig().Is(nm) { rs.Iter(func(_ int, item *rosters.Peer) { fmt.Printf("-> %s\n", item.Dump()) }) } } fmt.Printf(" ************************************** \n") fmt.Println() } func isNominallyVisible(p *rosters.Peer) bool { return (p.Subscription != "none" && p.Subscription != "") || p.PendingSubscribeID != "" } func shouldDisplay(p *rosters.Peer, showOffline bool) bool { return isNominallyVisible(p) && (showOffline || p.Online) } func isAway(p *rosters.Peer) bool { switch p.Status { case "dnd", "xa", "away": return true } return false } func isOnline(p *rosters.Peer) bool { return p.PendingSubscribeID == "" && p.Online } func decideStatusFor(p *rosters.Peer) string { if p.PendingSubscribeID != "" { return "unknown" } if !p.Online { return "offline" } switch p.Status { case "dnd": return "busy" case "xa": return "extended-away" case "away": return "away" } return "available" } func decideColorFor(p *rosters.Peer) string { if !p.Online { return "#aaaaaa" } return "#000000" } func createGroupDisplayName(parentName string, counter *counter, isExpanded bool) string { name := parentName if !isExpanded { name = fmt.Sprintf("[%s]", name) } return fmt.Sprintf("%s (%d/%d)", name, counter.online, counter.total) } func createTooltipFor(item *rosters.Peer) string { pname := html.EscapeString(item.NameForPresentation()) jid := html.EscapeString(item.Jid) if pname != jid { return fmt.Sprintf("%s (%s)", pname, jid) } return jid } func (r *roster) addItem(item *rosters.Peer, parentIter *gtk.TreeIter, indent string) { iter := r.model.Append(parentIter) setAll(r.model, iter, item.Jid, fmt.Sprintf("%s %s", indent, item.NameForPresentation()), item.BelongsTo, decideColorFor(item), "#ffffff", nil, createTooltipFor(item), ) r.model.SetValue(iter, indexRowType, "peer") r.model.SetValue(iter, indexStatusIcon, statusIcons[decideStatusFor(item)].getPixbuf()) } func (r *roster) redrawMerged() { showOffline := !r.ui.config.Display.ShowOnlyOnline r.ui.accountManager.RLock() defer r.ui.accountManager.RUnlock() r.toCollapse = nil grp := rosters.TopLevelGroup() for account, contacts := range r.ui.accountManager.getAllContacts() { contacts.AddTo(grp, account.session.GroupDelimiter) } accountCounter := &counter{} r.displayGroup(grp, nil, accountCounter, showOffline, "") r.view.ExpandAll() for _, path := range r.toCollapse { r.view.CollapseRow(path) } } type counter struct { total int online int } func (c *counter) inc(total, online bool) { if total { c.total++ } if online { c.online++ } } func (r *roster) displayGroup(g *rosters.Group, parentIter *gtk.TreeIter, accountCounter *counter, showOffline bool, accountName string) { pi := parentIter groupCounter := &counter{} groupID := accountName + "//" + g.FullGroupName() if g.GroupName != "" { pi = r.model.Append(parentIter) r.model.SetValue(pi, indexParentJid, groupID) r.model.SetValue(pi, indexRowType, "group") r.model.SetValue(pi, indexWeight, 500) r.model.SetValue(pi, indexBackgroundColor, "#e9e7f3") } for _, item := range g.Peers() { vs := isNominallyVisible(item) o := isOnline(item) accountCounter.inc(vs, vs && o) groupCounter.inc(vs, vs && o) if shouldDisplay(item, showOffline) { r.addItem(item, pi, "") } } for _, gr := range g.Groups() { r.displayGroup(gr, pi, accountCounter, showOffline, accountName) } if g.GroupName != "" { parentPath, _ := r.model.GetPath(pi) shouldCollapse, ok := r.isCollapsed[groupID] isExpanded := true if ok && shouldCollapse { isExpanded = false r.toCollapse = append(r.toCollapse, parentPath) } r.model.SetValue(pi, indexParentDisplayName, createGroupDisplayName(g.FullGroupName(), groupCounter, isExpanded)) } } func (r *roster) redrawSeparateAccount(account *account, contacts *rosters.List, showOffline bool) { parentIter := r.model.Append(nil) accountCounter := &counter{} grp := contacts.Grouped(account.session.GroupDelimiter) parentName := account.session.GetConfig().Account r.displayGroup(grp, parentIter, accountCounter, showOffline, parentName) r.model.SetValue(parentIter, indexParentJid, parentName) r.model.SetValue(parentIter, indexAccountID, account.session.GetConfig().ID())
r.model.SetValue(parentIter, indexRowType, "account")
random_line_split
roster.go
int) string { val, _ := m.GetValue(iter, index) v, _ := val.GetString() return v } func (r *roster) getAccountAndJidFromEvent(bt *gdk.EventButton) (jid string, account *account, rowType string, ok bool) { x := bt.X() y := bt.Y() path := new(gtk.TreePath) found := r.view.GetPathAtPos(int(x), int(y), path, nil, nil, nil) if !found { return "", nil, "", false } iter, err := r.model.GetIter(path) if err != nil { return "", nil, "", false } jid = getFromModelIter(r.model, iter, indexJid) accountID := getFromModelIter(r.model, iter, indexAccountID) rowType = getFromModelIter(r.model, iter, indexRowType) account, ok = r.getAccount(accountID) return jid, account, rowType, ok } func (r *roster) createAccountPeerPopup(jid string, account *account, bt *gdk.EventButton) { builder := builderForDefinition("ContactPopupMenu") obj, _ := builder.GetObject("contactMenu") mn := obj.(*gtk.Menu) builder.ConnectSignals(map[string]interface{}{ "on_remove_contact": func() { account.session.RemoveContact(jid) r.ui.removePeer(account, jid) r.redraw() }, "on_allow_contact_to_see_status": func() { account.session.ApprovePresenceSubscription(jid, "" /* generate id */) }, "on_forbid_contact_to_see_status": func() { account.session.DenyPresenceSubscription(jid, "" /* generate id */) }, "on_ask_contact_to_see_status": func() { account.session.RequestPresenceSubscription(jid) }, "on_dump_info": func() { r.debugPrintRosterFor(account.session.GetConfig().Account) }, }) mn.ShowAll() mn.PopupAtMouseCursor(nil, nil, int(bt.Button()), bt.Time()) } func (r *roster) createAccountPopup(jid string, account *account, bt *gdk.EventButton) { builder := builderForDefinition("AccountPopupMenu") obj, _ := builder.GetObject("accountMenu") mn := obj.(*gtk.Menu) builder.ConnectSignals(map[string]interface{}{ "on_connect": func() { account.connect() }, "on_disconnect": func() { account.disconnect() }, "on_dump_info": func() { r.debugPrintRosterFor(account.session.GetConfig().Account) }, }) connx, _ := builder.GetObject("connectMenuItem") connect := connx.(*gtk.MenuItem) dconnx, _ := builder.GetObject("disconnectMenuItem") disconnect := dconnx.(*gtk.MenuItem) connect.SetSensitive(account.session.IsDisconnected()) disconnect.SetSensitive(account.session.IsConnected()) mn.ShowAll() mn.PopupAtMouseCursor(nil, nil, int(bt.Button()), bt.Time()) } func (r *roster) onButtonPress(view *gtk.TreeView, ev *gdk.Event) bool
func (r *roster) onActivateBuddy(v *gtk.TreeView, path *gtk.TreePath) { selection, _ := v.GetSelection() defer selection.UnselectPath(path) iter, err := r.model.GetIter(path) if err != nil { return } jid := getFromModelIter(r.model, iter, indexJid) accountID := getFromModelIter(r.model, iter, indexAccountID) rowType := getFromModelIter(r.model, iter, indexRowType) if rowType != "peer" { r.isCollapsed[jid] = !r.isCollapsed[jid] r.redraw() return } account, ok := r.getAccount(accountID) if !ok { return } r.openConversationWindow(account, jid) } func (r *roster) openConversationWindow(account *account, to string) (*conversationWindow, error) { c, ok := account.getConversationWith(to) if !ok { textBuffer := r.ui.getTags().createTextBuffer() c = account.createConversationWindow(to, r.ui.displaySettings, textBuffer) r.ui.connectShortcutsChildWindow(c.win) r.ui.connectShortcutsConversationWindow(c) c.parentWin = r.ui.window } c.Show() return c, nil } func (r *roster) displayNameFor(account *account, from string) string { p, ok := r.ui.getPeer(account, from) if !ok { return from } return p.NameForPresentation() } func (r *roster) presenceUpdated(account *account, from, show, showStatus string, gone bool) { c, ok := account.getConversationWith(from) if !ok { return } doInUIThread(func() { c.appendStatus(r.displayNameFor(account, from), time.Now(), show, showStatus, gone) }) } func (r *roster) messageReceived(account *account, from string, timestamp time.Time, encrypted bool, message []byte) { doInUIThread(func() { conv, err := r.openConversationWindow(account, from) if err != nil { return } conv.appendMessage(r.displayNameFor(account, from), timestamp, encrypted, ui.StripHTML(message), false) }) } func (r *roster) update(account *account, entries *rosters.List) { r.ui.accountManager.Lock() defer r.ui.accountManager.Unlock() r.ui.accountManager.setContacts(account, entries) } func (r *roster) debugPrintRosterFor(nm string) { r.ui.accountManager.RLock() defer r.ui.accountManager.RUnlock() for account, rs := range r.ui.accountManager.getAllContacts() { if account.session.GetConfig().Is(nm) { rs.Iter(func(_ int, item *rosters.Peer) { fmt.Printf("-> %s\n", item.Dump()) }) } } fmt.Printf(" ************************************** \n") fmt.Println() } func isNominallyVisible(p *rosters.Peer) bool { return (p.Subscription != "none" && p.Subscription != "") || p.PendingSubscribeID != "" } func shouldDisplay(p *rosters.Peer, showOffline bool) bool { return isNominallyVisible(p) && (showOffline || p.Online) } func isAway(p *rosters.Peer) bool { switch p.Status { case "dnd", "xa", "away": return true } return false } func isOnline(p *rosters.Peer) bool { return p.PendingSubscribeID == "" && p.Online } func decideStatusFor(p *rosters.Peer) string { if p.PendingSubscribeID != "" { return "unknown" } if !p.Online { return "offline" } switch p.Status { case "dnd": return "busy" case "xa": return "extended-away" case "away": return "away" } return "available" } func decideColorFor(p *rosters.Peer) string { if !p.Online { return "#aaaaaa" } return "#000000" } func createGroupDisplayName(parentName string, counter *counter, isExpanded bool) string { name := parentName if !isExpanded { name = fmt.Sprintf("[%s]", name) } return fmt.Sprintf("%s (%d/%d)", name, counter.online, counter.total) } func createTooltipFor(item *rosters.Peer) string { pname := html.EscapeString(item.NameForPresentation()) jid := html.EscapeString(item.Jid) if pname != jid { return fmt.Sprintf("%s (%s)", pname, jid) } return jid } func (r *roster) addItem(item *rosters.Peer, parentIter *gtk.TreeIter, indent string) { iter := r.model.Append(parentIter) setAll(r.model, iter, item.Jid, fmt.Sprintf("%s %s", indent, item.NameForPresentation()), item.BelongsTo, decideColorFor(item), "#ffffff", nil, createTooltipFor(item), ) r.model.SetValue(iter, indexRowType, "peer") r.model.SetValue(iter, indexStatusIcon, statusIcons[decideStatusFor(item)].getPixbuf()) } func (r *roster) redrawMerged() { showOffline := !r.ui.config.Display.ShowOnlyOnline r.ui.accountManager.RLock() defer r.ui.accountManager.RUnlock() r.toCollapse = nil grp := rosters.TopLevelGroup() for account, contacts := range r.ui.accountManager.getAllContacts() { contacts.AddTo(grp, account.session.GroupDelimiter) } accountCounter := &counter{} r.displayGroup(grp, nil, accountCounter, showOffline, "") r.view.ExpandAll() for _, path := range r.toCollapse { r.view.CollapseRow(path) } } type counter struct { total
{ bt := &gdk.EventButton{ev} if bt.Button() == 0x03 { jid, account, rowType, ok := r.getAccountAndJidFromEvent(bt) if ok { switch rowType { case "peer": r.createAccountPeerPopup(jid, account, bt) case "account": r.createAccountPopup(jid, account, bt) } } } return false }
identifier_body
roster.go
int) string { val, _ := m.GetValue(iter, index) v, _ := val.GetString() return v } func (r *roster) getAccountAndJidFromEvent(bt *gdk.EventButton) (jid string, account *account, rowType string, ok bool) { x := bt.X() y := bt.Y() path := new(gtk.TreePath) found := r.view.GetPathAtPos(int(x), int(y), path, nil, nil, nil) if !found { return "", nil, "", false } iter, err := r.model.GetIter(path) if err != nil { return "", nil, "", false } jid = getFromModelIter(r.model, iter, indexJid) accountID := getFromModelIter(r.model, iter, indexAccountID) rowType = getFromModelIter(r.model, iter, indexRowType) account, ok = r.getAccount(accountID) return jid, account, rowType, ok } func (r *roster) createAccountPeerPopup(jid string, account *account, bt *gdk.EventButton) { builder := builderForDefinition("ContactPopupMenu") obj, _ := builder.GetObject("contactMenu") mn := obj.(*gtk.Menu) builder.ConnectSignals(map[string]interface{}{ "on_remove_contact": func() { account.session.RemoveContact(jid) r.ui.removePeer(account, jid) r.redraw() }, "on_allow_contact_to_see_status": func() { account.session.ApprovePresenceSubscription(jid, "" /* generate id */) }, "on_forbid_contact_to_see_status": func() { account.session.DenyPresenceSubscription(jid, "" /* generate id */) }, "on_ask_contact_to_see_status": func() { account.session.RequestPresenceSubscription(jid) }, "on_dump_info": func() { r.debugPrintRosterFor(account.session.GetConfig().Account) }, }) mn.ShowAll() mn.PopupAtMouseCursor(nil, nil, int(bt.Button()), bt.Time()) } func (r *roster) createAccountPopup(jid string, account *account, bt *gdk.EventButton) { builder := builderForDefinition("AccountPopupMenu") obj, _ := builder.GetObject("accountMenu") mn := obj.(*gtk.Menu) builder.ConnectSignals(map[string]interface{}{ "on_connect": func() { account.connect() }, "on_disconnect": func() { account.disconnect() }, "on_dump_info": func() { r.debugPrintRosterFor(account.session.GetConfig().Account) }, }) connx, _ := builder.GetObject("connectMenuItem") connect := connx.(*gtk.MenuItem) dconnx, _ := builder.GetObject("disconnectMenuItem") disconnect := dconnx.(*gtk.MenuItem) connect.SetSensitive(account.session.IsDisconnected()) disconnect.SetSensitive(account.session.IsConnected()) mn.ShowAll() mn.PopupAtMouseCursor(nil, nil, int(bt.Button()), bt.Time()) } func (r *roster) onButtonPress(view *gtk.TreeView, ev *gdk.Event) bool { bt := &gdk.EventButton{ev} if bt.Button() == 0x03 { jid, account, rowType, ok := r.getAccountAndJidFromEvent(bt) if ok { switch rowType { case "peer": r.createAccountPeerPopup(jid, account, bt) case "account": r.createAccountPopup(jid, account, bt) } } } return false } func (r *roster) onActivateBuddy(v *gtk.TreeView, path *gtk.TreePath) { selection, _ := v.GetSelection() defer selection.UnselectPath(path) iter, err := r.model.GetIter(path) if err != nil { return } jid := getFromModelIter(r.model, iter, indexJid) accountID := getFromModelIter(r.model, iter, indexAccountID) rowType := getFromModelIter(r.model, iter, indexRowType) if rowType != "peer" { r.isCollapsed[jid] = !r.isCollapsed[jid] r.redraw() return } account, ok := r.getAccount(accountID) if !ok { return } r.openConversationWindow(account, jid) } func (r *roster) openConversationWindow(account *account, to string) (*conversationWindow, error) { c, ok := account.getConversationWith(to) if !ok { textBuffer := r.ui.getTags().createTextBuffer() c = account.createConversationWindow(to, r.ui.displaySettings, textBuffer) r.ui.connectShortcutsChildWindow(c.win) r.ui.connectShortcutsConversationWindow(c) c.parentWin = r.ui.window } c.Show() return c, nil } func (r *roster) displayNameFor(account *account, from string) string { p, ok := r.ui.getPeer(account, from) if !ok { return from } return p.NameForPresentation() } func (r *roster) presenceUpdated(account *account, from, show, showStatus string, gone bool) { c, ok := account.getConversationWith(from) if !ok { return } doInUIThread(func() { c.appendStatus(r.displayNameFor(account, from), time.Now(), show, showStatus, gone) }) } func (r *roster) messageReceived(account *account, from string, timestamp time.Time, encrypted bool, message []byte) { doInUIThread(func() { conv, err := r.openConversationWindow(account, from) if err != nil { return } conv.appendMessage(r.displayNameFor(account, from), timestamp, encrypted, ui.StripHTML(message), false) }) } func (r *roster) update(account *account, entries *rosters.List) { r.ui.accountManager.Lock() defer r.ui.accountManager.Unlock() r.ui.accountManager.setContacts(account, entries) } func (r *roster) debugPrintRosterFor(nm string) { r.ui.accountManager.RLock() defer r.ui.accountManager.RUnlock() for account, rs := range r.ui.accountManager.getAllContacts() { if account.session.GetConfig().Is(nm) { rs.Iter(func(_ int, item *rosters.Peer) { fmt.Printf("-> %s\n", item.Dump()) }) } } fmt.Printf(" ************************************** \n") fmt.Println() } func isNominallyVisible(p *rosters.Peer) bool { return (p.Subscription != "none" && p.Subscription != "") || p.PendingSubscribeID != "" } func shouldDisplay(p *rosters.Peer, showOffline bool) bool { return isNominallyVisible(p) && (showOffline || p.Online) } func isAway(p *rosters.Peer) bool { switch p.Status { case "dnd", "xa", "away": return true } return false } func isOnline(p *rosters.Peer) bool { return p.PendingSubscribeID == "" && p.Online } func decideStatusFor(p *rosters.Peer) string { if p.PendingSubscribeID != ""
if !p.Online { return "offline" } switch p.Status { case "dnd": return "busy" case "xa": return "extended-away" case "away": return "away" } return "available" } func decideColorFor(p *rosters.Peer) string { if !p.Online { return "#aaaaaa" } return "#000000" } func createGroupDisplayName(parentName string, counter *counter, isExpanded bool) string { name := parentName if !isExpanded { name = fmt.Sprintf("[%s]", name) } return fmt.Sprintf("%s (%d/%d)", name, counter.online, counter.total) } func createTooltipFor(item *rosters.Peer) string { pname := html.EscapeString(item.NameForPresentation()) jid := html.EscapeString(item.Jid) if pname != jid { return fmt.Sprintf("%s (%s)", pname, jid) } return jid } func (r *roster) addItem(item *rosters.Peer, parentIter *gtk.TreeIter, indent string) { iter := r.model.Append(parentIter) setAll(r.model, iter, item.Jid, fmt.Sprintf("%s %s", indent, item.NameForPresentation()), item.BelongsTo, decideColorFor(item), "#ffffff", nil, createTooltipFor(item), ) r.model.SetValue(iter, indexRowType, "peer") r.model.SetValue(iter, indexStatusIcon, statusIcons[decideStatusFor(item)].getPixbuf()) } func (r *roster) redrawMerged() { showOffline := !r.ui.config.Display.ShowOnlyOnline r.ui.accountManager.RLock() defer r.ui.accountManager.RUnlock() r.toCollapse = nil grp := rosters.TopLevelGroup() for account, contacts := range r.ui.accountManager.getAllContacts() { contacts.AddTo(grp, account.session.GroupDelimiter) } accountCounter := &counter{} r.displayGroup(grp, nil, accountCounter, showOffline, "") r.view.ExpandAll() for _, path := range r.toCollapse { r.view.CollapseRow(path) } } type counter struct { total
{ return "unknown" }
conditional_block
models.py
. :) def create_database(db_settings: Dict[str, Any], database_name: str) -> None: with connect(**db_settings) as connection: cursor = connection.cursor() cursor.execute(f'CREATE DATABASE IF NOT EXISTS {database_name}') def remove_database(db_settings: Dict[str, Any], database_name: str) -> None: with connect(**db_settings) as connection: cursor = connection.cursor() cursor.execute(f'DROP DATABASE IF EXISTS {database_name}') class Keywords(str, Enum): ID = 'ID' Properties = 'properties' And = 'AND' From = 'FROM' InsertInto = 'INSERT INTO' Set = 'SET' Select = 'SELECT' Update = 'UPDATE' Values = 'VALUES' Where = 'WHERE' class EmptyQuery(Exception): """Raised when `.execute()` is called without prior select/filter.""" class SchemaNotFound(Exception): """Raised when the requested Schema is not found.""" class BaseModel: db_settings: Dict[str, Any] = DB_SETTINGS table_name: str schema: Type[ISchema] _table_created: Dict[str, bool] = {} BLANK_QUERY: str = '' def _create_table_if_not_exists(self) -> None: """Automatically create the provided schema table if it does not exist. For example:: { "title":"SportSchema", "type":"object", "properties":{ "Name":{ "title":"Name", "type":"string" }, "Slug":{ "title":"Slug", "type":"string" }, "Active":{ "title":"Active", "type":"boolean"
"Active" ] } Would result in the following create table query:: CREATE TABLE IF NOT EXISTS sports (ID INTEGER PRIMARY KEY AUTO_INCREMENT, Name VARCHAR(255), Slug VARCHAR(255), Active BOOLEAN) """ COLUMN_DEFINITIONS = 'definitions' COLUMN_TYPE = 'type' KEY_REF = '$ref' TYPE_LOOKUP = { 'string': 'VARCHAR(255)', 'integer': 'INTEGER', 'boolean': 'BOOLEAN', 'number': 'INTEGER', } def ref_lookup( property: Dict[str, Any], fields: Dict[str, Any] ) -> Dict[str, Any]: ref = property[KEY_REF] property_lookup_name = ref[ref.rfind('/') + 1 :] return fields[COLUMN_DEFINITIONS][property_lookup_name] field_queries = [] fields = json.loads(self.schema.schema_json()) del fields[Keywords.Properties.value][ Keywords.ID.value ] # Remove primary key field. It is handled with auto increment below. for property_name, property in fields[Keywords.Properties.value].items(): if KEY_REF in property: property = ref_lookup(property, fields) field_queries.append( f'{property_name} {TYPE_LOOKUP[property[COLUMN_TYPE]]}' ) table_columns = ', '.join(field_queries) with connect(**BaseModel.db_settings) as connection: cursor = connection.cursor() cursor.execute( f'CREATE TABLE IF NOT EXISTS {self.table_name} (ID INTEGER PRIMARY KEY AUTO_INCREMENT, {table_columns})' ) self._table_created[self.table_name] = True def __init__(self) -> None: if not self._table_created.get(self.table_name): self._create_table_if_not_exists() self._query: str = BaseModel.BLANK_QUERY self._last_method_called: Optional[function] = None def _clean_selected_fields(self, field_names: Tuple[str, ...]) -> Tuple[str, ...]: """Remove duplicates, e.g. 'ID' field requested twice. Maintains order. Using a set doesn't maintain order. """ list_field_names = [Keywords.ID.value] for field in field_names: if field in list_field_names: continue list_field_names.append(field) return tuple(list_field_names) def _append_to_query(self, statement: str) -> None: if self._query == BaseModel.BLANK_QUERY: fields = json.loads(self.schema.schema_json()) field_names = ', '.join(fields[Keywords.Properties.value].keys()) self._query = f'{Keywords.Select.value} {field_names} {Keywords.From.value} {self.table_name}' self._query += f' {statement}' def _map_results_to_schema( self, field_names: Iterable[str], results: List[Tuple[Any, ...]] ) -> List[ISchema]: schema_objects: List[ISchema] = [] for result in results: row_data_mapped_to_fields = dict(zip(field_names, result)) schema_objects.append(self.schema.construct(**row_data_mapped_to_fields)) return schema_objects def _fields_from_schema(self, schema: ISchema) -> List[str]: return cast(List[str], schema.dict().keys()) # KeysView[str] def _values_from_schema(self, schema: ISchema) -> List[Any]: return cast(List[Any], schema.dict().values()) # KeysView[Any] def select_fields(self, *field_names) -> List[ISchema]: field_names = self._clean_selected_fields(field_names) fields_formatted = ', '.join(field_names) query = f'{Keywords.Select.value} {fields_formatted} {Keywords.From.value} {self.table_name}' with connect(**BaseModel.db_settings) as connection: cursor = connection.cursor() cursor.execute(query) results = cursor.fetchall() return self._map_results_to_schema(field_names, results) def insert(self, schema: ISchema) -> ISchema: fields = self._fields_from_schema(schema) field_names = ', '.join(self._fields_from_schema(schema)) fields_placeholder = ('%s, ' * len(fields))[:-2] # Remove trailing , . values = tuple(self._values_from_schema(schema)) query = f'{Keywords.InsertInto.value} {self.table_name} ({field_names}) {Keywords.Values.value} ({fields_placeholder})' with connect(**BaseModel.db_settings) as connection: cursor = connection.cursor() cursor.execute(query, values) connection.commit() schema.set_id(cursor.lastrowid) return schema def update(self, schema: ISchema) -> ISchema: REMOVE_ID_FIELD_WITH_INDEX = 1 field_names = list(self._fields_from_schema(schema))[ REMOVE_ID_FIELD_WITH_INDEX: ] values = list(self._values_from_schema(schema))[REMOVE_ID_FIELD_WITH_INDEX:] fields_placeholder = ', '.join( [f'{field_name} = %s' for field_name in field_names] ) query = f"{Keywords.Update.value} {self.table_name} {Keywords.Set.value} {fields_placeholder} {Keywords.Where.value} {Keywords.ID.value} = '{schema.get_id()}'" with connect(**BaseModel.db_settings) as connection: cursor = connection.cursor() cursor.execute(query, values) connection.commit() return schema def select(self, *field_names) -> 'BaseModel': field_names = self._clean_selected_fields(field_names) fields_formatted = ', '.join(field_names) self._query = f'{Keywords.Select.value} {fields_formatted} {Keywords.From.value} {self.table_name}' self._last_method_called = self.select return self def filter(self, field_name: str, operator: Operators, value: Any) -> 'BaseModel': expression = ( Keywords.And.value if self._last_method_called == self.filter else Keywords.Where.value ) query = f"{expression} {field_name} {operator.value} '{value}'" self._append_to_query(query) self._last_method_called = self.filter return self def execute(self) -> List[ISchema]: if self._query == BaseModel.BLANK_QUERY: raise EmptyQuery() field_names = list( map( str.strip, self._query[ self._query.find(Keywords.Select.value) + len(Keywords.Select.value) : self._query.find(Keywords.From.value) ] .strip() .split(','), ) ) with connect(**BaseModel.db_settings) as connection: cursor = connection.cursor() cursor.execute(self._query) results = cursor.fetchall() self._query = BaseModel.BLANK_QUERY self._last_method_called = None return self._map_results_to_schema(field_names, results) def find(self, id: int) -> ISchema: self.filter(Keywords.ID.value, Operators.Equals, id) result = self.execute() if result: return result[0] raise SchemaNotFound(f'Not found, ID: {id}.') def get_query(self) -> str: return self._query class SportModel(BaseModel): schema = SportSchema table_name = 'sports' class EventModel(BaseModel): schema = EventSchema table_name = 'events' def update(self, schema: ISchema) -> ISchema: """When all the events of a sport are inactive, the sport becomes inactive """ schema = super().update(schema) schema = cast(EventSchema, schema) if not schema.Active and schema.Sport > 0: self.select('ID', 'Sport', 'Active').filter( 'Sport', Operators.Equals, schema.Sport, ).filter('Active', Operators
} }, "required":[ "Name", "Slug",
random_line_split
models.py
. :) def create_database(db_settings: Dict[str, Any], database_name: str) -> None: with connect(**db_settings) as connection: cursor = connection.cursor() cursor.execute(f'CREATE DATABASE IF NOT EXISTS {database_name}') def remove_database(db_settings: Dict[str, Any], database_name: str) -> None: with connect(**db_settings) as connection: cursor = connection.cursor() cursor.execute(f'DROP DATABASE IF EXISTS {database_name}') class Keywords(str, Enum): ID = 'ID' Properties = 'properties' And = 'AND' From = 'FROM' InsertInto = 'INSERT INTO' Set = 'SET' Select = 'SELECT' Update = 'UPDATE' Values = 'VALUES' Where = 'WHERE' class EmptyQuery(Exception): """Raised when `.execute()` is called without prior select/filter.""" class SchemaNotFound(Exception): """Raised when the requested Schema is not found.""" class BaseModel: db_settings: Dict[str, Any] = DB_SETTINGS table_name: str schema: Type[ISchema] _table_created: Dict[str, bool] = {} BLANK_QUERY: str = '' def _create_table_if_not_exists(self) -> None: """Automatically create the provided schema table if it does not exist. For example:: { "title":"SportSchema", "type":"object", "properties":{ "Name":{ "title":"Name", "type":"string" }, "Slug":{ "title":"Slug", "type":"string" }, "Active":{ "title":"Active", "type":"boolean" } }, "required":[ "Name", "Slug", "Active" ] } Would result in the following create table query:: CREATE TABLE IF NOT EXISTS sports (ID INTEGER PRIMARY KEY AUTO_INCREMENT, Name VARCHAR(255), Slug VARCHAR(255), Active BOOLEAN) """ COLUMN_DEFINITIONS = 'definitions' COLUMN_TYPE = 'type' KEY_REF = '$ref' TYPE_LOOKUP = { 'string': 'VARCHAR(255)', 'integer': 'INTEGER', 'boolean': 'BOOLEAN', 'number': 'INTEGER', } def ref_lookup( property: Dict[str, Any], fields: Dict[str, Any] ) -> Dict[str, Any]: ref = property[KEY_REF] property_lookup_name = ref[ref.rfind('/') + 1 :] return fields[COLUMN_DEFINITIONS][property_lookup_name] field_queries = [] fields = json.loads(self.schema.schema_json()) del fields[Keywords.Properties.value][ Keywords.ID.value ] # Remove primary key field. It is handled with auto increment below. for property_name, property in fields[Keywords.Properties.value].items(): if KEY_REF in property: property = ref_lookup(property, fields) field_queries.append( f'{property_name} {TYPE_LOOKUP[property[COLUMN_TYPE]]}' ) table_columns = ', '.join(field_queries) with connect(**BaseModel.db_settings) as connection: cursor = connection.cursor() cursor.execute( f'CREATE TABLE IF NOT EXISTS {self.table_name} (ID INTEGER PRIMARY KEY AUTO_INCREMENT, {table_columns})' ) self._table_created[self.table_name] = True def __init__(self) -> None: if not self._table_created.get(self.table_name): self._create_table_if_not_exists() self._query: str = BaseModel.BLANK_QUERY self._last_method_called: Optional[function] = None def _clean_selected_fields(self, field_names: Tuple[str, ...]) -> Tuple[str, ...]: """Remove duplicates, e.g. 'ID' field requested twice. Maintains order. Using a set doesn't maintain order. """ list_field_names = [Keywords.ID.value] for field in field_names: if field in list_field_names: continue list_field_names.append(field) return tuple(list_field_names) def _append_to_query(self, statement: str) -> None: if self._query == BaseModel.BLANK_QUERY: fields = json.loads(self.schema.schema_json()) field_names = ', '.join(fields[Keywords.Properties.value].keys()) self._query = f'{Keywords.Select.value} {field_names} {Keywords.From.value} {self.table_name}' self._query += f' {statement}' def _map_results_to_schema( self, field_names: Iterable[str], results: List[Tuple[Any, ...]] ) -> List[ISchema]:
def _fields_from_schema(self, schema: ISchema) -> List[str]: return cast(List[str], schema.dict().keys()) # KeysView[str] def _values_from_schema(self, schema: ISchema) -> List[Any]: return cast(List[Any], schema.dict().values()) # KeysView[Any] def select_fields(self, *field_names) -> List[ISchema]: field_names = self._clean_selected_fields(field_names) fields_formatted = ', '.join(field_names) query = f'{Keywords.Select.value} {fields_formatted} {Keywords.From.value} {self.table_name}' with connect(**BaseModel.db_settings) as connection: cursor = connection.cursor() cursor.execute(query) results = cursor.fetchall() return self._map_results_to_schema(field_names, results) def insert(self, schema: ISchema) -> ISchema: fields = self._fields_from_schema(schema) field_names = ', '.join(self._fields_from_schema(schema)) fields_placeholder = ('%s, ' * len(fields))[:-2] # Remove trailing , . values = tuple(self._values_from_schema(schema)) query = f'{Keywords.InsertInto.value} {self.table_name} ({field_names}) {Keywords.Values.value} ({fields_placeholder})' with connect(**BaseModel.db_settings) as connection: cursor = connection.cursor() cursor.execute(query, values) connection.commit() schema.set_id(cursor.lastrowid) return schema def update(self, schema: ISchema) -> ISchema: REMOVE_ID_FIELD_WITH_INDEX = 1 field_names = list(self._fields_from_schema(schema))[ REMOVE_ID_FIELD_WITH_INDEX: ] values = list(self._values_from_schema(schema))[REMOVE_ID_FIELD_WITH_INDEX:] fields_placeholder = ', '.join( [f'{field_name} = %s' for field_name in field_names] ) query = f"{Keywords.Update.value} {self.table_name} {Keywords.Set.value} {fields_placeholder} {Keywords.Where.value} {Keywords.ID.value} = '{schema.get_id()}'" with connect(**BaseModel.db_settings) as connection: cursor = connection.cursor() cursor.execute(query, values) connection.commit() return schema def select(self, *field_names) -> 'BaseModel': field_names = self._clean_selected_fields(field_names) fields_formatted = ', '.join(field_names) self._query = f'{Keywords.Select.value} {fields_formatted} {Keywords.From.value} {self.table_name}' self._last_method_called = self.select return self def filter(self, field_name: str, operator: Operators, value: Any) -> 'BaseModel': expression = ( Keywords.And.value if self._last_method_called == self.filter else Keywords.Where.value ) query = f"{expression} {field_name} {operator.value} '{value}'" self._append_to_query(query) self._last_method_called = self.filter return self def execute(self) -> List[ISchema]: if self._query == BaseModel.BLANK_QUERY: raise EmptyQuery() field_names = list( map( str.strip, self._query[ self._query.find(Keywords.Select.value) + len(Keywords.Select.value) : self._query.find(Keywords.From.value) ] .strip() .split(','), ) ) with connect(**BaseModel.db_settings) as connection: cursor = connection.cursor() cursor.execute(self._query) results = cursor.fetchall() self._query = BaseModel.BLANK_QUERY self._last_method_called = None return self._map_results_to_schema(field_names, results) def find(self, id: int) -> ISchema: self.filter(Keywords.ID.value, Operators.Equals, id) result = self.execute() if result: return result[0] raise SchemaNotFound(f'Not found, ID: {id}.') def get_query(self) -> str: return self._query class SportModel(BaseModel): schema = SportSchema table_name = 'sports' class EventModel(BaseModel): schema = EventSchema table_name = 'events' def update(self, schema: ISchema) -> ISchema: """When all the events of a sport are inactive, the sport becomes inactive """ schema = super().update(schema) schema = cast(EventSchema, schema) if not schema.Active and schema.Sport > 0: self.select('ID', 'Sport', 'Active').filter( 'Sport', Operators.Equals, schema.Sport, ).filter('Active', Operators
schema_objects: List[ISchema] = [] for result in results: row_data_mapped_to_fields = dict(zip(field_names, result)) schema_objects.append(self.schema.construct(**row_data_mapped_to_fields)) return schema_objects
identifier_body
models.py
:) def create_database(db_settings: Dict[str, Any], database_name: str) -> None: with connect(**db_settings) as connection: cursor = connection.cursor() cursor.execute(f'CREATE DATABASE IF NOT EXISTS {database_name}') def remove_database(db_settings: Dict[str, Any], database_name: str) -> None: with connect(**db_settings) as connection: cursor = connection.cursor() cursor.execute(f'DROP DATABASE IF EXISTS {database_name}') class Keywords(str, Enum): ID = 'ID' Properties = 'properties' And = 'AND' From = 'FROM' InsertInto = 'INSERT INTO' Set = 'SET' Select = 'SELECT' Update = 'UPDATE' Values = 'VALUES' Where = 'WHERE' class EmptyQuery(Exception): """Raised when `.execute()` is called without prior select/filter.""" class SchemaNotFound(Exception): """Raised when the requested Schema is not found.""" class BaseModel: db_settings: Dict[str, Any] = DB_SETTINGS table_name: str schema: Type[ISchema] _table_created: Dict[str, bool] = {} BLANK_QUERY: str = '' def _create_table_if_not_exists(self) -> None: """Automatically create the provided schema table if it does not exist. For example:: { "title":"SportSchema", "type":"object", "properties":{ "Name":{ "title":"Name", "type":"string" }, "Slug":{ "title":"Slug", "type":"string" }, "Active":{ "title":"Active", "type":"boolean" } }, "required":[ "Name", "Slug", "Active" ] } Would result in the following create table query:: CREATE TABLE IF NOT EXISTS sports (ID INTEGER PRIMARY KEY AUTO_INCREMENT, Name VARCHAR(255), Slug VARCHAR(255), Active BOOLEAN) """ COLUMN_DEFINITIONS = 'definitions' COLUMN_TYPE = 'type' KEY_REF = '$ref' TYPE_LOOKUP = { 'string': 'VARCHAR(255)', 'integer': 'INTEGER', 'boolean': 'BOOLEAN', 'number': 'INTEGER', } def ref_lookup( property: Dict[str, Any], fields: Dict[str, Any] ) -> Dict[str, Any]: ref = property[KEY_REF] property_lookup_name = ref[ref.rfind('/') + 1 :] return fields[COLUMN_DEFINITIONS][property_lookup_name] field_queries = [] fields = json.loads(self.schema.schema_json()) del fields[Keywords.Properties.value][ Keywords.ID.value ] # Remove primary key field. It is handled with auto increment below. for property_name, property in fields[Keywords.Properties.value].items():
table_columns = ', '.join(field_queries) with connect(**BaseModel.db_settings) as connection: cursor = connection.cursor() cursor.execute( f'CREATE TABLE IF NOT EXISTS {self.table_name} (ID INTEGER PRIMARY KEY AUTO_INCREMENT, {table_columns})' ) self._table_created[self.table_name] = True def __init__(self) -> None: if not self._table_created.get(self.table_name): self._create_table_if_not_exists() self._query: str = BaseModel.BLANK_QUERY self._last_method_called: Optional[function] = None def _clean_selected_fields(self, field_names: Tuple[str, ...]) -> Tuple[str, ...]: """Remove duplicates, e.g. 'ID' field requested twice. Maintains order. Using a set doesn't maintain order. """ list_field_names = [Keywords.ID.value] for field in field_names: if field in list_field_names: continue list_field_names.append(field) return tuple(list_field_names) def _append_to_query(self, statement: str) -> None: if self._query == BaseModel.BLANK_QUERY: fields = json.loads(self.schema.schema_json()) field_names = ', '.join(fields[Keywords.Properties.value].keys()) self._query = f'{Keywords.Select.value} {field_names} {Keywords.From.value} {self.table_name}' self._query += f' {statement}' def _map_results_to_schema( self, field_names: Iterable[str], results: List[Tuple[Any, ...]] ) -> List[ISchema]: schema_objects: List[ISchema] = [] for result in results: row_data_mapped_to_fields = dict(zip(field_names, result)) schema_objects.append(self.schema.construct(**row_data_mapped_to_fields)) return schema_objects def _fields_from_schema(self, schema: ISchema) -> List[str]: return cast(List[str], schema.dict().keys()) # KeysView[str] def _values_from_schema(self, schema: ISchema) -> List[Any]: return cast(List[Any], schema.dict().values()) # KeysView[Any] def select_fields(self, *field_names) -> List[ISchema]: field_names = self._clean_selected_fields(field_names) fields_formatted = ', '.join(field_names) query = f'{Keywords.Select.value} {fields_formatted} {Keywords.From.value} {self.table_name}' with connect(**BaseModel.db_settings) as connection: cursor = connection.cursor() cursor.execute(query) results = cursor.fetchall() return self._map_results_to_schema(field_names, results) def insert(self, schema: ISchema) -> ISchema: fields = self._fields_from_schema(schema) field_names = ', '.join(self._fields_from_schema(schema)) fields_placeholder = ('%s, ' * len(fields))[:-2] # Remove trailing , . values = tuple(self._values_from_schema(schema)) query = f'{Keywords.InsertInto.value} {self.table_name} ({field_names}) {Keywords.Values.value} ({fields_placeholder})' with connect(**BaseModel.db_settings) as connection: cursor = connection.cursor() cursor.execute(query, values) connection.commit() schema.set_id(cursor.lastrowid) return schema def update(self, schema: ISchema) -> ISchema: REMOVE_ID_FIELD_WITH_INDEX = 1 field_names = list(self._fields_from_schema(schema))[ REMOVE_ID_FIELD_WITH_INDEX: ] values = list(self._values_from_schema(schema))[REMOVE_ID_FIELD_WITH_INDEX:] fields_placeholder = ', '.join( [f'{field_name} = %s' for field_name in field_names] ) query = f"{Keywords.Update.value} {self.table_name} {Keywords.Set.value} {fields_placeholder} {Keywords.Where.value} {Keywords.ID.value} = '{schema.get_id()}'" with connect(**BaseModel.db_settings) as connection: cursor = connection.cursor() cursor.execute(query, values) connection.commit() return schema def select(self, *field_names) -> 'BaseModel': field_names = self._clean_selected_fields(field_names) fields_formatted = ', '.join(field_names) self._query = f'{Keywords.Select.value} {fields_formatted} {Keywords.From.value} {self.table_name}' self._last_method_called = self.select return self def filter(self, field_name: str, operator: Operators, value: Any) -> 'BaseModel': expression = ( Keywords.And.value if self._last_method_called == self.filter else Keywords.Where.value ) query = f"{expression} {field_name} {operator.value} '{value}'" self._append_to_query(query) self._last_method_called = self.filter return self def execute(self) -> List[ISchema]: if self._query == BaseModel.BLANK_QUERY: raise EmptyQuery() field_names = list( map( str.strip, self._query[ self._query.find(Keywords.Select.value) + len(Keywords.Select.value) : self._query.find(Keywords.From.value) ] .strip() .split(','), ) ) with connect(**BaseModel.db_settings) as connection: cursor = connection.cursor() cursor.execute(self._query) results = cursor.fetchall() self._query = BaseModel.BLANK_QUERY self._last_method_called = None return self._map_results_to_schema(field_names, results) def find(self, id: int) -> ISchema: self.filter(Keywords.ID.value, Operators.Equals, id) result = self.execute() if result: return result[0] raise SchemaNotFound(f'Not found, ID: {id}.') def get_query(self) -> str: return self._query class SportModel(BaseModel): schema = SportSchema table_name = 'sports' class EventModel(BaseModel): schema = EventSchema table_name = 'events' def update(self, schema: ISchema) -> ISchema: """When all the events of a sport are inactive, the sport becomes inactive """ schema = super().update(schema) schema = cast(EventSchema, schema) if not schema.Active and schema.Sport > 0: self.select('ID', 'Sport', 'Active').filter( 'Sport', Operators.Equals, schema.Sport, ).filter('Active', Operators
if KEY_REF in property: property = ref_lookup(property, fields) field_queries.append( f'{property_name} {TYPE_LOOKUP[property[COLUMN_TYPE]]}' )
conditional_block
models.py
. :) def create_database(db_settings: Dict[str, Any], database_name: str) -> None: with connect(**db_settings) as connection: cursor = connection.cursor() cursor.execute(f'CREATE DATABASE IF NOT EXISTS {database_name}') def remove_database(db_settings: Dict[str, Any], database_name: str) -> None: with connect(**db_settings) as connection: cursor = connection.cursor() cursor.execute(f'DROP DATABASE IF EXISTS {database_name}') class Keywords(str, Enum): ID = 'ID' Properties = 'properties' And = 'AND' From = 'FROM' InsertInto = 'INSERT INTO' Set = 'SET' Select = 'SELECT' Update = 'UPDATE' Values = 'VALUES' Where = 'WHERE' class EmptyQuery(Exception): """Raised when `.execute()` is called without prior select/filter.""" class SchemaNotFound(Exception): """Raised when the requested Schema is not found.""" class BaseModel: db_settings: Dict[str, Any] = DB_SETTINGS table_name: str schema: Type[ISchema] _table_created: Dict[str, bool] = {} BLANK_QUERY: str = '' def _create_table_if_not_exists(self) -> None: """Automatically create the provided schema table if it does not exist. For example:: { "title":"SportSchema", "type":"object", "properties":{ "Name":{ "title":"Name", "type":"string" }, "Slug":{ "title":"Slug", "type":"string" }, "Active":{ "title":"Active", "type":"boolean" } }, "required":[ "Name", "Slug", "Active" ] } Would result in the following create table query:: CREATE TABLE IF NOT EXISTS sports (ID INTEGER PRIMARY KEY AUTO_INCREMENT, Name VARCHAR(255), Slug VARCHAR(255), Active BOOLEAN) """ COLUMN_DEFINITIONS = 'definitions' COLUMN_TYPE = 'type' KEY_REF = '$ref' TYPE_LOOKUP = { 'string': 'VARCHAR(255)', 'integer': 'INTEGER', 'boolean': 'BOOLEAN', 'number': 'INTEGER', } def ref_lookup( property: Dict[str, Any], fields: Dict[str, Any] ) -> Dict[str, Any]: ref = property[KEY_REF] property_lookup_name = ref[ref.rfind('/') + 1 :] return fields[COLUMN_DEFINITIONS][property_lookup_name] field_queries = [] fields = json.loads(self.schema.schema_json()) del fields[Keywords.Properties.value][ Keywords.ID.value ] # Remove primary key field. It is handled with auto increment below. for property_name, property in fields[Keywords.Properties.value].items(): if KEY_REF in property: property = ref_lookup(property, fields) field_queries.append( f'{property_name} {TYPE_LOOKUP[property[COLUMN_TYPE]]}' ) table_columns = ', '.join(field_queries) with connect(**BaseModel.db_settings) as connection: cursor = connection.cursor() cursor.execute( f'CREATE TABLE IF NOT EXISTS {self.table_name} (ID INTEGER PRIMARY KEY AUTO_INCREMENT, {table_columns})' ) self._table_created[self.table_name] = True def __init__(self) -> None: if not self._table_created.get(self.table_name): self._create_table_if_not_exists() self._query: str = BaseModel.BLANK_QUERY self._last_method_called: Optional[function] = None def _clean_selected_fields(self, field_names: Tuple[str, ...]) -> Tuple[str, ...]: """Remove duplicates, e.g. 'ID' field requested twice. Maintains order. Using a set doesn't maintain order. """ list_field_names = [Keywords.ID.value] for field in field_names: if field in list_field_names: continue list_field_names.append(field) return tuple(list_field_names) def _append_to_query(self, statement: str) -> None: if self._query == BaseModel.BLANK_QUERY: fields = json.loads(self.schema.schema_json()) field_names = ', '.join(fields[Keywords.Properties.value].keys()) self._query = f'{Keywords.Select.value} {field_names} {Keywords.From.value} {self.table_name}' self._query += f' {statement}' def _map_results_to_schema( self, field_names: Iterable[str], results: List[Tuple[Any, ...]] ) -> List[ISchema]: schema_objects: List[ISchema] = [] for result in results: row_data_mapped_to_fields = dict(zip(field_names, result)) schema_objects.append(self.schema.construct(**row_data_mapped_to_fields)) return schema_objects def _fields_from_schema(self, schema: ISchema) -> List[str]: return cast(List[str], schema.dict().keys()) # KeysView[str] def
(self, schema: ISchema) -> List[Any]: return cast(List[Any], schema.dict().values()) # KeysView[Any] def select_fields(self, *field_names) -> List[ISchema]: field_names = self._clean_selected_fields(field_names) fields_formatted = ', '.join(field_names) query = f'{Keywords.Select.value} {fields_formatted} {Keywords.From.value} {self.table_name}' with connect(**BaseModel.db_settings) as connection: cursor = connection.cursor() cursor.execute(query) results = cursor.fetchall() return self._map_results_to_schema(field_names, results) def insert(self, schema: ISchema) -> ISchema: fields = self._fields_from_schema(schema) field_names = ', '.join(self._fields_from_schema(schema)) fields_placeholder = ('%s, ' * len(fields))[:-2] # Remove trailing , . values = tuple(self._values_from_schema(schema)) query = f'{Keywords.InsertInto.value} {self.table_name} ({field_names}) {Keywords.Values.value} ({fields_placeholder})' with connect(**BaseModel.db_settings) as connection: cursor = connection.cursor() cursor.execute(query, values) connection.commit() schema.set_id(cursor.lastrowid) return schema def update(self, schema: ISchema) -> ISchema: REMOVE_ID_FIELD_WITH_INDEX = 1 field_names = list(self._fields_from_schema(schema))[ REMOVE_ID_FIELD_WITH_INDEX: ] values = list(self._values_from_schema(schema))[REMOVE_ID_FIELD_WITH_INDEX:] fields_placeholder = ', '.join( [f'{field_name} = %s' for field_name in field_names] ) query = f"{Keywords.Update.value} {self.table_name} {Keywords.Set.value} {fields_placeholder} {Keywords.Where.value} {Keywords.ID.value} = '{schema.get_id()}'" with connect(**BaseModel.db_settings) as connection: cursor = connection.cursor() cursor.execute(query, values) connection.commit() return schema def select(self, *field_names) -> 'BaseModel': field_names = self._clean_selected_fields(field_names) fields_formatted = ', '.join(field_names) self._query = f'{Keywords.Select.value} {fields_formatted} {Keywords.From.value} {self.table_name}' self._last_method_called = self.select return self def filter(self, field_name: str, operator: Operators, value: Any) -> 'BaseModel': expression = ( Keywords.And.value if self._last_method_called == self.filter else Keywords.Where.value ) query = f"{expression} {field_name} {operator.value} '{value}'" self._append_to_query(query) self._last_method_called = self.filter return self def execute(self) -> List[ISchema]: if self._query == BaseModel.BLANK_QUERY: raise EmptyQuery() field_names = list( map( str.strip, self._query[ self._query.find(Keywords.Select.value) + len(Keywords.Select.value) : self._query.find(Keywords.From.value) ] .strip() .split(','), ) ) with connect(**BaseModel.db_settings) as connection: cursor = connection.cursor() cursor.execute(self._query) results = cursor.fetchall() self._query = BaseModel.BLANK_QUERY self._last_method_called = None return self._map_results_to_schema(field_names, results) def find(self, id: int) -> ISchema: self.filter(Keywords.ID.value, Operators.Equals, id) result = self.execute() if result: return result[0] raise SchemaNotFound(f'Not found, ID: {id}.') def get_query(self) -> str: return self._query class SportModel(BaseModel): schema = SportSchema table_name = 'sports' class EventModel(BaseModel): schema = EventSchema table_name = 'events' def update(self, schema: ISchema) -> ISchema: """When all the events of a sport are inactive, the sport becomes inactive """ schema = super().update(schema) schema = cast(EventSchema, schema) if not schema.Active and schema.Sport > 0: self.select('ID', 'Sport', 'Active').filter( 'Sport', Operators.Equals, schema.Sport, ).filter('Active', Operators
_values_from_schema
identifier_name
wgl.rs
wgl") .encode_wide() .chain(Some(0).into_iter()) .collect::<Vec<_>>(); class.cbSize = mem::size_of::<WNDCLASSEXW>() as UINT; class.lpszClassName = class_name.as_ptr(); class.lpfnWndProc = Some(DefWindowProcW); RegisterClassExW(&class); let hwnd = CreateWindowExW( 0, class_name.as_ptr(), std::ptr::null(), 0, CW_USEDEFAULT, CW_USEDEFAULT, CW_USEDEFAULT, CW_USEDEFAULT, std::ptr::null_mut(), std::ptr::null_mut(), instance, std::ptr::null_mut(), ); let hdc = GetDC(hwnd); let desc = PIXELFORMATDESCRIPTOR { nSize: std::mem::size_of::<PIXELFORMATDESCRIPTOR>() as u16, nVersion: 1, dwFlags: PFD_SUPPORT_OPENGL, iPixelType: PFD_TYPE_RGBA, cColorBits: 8, cRedBits: 0, cRedShift: 0, cGreenBits: 0, cGreenShift: 0, cBlueBits: 0, cBlueShift: 0, cAlphaBits: 8, cAlphaShift: 0, cAccumBits: 0, cAccumRedBits: 0, cAccumGreenBits: 0, cAccumBlueBits: 0, cAccumAlphaBits: 0, cDepthBits: 0, cStencilBits: 0, cAuxBuffers: 0, iLayerType: PFD_MAIN_PLANE, bReserved: 0, dwLayerMask: 0, dwVisibleMask: 0, dwDamageMask: 0, }; let format_id = ChoosePixelFormat(hdc, &desc); SetPixelFormat(hdc, format_id, &desc); let hglrc = wglCreateContext(hdc); println!("{:?}", (hwnd, hdc, format_id, hglrc)); wglMakeCurrent(hdc, hglrc); let name = OsStr::new("opengl32.dll") .encode_wide() .chain(Some(0).into_iter()) .collect::<Vec<_>>(); let lib = LoadLibraryW(name.as_ptr()); let wgl = wgl_ext_sys::Wgl::load_with(|sym| { let sym = CString::new(sym.as_bytes()).unwrap(); let addr = wgl_sys::GetProcAddress(sym.as_ptr()) as *const (); if !addr.is_null() { addr as *const _ } else { GetProcAddress(lib, sym.as_ptr()) as *const _ } }); Entry { hwnd, hdc: hdc as _, wgl, lib, } } } } impl Drop for Entry { fn drop(&mut self) { unsafe { DestroyWindow(self.hwnd); } } } lazy_static! { // Entry function pointers pub(crate) static ref WGL_ENTRY: Entry = Entry::new(); } pub struct Instance { pub(crate) ctxt: DeviceContext, } impl Instance { pub fn create(_name: &str, version: u32) -> Result<Self, hal::UnsupportedBackend> { unsafe { let glrc = WGL_ENTRY.wgl.CreateContextAttribsARB( WGL_ENTRY.hdc as *const _, ptr::null(), ptr::null(), ) as HGLRC; wglMakeCurrent(WGL_ENTRY.hdc as *mut _, glrc); Ok(Instance { ctxt: DeviceContext { ctxt: Context { glrc }, hdc: WGL_ENTRY.hdc, }, }) } } #[cfg(windows)] pub fn create_surface_from_hwnd(&self, hwnd: *mut c_void) -> Surface { Surface { hwnd: hwnd as *mut _, swapchain: None, renderbuffer: None, } } #[cfg(feature = "winit")] pub fn create_surface(&self, window: &winit::window::Window) -> Surface { use winit::platform::windows::WindowExtWindows; let hwnd = window.hwnd(); self.create_surface_from_hwnd(hwnd as *mut _) } } impl hal::Instance for Instance { type Backend = Backend; fn enumerate_adapters(&self) -> Vec<Adapter<Backend>> { let gl_container = GlContainer::from_fn_proc(|s| unsafe { let sym = CString::new(s.as_bytes()).unwrap(); let addr = wgl_sys::GetProcAddress(sym.as_ptr()) as *const (); if !addr.is_null() { addr as *const _ } else { GetProcAddress(WGL_ENTRY.lib, sym.as_ptr()) as *const _ } }); let adapter = PhysicalDevice::new_adapter(self.ctxt, gl_container); vec![adapter] } } #[derive(Debug)] pub struct Surface { pub(crate) hwnd: HWND, pub(crate) swapchain: Option<Swapchain>, renderbuffer: Option<native::Renderbuffer>, } // TODO: high -msiglreith unsafe impl Send for Surface {} unsafe impl Sync for Surface {} impl window::Surface<Backend> for Surface { fn compatibility( &self, physical_device: &PhysicalDevice, ) -> ( window::SurfaceCapabilities, Option<Vec<f::Format>>, Vec<window::PresentMode>, ) { let extent = unsafe { let mut rect: RECT = mem::zeroed(); GetClientRect(self.hwnd, &mut rect); window::Extent2D { width: (rect.right - rect.left) as _, height: (rect.bottom - rect.top) as _, } }; let caps = window::SurfaceCapabilities { image_count: 2 ..= 2, current_extent: Some(extent), extents: extent ..= extent, max_image_layers: 1, usage: image::Usage::COLOR_ATTACHMENT | image::Usage::TRANSFER_SRC, composite_alpha: window::CompositeAlpha::OPAQUE, //TODO }; let present_modes = vec![ window::PresentMode::Fifo, //TODO ]; ( caps, Some(vec![f::Format::Rgba8Srgb, f::Format::Bgra8Srgb]), present_modes, ) } fn supports_queue_family(&self, _queue_family: &QueueFamily) -> bool { true } } impl window::PresentationSurface<Backend> for Surface { type SwapchainImage = native::ImageView; unsafe fn configure_swapchain( &mut self, device: &Device, config: window::SwapchainConfig, ) -> Result<(), window::CreationError> { let gl = &device.share.context; let context = match self.swapchain.take() { Some(old) => { for fbo in old.fbos { gl.delete_framebuffer(fbo); } old.context } None => PresentContext::new(self, &device.share.instance_context), }; context.make_current(); if self.renderbuffer.is_none() { self.renderbuffer = Some(gl.create_renderbuffer().unwrap()); } let desc = conv::describe_format(config.format).unwrap(); gl.bind_renderbuffer(glow::RENDERBUFFER, self.renderbuffer); gl.renderbuffer_storage( glow::RENDERBUFFER, desc.tex_internal, config.extent.width as i32, config.extent.height as i32, ); let fbo = gl.create_framebuffer().unwrap(); gl.bind_framebuffer(glow::READ_FRAMEBUFFER, Some(fbo)); gl.framebuffer_renderbuffer( glow::READ_FRAMEBUFFER, glow::COLOR_ATTACHMENT0, glow::RENDERBUFFER, self.renderbuffer,
extent: config.extent, fbos: iter::once(fbo).collect(), }); Ok(()) } unsafe fn unconfigure_swapchain(&mut self, device: &Device) { let gl = &device.share.context; if let Some(old) = self.swapchain.take() { for fbo in old.fbos { gl.delete_framebuffer(fbo); } } if let Some(rbo) = self.renderbuffer.take() { gl.delete_renderbuffer(rbo); } } unsafe fn acquire_image( &mut self, _timeout_ns: u64, ) -> Result<(Self::SwapchainImage, Option<window::Suboptimal>), window::AcquireError> { let image = native::ImageView::Renderbuffer(self.renderbuffer.unwrap()); Ok((image, None)) } } #[derive(Debug)] pub struct Swapchain { pub(crate) fbos: ArrayVec<[native::RawFrameBuffer; 3]>, pub(crate) context: PresentContext, pub(crate) extent: window::Extent2D, } impl Swapchain { pub(crate) fn make_current(&self) { self.context.make_current(); } pub(crate) fn swap_buffers(&self) { self.context.swap_buffers(); } } impl window::Swapchain<Backend> for Swapchain { unsafe fn acquire_image( &mut self, _timeout_ns: u64, _semaphore: Option
); self.swapchain = Some(Swapchain { context,
random_line_split
wgl.rs
wgl") .encode_wide() .chain(Some(0).into_iter()) .collect::<Vec<_>>(); class.cbSize = mem::size_of::<WNDCLASSEXW>() as UINT; class.lpszClassName = class_name.as_ptr(); class.lpfnWndProc = Some(DefWindowProcW); RegisterClassExW(&class); let hwnd = CreateWindowExW( 0, class_name.as_ptr(), std::ptr::null(), 0, CW_USEDEFAULT, CW_USEDEFAULT, CW_USEDEFAULT, CW_USEDEFAULT, std::ptr::null_mut(), std::ptr::null_mut(), instance, std::ptr::null_mut(), ); let hdc = GetDC(hwnd); let desc = PIXELFORMATDESCRIPTOR { nSize: std::mem::size_of::<PIXELFORMATDESCRIPTOR>() as u16, nVersion: 1, dwFlags: PFD_SUPPORT_OPENGL, iPixelType: PFD_TYPE_RGBA, cColorBits: 8, cRedBits: 0, cRedShift: 0, cGreenBits: 0, cGreenShift: 0, cBlueBits: 0, cBlueShift: 0, cAlphaBits: 8, cAlphaShift: 0, cAccumBits: 0, cAccumRedBits: 0, cAccumGreenBits: 0, cAccumBlueBits: 0, cAccumAlphaBits: 0, cDepthBits: 0, cStencilBits: 0, cAuxBuffers: 0, iLayerType: PFD_MAIN_PLANE, bReserved: 0, dwLayerMask: 0, dwVisibleMask: 0, dwDamageMask: 0, }; let format_id = ChoosePixelFormat(hdc, &desc); SetPixelFormat(hdc, format_id, &desc); let hglrc = wglCreateContext(hdc); println!("{:?}", (hwnd, hdc, format_id, hglrc)); wglMakeCurrent(hdc, hglrc); let name = OsStr::new("opengl32.dll") .encode_wide() .chain(Some(0).into_iter()) .collect::<Vec<_>>(); let lib = LoadLibraryW(name.as_ptr()); let wgl = wgl_ext_sys::Wgl::load_with(|sym| { let sym = CString::new(sym.as_bytes()).unwrap(); let addr = wgl_sys::GetProcAddress(sym.as_ptr()) as *const (); if !addr.is_null()
else { GetProcAddress(lib, sym.as_ptr()) as *const _ } }); Entry { hwnd, hdc: hdc as _, wgl, lib, } } } } impl Drop for Entry { fn drop(&mut self) { unsafe { DestroyWindow(self.hwnd); } } } lazy_static! { // Entry function pointers pub(crate) static ref WGL_ENTRY: Entry = Entry::new(); } pub struct Instance { pub(crate) ctxt: DeviceContext, } impl Instance { pub fn create(_name: &str, version: u32) -> Result<Self, hal::UnsupportedBackend> { unsafe { let glrc = WGL_ENTRY.wgl.CreateContextAttribsARB( WGL_ENTRY.hdc as *const _, ptr::null(), ptr::null(), ) as HGLRC; wglMakeCurrent(WGL_ENTRY.hdc as *mut _, glrc); Ok(Instance { ctxt: DeviceContext { ctxt: Context { glrc }, hdc: WGL_ENTRY.hdc, }, }) } } #[cfg(windows)] pub fn create_surface_from_hwnd(&self, hwnd: *mut c_void) -> Surface { Surface { hwnd: hwnd as *mut _, swapchain: None, renderbuffer: None, } } #[cfg(feature = "winit")] pub fn create_surface(&self, window: &winit::window::Window) -> Surface { use winit::platform::windows::WindowExtWindows; let hwnd = window.hwnd(); self.create_surface_from_hwnd(hwnd as *mut _) } } impl hal::Instance for Instance { type Backend = Backend; fn enumerate_adapters(&self) -> Vec<Adapter<Backend>> { let gl_container = GlContainer::from_fn_proc(|s| unsafe { let sym = CString::new(s.as_bytes()).unwrap(); let addr = wgl_sys::GetProcAddress(sym.as_ptr()) as *const (); if !addr.is_null() { addr as *const _ } else { GetProcAddress(WGL_ENTRY.lib, sym.as_ptr()) as *const _ } }); let adapter = PhysicalDevice::new_adapter(self.ctxt, gl_container); vec![adapter] } } #[derive(Debug)] pub struct Surface { pub(crate) hwnd: HWND, pub(crate) swapchain: Option<Swapchain>, renderbuffer: Option<native::Renderbuffer>, } // TODO: high -msiglreith unsafe impl Send for Surface {} unsafe impl Sync for Surface {} impl window::Surface<Backend> for Surface { fn compatibility( &self, physical_device: &PhysicalDevice, ) -> ( window::SurfaceCapabilities, Option<Vec<f::Format>>, Vec<window::PresentMode>, ) { let extent = unsafe { let mut rect: RECT = mem::zeroed(); GetClientRect(self.hwnd, &mut rect); window::Extent2D { width: (rect.right - rect.left) as _, height: (rect.bottom - rect.top) as _, } }; let caps = window::SurfaceCapabilities { image_count: 2 ..= 2, current_extent: Some(extent), extents: extent ..= extent, max_image_layers: 1, usage: image::Usage::COLOR_ATTACHMENT | image::Usage::TRANSFER_SRC, composite_alpha: window::CompositeAlpha::OPAQUE, //TODO }; let present_modes = vec![ window::PresentMode::Fifo, //TODO ]; ( caps, Some(vec![f::Format::Rgba8Srgb, f::Format::Bgra8Srgb]), present_modes, ) } fn supports_queue_family(&self, _queue_family: &QueueFamily) -> bool { true } } impl window::PresentationSurface<Backend> for Surface { type SwapchainImage = native::ImageView; unsafe fn configure_swapchain( &mut self, device: &Device, config: window::SwapchainConfig, ) -> Result<(), window::CreationError> { let gl = &device.share.context; let context = match self.swapchain.take() { Some(old) => { for fbo in old.fbos { gl.delete_framebuffer(fbo); } old.context } None => PresentContext::new(self, &device.share.instance_context), }; context.make_current(); if self.renderbuffer.is_none() { self.renderbuffer = Some(gl.create_renderbuffer().unwrap()); } let desc = conv::describe_format(config.format).unwrap(); gl.bind_renderbuffer(glow::RENDERBUFFER, self.renderbuffer); gl.renderbuffer_storage( glow::RENDERBUFFER, desc.tex_internal, config.extent.width as i32, config.extent.height as i32, ); let fbo = gl.create_framebuffer().unwrap(); gl.bind_framebuffer(glow::READ_FRAMEBUFFER, Some(fbo)); gl.framebuffer_renderbuffer( glow::READ_FRAMEBUFFER, glow::COLOR_ATTACHMENT0, glow::RENDERBUFFER, self.renderbuffer, ); self.swapchain = Some(Swapchain { context, extent: config.extent, fbos: iter::once(fbo).collect(), }); Ok(()) } unsafe fn unconfigure_swapchain(&mut self, device: &Device) { let gl = &device.share.context; if let Some(old) = self.swapchain.take() { for fbo in old.fbos { gl.delete_framebuffer(fbo); } } if let Some(rbo) = self.renderbuffer.take() { gl.delete_renderbuffer(rbo); } } unsafe fn acquire_image( &mut self, _timeout_ns: u64, ) -> Result<(Self::SwapchainImage, Option<window::Suboptimal>), window::AcquireError> { let image = native::ImageView::Renderbuffer(self.renderbuffer.unwrap()); Ok((image, None)) } } #[derive(Debug)] pub struct Swapchain { pub(crate) fbos: ArrayVec<[native::RawFrameBuffer; 3]>, pub(crate) context: PresentContext, pub(crate) extent: window::Extent2D, } impl Swapchain { pub(crate) fn make_current(&self) { self.context.make_current(); } pub(crate) fn swap_buffers(&self) { self.context.swap_buffers(); } } impl window::Swapchain<Backend> for Swapchain { unsafe fn acquire_image( &mut self, _timeout_ns: u64, _semaphore:
{ addr as *const _ }
conditional_block
wgl.rs
gl") .encode_wide() .chain(Some(0).into_iter()) .collect::<Vec<_>>(); class.cbSize = mem::size_of::<WNDCLASSEXW>() as UINT; class.lpszClassName = class_name.as_ptr(); class.lpfnWndProc = Some(DefWindowProcW); RegisterClassExW(&class); let hwnd = CreateWindowExW( 0, class_name.as_ptr(), std::ptr::null(), 0, CW_USEDEFAULT, CW_USEDEFAULT, CW_USEDEFAULT, CW_USEDEFAULT, std::ptr::null_mut(), std::ptr::null_mut(), instance, std::ptr::null_mut(), ); let hdc = GetDC(hwnd); let desc = PIXELFORMATDESCRIPTOR { nSize: std::mem::size_of::<PIXELFORMATDESCRIPTOR>() as u16, nVersion: 1, dwFlags: PFD_SUPPORT_OPENGL, iPixelType: PFD_TYPE_RGBA, cColorBits: 8, cRedBits: 0, cRedShift: 0, cGreenBits: 0, cGreenShift: 0, cBlueBits: 0, cBlueShift: 0, cAlphaBits: 8, cAlphaShift: 0, cAccumBits: 0, cAccumRedBits: 0, cAccumGreenBits: 0, cAccumBlueBits: 0, cAccumAlphaBits: 0, cDepthBits: 0, cStencilBits: 0, cAuxBuffers: 0, iLayerType: PFD_MAIN_PLANE, bReserved: 0, dwLayerMask: 0, dwVisibleMask: 0, dwDamageMask: 0, }; let format_id = ChoosePixelFormat(hdc, &desc); SetPixelFormat(hdc, format_id, &desc); let hglrc = wglCreateContext(hdc); println!("{:?}", (hwnd, hdc, format_id, hglrc)); wglMakeCurrent(hdc, hglrc); let name = OsStr::new("opengl32.dll") .encode_wide() .chain(Some(0).into_iter()) .collect::<Vec<_>>(); let lib = LoadLibraryW(name.as_ptr()); let wgl = wgl_ext_sys::Wgl::load_with(|sym| { let sym = CString::new(sym.as_bytes()).unwrap(); let addr = wgl_sys::GetProcAddress(sym.as_ptr()) as *const (); if !addr.is_null() { addr as *const _ } else { GetProcAddress(lib, sym.as_ptr()) as *const _ } }); Entry { hwnd, hdc: hdc as _, wgl, lib, } } } } impl Drop for Entry { fn drop(&mut self) { unsafe { DestroyWindow(self.hwnd); } } } lazy_static! { // Entry function pointers pub(crate) static ref WGL_ENTRY: Entry = Entry::new(); } pub struct Instance { pub(crate) ctxt: DeviceContext, } impl Instance { pub fn create(_name: &str, version: u32) -> Result<Self, hal::UnsupportedBackend> { unsafe { let glrc = WGL_ENTRY.wgl.CreateContextAttribsARB( WGL_ENTRY.hdc as *const _, ptr::null(), ptr::null(), ) as HGLRC; wglMakeCurrent(WGL_ENTRY.hdc as *mut _, glrc); Ok(Instance { ctxt: DeviceContext { ctxt: Context { glrc }, hdc: WGL_ENTRY.hdc, }, }) } } #[cfg(windows)] pub fn create_surface_from_hwnd(&self, hwnd: *mut c_void) -> Surface { Surface { hwnd: hwnd as *mut _, swapchain: None, renderbuffer: None, } } #[cfg(feature = "winit")] pub fn create_surface(&self, window: &winit::window::Window) -> Surface { use winit::platform::windows::WindowExtWindows; let hwnd = window.hwnd(); self.create_surface_from_hwnd(hwnd as *mut _) } } impl hal::Instance for Instance { type Backend = Backend; fn enumerate_adapters(&self) -> Vec<Adapter<Backend>> { let gl_container = GlContainer::from_fn_proc(|s| unsafe { let sym = CString::new(s.as_bytes()).unwrap(); let addr = wgl_sys::GetProcAddress(sym.as_ptr()) as *const (); if !addr.is_null() { addr as *const _ } else { GetProcAddress(WGL_ENTRY.lib, sym.as_ptr()) as *const _ } }); let adapter = PhysicalDevice::new_adapter(self.ctxt, gl_container); vec![adapter] } } #[derive(Debug)] pub struct Surface { pub(crate) hwnd: HWND, pub(crate) swapchain: Option<Swapchain>, renderbuffer: Option<native::Renderbuffer>, } // TODO: high -msiglreith unsafe impl Send for Surface {} unsafe impl Sync for Surface {} impl window::Surface<Backend> for Surface { fn compatibility( &self, physical_device: &PhysicalDevice, ) -> ( window::SurfaceCapabilities, Option<Vec<f::Format>>, Vec<window::PresentMode>, ) { let extent = unsafe { let mut rect: RECT = mem::zeroed(); GetClientRect(self.hwnd, &mut rect); window::Extent2D { width: (rect.right - rect.left) as _, height: (rect.bottom - rect.top) as _, } }; let caps = window::SurfaceCapabilities { image_count: 2 ..= 2, current_extent: Some(extent), extents: extent ..= extent, max_image_layers: 1, usage: image::Usage::COLOR_ATTACHMENT | image::Usage::TRANSFER_SRC, composite_alpha: window::CompositeAlpha::OPAQUE, //TODO }; let present_modes = vec![ window::PresentMode::Fifo, //TODO ]; ( caps, Some(vec![f::Format::Rgba8Srgb, f::Format::Bgra8Srgb]), present_modes, ) } fn supports_queue_family(&self, _queue_family: &QueueFamily) -> bool
} impl window::PresentationSurface<Backend> for Surface { type SwapchainImage = native::ImageView; unsafe fn configure_swapchain( &mut self, device: &Device, config: window::SwapchainConfig, ) -> Result<(), window::CreationError> { let gl = &device.share.context; let context = match self.swapchain.take() { Some(old) => { for fbo in old.fbos { gl.delete_framebuffer(fbo); } old.context } None => PresentContext::new(self, &device.share.instance_context), }; context.make_current(); if self.renderbuffer.is_none() { self.renderbuffer = Some(gl.create_renderbuffer().unwrap()); } let desc = conv::describe_format(config.format).unwrap(); gl.bind_renderbuffer(glow::RENDERBUFFER, self.renderbuffer); gl.renderbuffer_storage( glow::RENDERBUFFER, desc.tex_internal, config.extent.width as i32, config.extent.height as i32, ); let fbo = gl.create_framebuffer().unwrap(); gl.bind_framebuffer(glow::READ_FRAMEBUFFER, Some(fbo)); gl.framebuffer_renderbuffer( glow::READ_FRAMEBUFFER, glow::COLOR_ATTACHMENT0, glow::RENDERBUFFER, self.renderbuffer, ); self.swapchain = Some(Swapchain { context, extent: config.extent, fbos: iter::once(fbo).collect(), }); Ok(()) } unsafe fn unconfigure_swapchain(&mut self, device: &Device) { let gl = &device.share.context; if let Some(old) = self.swapchain.take() { for fbo in old.fbos { gl.delete_framebuffer(fbo); } } if let Some(rbo) = self.renderbuffer.take() { gl.delete_renderbuffer(rbo); } } unsafe fn acquire_image( &mut self, _timeout_ns: u64, ) -> Result<(Self::SwapchainImage, Option<window::Suboptimal>), window::AcquireError> { let image = native::ImageView::Renderbuffer(self.renderbuffer.unwrap()); Ok((image, None)) } } #[derive(Debug)] pub struct Swapchain { pub(crate) fbos: ArrayVec<[native::RawFrameBuffer; 3]>, pub(crate) context: PresentContext, pub(crate) extent: window::Extent2D, } impl Swapchain { pub(crate) fn make_current(&self) { self.context.make_current(); } pub(crate) fn swap_buffers(&self) { self.context.swap_buffers(); } } impl window::Swapchain<Backend> for Swapchain { unsafe fn acquire_image( &mut self, _timeout_ns: u64, _semaphore:
{ true }
identifier_body
wgl.rs
FD_TYPE_RGBA, cColorBits: 8, cRedBits: 0, cRedShift: 0, cGreenBits: 0, cGreenShift: 0, cBlueBits: 0, cBlueShift: 0, cAlphaBits: 8, cAlphaShift: 0, cAccumBits: 0, cAccumRedBits: 0, cAccumGreenBits: 0, cAccumBlueBits: 0, cAccumAlphaBits: 0, cDepthBits: 0, cStencilBits: 0, cAuxBuffers: 0, iLayerType: PFD_MAIN_PLANE, bReserved: 0, dwLayerMask: 0, dwVisibleMask: 0, dwDamageMask: 0, }; let format_id = ChoosePixelFormat(hdc, &desc); SetPixelFormat(hdc, format_id, &desc); let hglrc = wglCreateContext(hdc); println!("{:?}", (hwnd, hdc, format_id, hglrc)); wglMakeCurrent(hdc, hglrc); let name = OsStr::new("opengl32.dll") .encode_wide() .chain(Some(0).into_iter()) .collect::<Vec<_>>(); let lib = LoadLibraryW(name.as_ptr()); let wgl = wgl_ext_sys::Wgl::load_with(|sym| { let sym = CString::new(sym.as_bytes()).unwrap(); let addr = wgl_sys::GetProcAddress(sym.as_ptr()) as *const (); if !addr.is_null() { addr as *const _ } else { GetProcAddress(lib, sym.as_ptr()) as *const _ } }); Entry { hwnd, hdc: hdc as _, wgl, lib, } } } } impl Drop for Entry { fn drop(&mut self) { unsafe { DestroyWindow(self.hwnd); } } } lazy_static! { // Entry function pointers pub(crate) static ref WGL_ENTRY: Entry = Entry::new(); } pub struct Instance { pub(crate) ctxt: DeviceContext, } impl Instance { pub fn create(_name: &str, version: u32) -> Result<Self, hal::UnsupportedBackend> { unsafe { let glrc = WGL_ENTRY.wgl.CreateContextAttribsARB( WGL_ENTRY.hdc as *const _, ptr::null(), ptr::null(), ) as HGLRC; wglMakeCurrent(WGL_ENTRY.hdc as *mut _, glrc); Ok(Instance { ctxt: DeviceContext { ctxt: Context { glrc }, hdc: WGL_ENTRY.hdc, }, }) } } #[cfg(windows)] pub fn create_surface_from_hwnd(&self, hwnd: *mut c_void) -> Surface { Surface { hwnd: hwnd as *mut _, swapchain: None, renderbuffer: None, } } #[cfg(feature = "winit")] pub fn create_surface(&self, window: &winit::window::Window) -> Surface { use winit::platform::windows::WindowExtWindows; let hwnd = window.hwnd(); self.create_surface_from_hwnd(hwnd as *mut _) } } impl hal::Instance for Instance { type Backend = Backend; fn enumerate_adapters(&self) -> Vec<Adapter<Backend>> { let gl_container = GlContainer::from_fn_proc(|s| unsafe { let sym = CString::new(s.as_bytes()).unwrap(); let addr = wgl_sys::GetProcAddress(sym.as_ptr()) as *const (); if !addr.is_null() { addr as *const _ } else { GetProcAddress(WGL_ENTRY.lib, sym.as_ptr()) as *const _ } }); let adapter = PhysicalDevice::new_adapter(self.ctxt, gl_container); vec![adapter] } } #[derive(Debug)] pub struct Surface { pub(crate) hwnd: HWND, pub(crate) swapchain: Option<Swapchain>, renderbuffer: Option<native::Renderbuffer>, } // TODO: high -msiglreith unsafe impl Send for Surface {} unsafe impl Sync for Surface {} impl window::Surface<Backend> for Surface { fn compatibility( &self, physical_device: &PhysicalDevice, ) -> ( window::SurfaceCapabilities, Option<Vec<f::Format>>, Vec<window::PresentMode>, ) { let extent = unsafe { let mut rect: RECT = mem::zeroed(); GetClientRect(self.hwnd, &mut rect); window::Extent2D { width: (rect.right - rect.left) as _, height: (rect.bottom - rect.top) as _, } }; let caps = window::SurfaceCapabilities { image_count: 2 ..= 2, current_extent: Some(extent), extents: extent ..= extent, max_image_layers: 1, usage: image::Usage::COLOR_ATTACHMENT | image::Usage::TRANSFER_SRC, composite_alpha: window::CompositeAlpha::OPAQUE, //TODO }; let present_modes = vec![ window::PresentMode::Fifo, //TODO ]; ( caps, Some(vec![f::Format::Rgba8Srgb, f::Format::Bgra8Srgb]), present_modes, ) } fn supports_queue_family(&self, _queue_family: &QueueFamily) -> bool { true } } impl window::PresentationSurface<Backend> for Surface { type SwapchainImage = native::ImageView; unsafe fn configure_swapchain( &mut self, device: &Device, config: window::SwapchainConfig, ) -> Result<(), window::CreationError> { let gl = &device.share.context; let context = match self.swapchain.take() { Some(old) => { for fbo in old.fbos { gl.delete_framebuffer(fbo); } old.context } None => PresentContext::new(self, &device.share.instance_context), }; context.make_current(); if self.renderbuffer.is_none() { self.renderbuffer = Some(gl.create_renderbuffer().unwrap()); } let desc = conv::describe_format(config.format).unwrap(); gl.bind_renderbuffer(glow::RENDERBUFFER, self.renderbuffer); gl.renderbuffer_storage( glow::RENDERBUFFER, desc.tex_internal, config.extent.width as i32, config.extent.height as i32, ); let fbo = gl.create_framebuffer().unwrap(); gl.bind_framebuffer(glow::READ_FRAMEBUFFER, Some(fbo)); gl.framebuffer_renderbuffer( glow::READ_FRAMEBUFFER, glow::COLOR_ATTACHMENT0, glow::RENDERBUFFER, self.renderbuffer, ); self.swapchain = Some(Swapchain { context, extent: config.extent, fbos: iter::once(fbo).collect(), }); Ok(()) } unsafe fn unconfigure_swapchain(&mut self, device: &Device) { let gl = &device.share.context; if let Some(old) = self.swapchain.take() { for fbo in old.fbos { gl.delete_framebuffer(fbo); } } if let Some(rbo) = self.renderbuffer.take() { gl.delete_renderbuffer(rbo); } } unsafe fn acquire_image( &mut self, _timeout_ns: u64, ) -> Result<(Self::SwapchainImage, Option<window::Suboptimal>), window::AcquireError> { let image = native::ImageView::Renderbuffer(self.renderbuffer.unwrap()); Ok((image, None)) } } #[derive(Debug)] pub struct Swapchain { pub(crate) fbos: ArrayVec<[native::RawFrameBuffer; 3]>, pub(crate) context: PresentContext, pub(crate) extent: window::Extent2D, } impl Swapchain { pub(crate) fn make_current(&self) { self.context.make_current(); } pub(crate) fn swap_buffers(&self) { self.context.swap_buffers(); } } impl window::Swapchain<Backend> for Swapchain { unsafe fn acquire_image( &mut self, _timeout_ns: u64, _semaphore: Option<&native::Semaphore>, _fence: Option<&native::Fence>, ) -> Result<(window::SwapImageIndex, Option<window::Suboptimal>), window::AcquireError> { Ok((0, None)) // TODO } } /// Basic abstraction for wgl context handles. #[derive(Debug, Copy, Clone)] struct Context { glrc: HGLRC, } impl Context { unsafe fn make_current(&self, hdc: HDC) { wglMakeCurrent(hdc, self.glrc); } } /// Owned context for devices and instances. #[derive(Debug, Copy, Clone)] pub(crate) struct DeviceContext { /// Owned wgl context. ctxt: Context, /// Device context owned by the corresponding instance. /// /// This refers to either a pbuffer or dummy window. Therefore not used for actual presentation. hdc: HDC, } // TODO unsafe impl Send for DeviceContext {} unsafe impl Sync for DeviceContext {} impl DeviceContext { pub(crate) fn
make_current
identifier_name
RunExperiment.py
parser = argparse.ArgumentParser(description='Run simple linear track experiment.') parser.add_argument('-P', '--serial-port', default='/dev/ttyACM0', help='TTY device for USB-serial interface (e.g., /dev/ttyUSB0 or COM10)') parser.add_argument('-C','--param-file', default='defaults.yaml', help='YAML file containing task parameters') parser.add_argument('-R','--random-seed', default=None, help='Random seed. If specified, this also overrides the YAML configuration file.') parser.add_argument('--output-dir', default=None, help='Directory to write output file (defaults to cwd)') parser.add_argument('--no-check-space', default=None, help='Exits if less than 10 GB of space is available.') args = parser.parse_args() print(args) if args.param_file == 'defaults.yaml': warnings.warn('Using default configuration file. That is almost certainly not what you want to do!') # YAML parameters: task settings with open(args.param_file, 'r') as f: Config = yaml.safe_load(f) # ------------------- Validate config file------------------------------------------------------------- if 'AuditoryStimuli' in Config: from treadmillio.soundstimulus import validate_sound_config validate_sound_config(Config['AuditoryStimuli']) # ------------------- Setup logging. ------------------------------------------------------------------ DoLogCommands = Config['Preferences'].get('LogCommands', True) if DoLogCommands: auto_log_directory = Config['Preferences'].get('AutoLogDirectory', True) if 'Preferences' in Config else True log_directory = Config['Preferences'].get('LogDirectory', None) if 'Preferences' in Config else None if log_directory is not None and args.output_dir is not None: warnings.warn('The configuration file specifies {} for logging, ' 'but command line has {}. Using command line!\n'.format(log_directory, args.output_dir)) log_directory = args.output_dir elif args.output_dir is not None: log_directory = args.output_dir elif auto_log_directory: now = datetime.datetime.now() log_root = Config['Preferences'].get('LogDirectoryRoot', '') if 'Preferences' in Config else '' log_directory = os.path.join(log_root, '{}{}'.format('ExperimentLog', now.strftime("%Y-%m-%d_%H%M"))) else: raise(ValueError('You did not specify a directory for experiment logs, and AutoLogDirectory is False.')) if not os.path.isabs(log_directory): log_directory = os.path.join(os.getcwd(), log_directory) orig_log_directory = log_directory k=1 while os.path.exists(log_directory): k = k + 1 log_directory = orig_log_directory + '_' + str(k) if log_directory != orig_log_directory: warnings.warn('Specified experiment logs directory {} exists, using {}'.format(orig_log_directory, log_directory)) print('Creating log directory: {}\n'.format(log_directory)) os.makedirs(log_directory) # Check for available space! if not args.no_check_space: disk_total, disk_used, disk_free = shutil.disk_usage(log_directory) if disk_free < 10*1024.0**3: # if less than 10 GB is available, exit print("\n!!!! Only {} MB available, exiting. Use the '--no-check-space' " "command line option to override. !!!!".format(disk_free/(1024.0**2))) os.removedirs(log_directory) exit(0) else: print('#'*80, '\n') print('Warning!!! Not logging!!!!') print('#'*80, '\n') log_directory = None EnableSound = Config['Preferences'].get('EnableSound', False) # Check for random seed on command line or in preferences if args.random_seed is not None: np.random.seed(int(args.random_seed)) print(f'Setting random seed to {args.random_seed}.') if 'RandomSeed' in Config['Preferences']: Config['Preferences']['RandomSeed'] = int(args.random_seed) print('Overwriting random seed in preferences file (true value will be logged).') elif 'RandomSeed' in Config['Preferences']: np.random.seed(Config['Preferences']['RandomSeed']) print(f"Setting random seed to {Config['Preferences']['RandomSeed']}.") with ExitStack() as stack: # -------------- Initialize Serial IO - Won't actually do anything until we call connect()! -------------------------- from treadmillio.serialinterface import SerialInterface gpio_config = Config.get('GPIO', None) if not gpio_config: warnings.warn("No GPIOs specified in config file. All IOs will be inputs.", RuntimeWarning) maze_config = Config.get('Maze', None) if 'Preferences' in Config: zmq_streaming = Config['Preferences'].get('DataStreamingPort', None) Interface = stack.enter_context(SerialInterface(SerialPort=args.serial_port, gpio_config=gpio_config, maze_config=maze_config, zmq_streaming=zmq_streaming)) #----------------------- Sound stimuli -------------- if 'AuditoryStimuli' in Config and EnableSound: from treadmillio.soundstimulus import SoundStimulusController SoundController = stack.enter_context(SoundStimulusController(Config['AuditoryStimuli'], Interface.virtual_track_length, Interface.maze_topology, log_directory)) else: SoundController = None if 'AuditoryStimuli' in Config: warnings.warn("Config file specified AuditoryStimuli, but EnableSound is False.", RuntimeWarning) # ------------------- Read in State Machine States. ------------------------------------------------------------------ if 'StateMachine' in Config: from treadmillio.taskstatemachine import TaskStateMachine # BUG: Should check to make sure states are all connected properly? StateMachine = stack.enter_context(TaskStateMachine(Config['StateMachine'], Interface, SoundController)) else: StateMachine = None # ------------------- Read in VR Reward Zones. ------------------------------------------------------------------ if 'RewardZones' in Config: from treadmillio.rewardzone import RewardZoneController RewardZones = RewardZoneController(Config['RewardZones'], Interface, SoundController) else: RewardZones = None if DoLogCommands: # -------------------------- Set up all the different log files ------------------------------------- # Log git diffs for provenance import git # gitpython repo = git.Repo(search_parent_directories=True) GitCommit = repo.head.object.hexsha GitChangedFiles = [fn.a_path for fn in repo.index.diff(None)] GitPatch = [fn.diff for fn in repo.index.diff(None, create_patch=True)] with open(os.path.join(log_directory, 'ExperimentCodeDiffs.txt'), 'w') as git_file: print(f' Git Commit: {GitCommit}',file=git_file) if GitChangedFiles: print(f' ChangedFiles: {GitChangedFiles}',file=git_file) print(f'Patch:\n{GitPatch}',file=git_file) # Log config file used with open(os.path.join(log_directory, 'ParsedConfig.yaml'), 'w') as yaml_file: yaml.dump(Config, yaml_file, indent=4) # Create data log file and write header log_file = stack.enter_context(open(os.path.join(log_directory, 'DataLog.csv'), 'w', newline='')) print(f'Experiment Data File.\n Version {NamedVersion}',file=log_file) log_writer = csv.writer(log_file) # logging is actually CSV format if StateMachine and DoLogCommands: # Create state machine log file and write header state_machine_log = stack.enter_context(open(os.path.join(log_directory, 'StatemachineLog.csv'), 'w', newline='')) print(f'State Machine Log File.\n Version {NamedVersion}',file=state_machine_log) state_log_writer = csv.writer(state_machine_log) if RewardZones and DoLogCommands: # Create state machine log file and write header reward_zone_log = stack.enter_context(open(os.path.join(log_directory, 'RewardzoneLog.csv'), 'w', newline='', buffering=1)) print(f'Reward Zone Log File.\n Version {NamedVersion}',file=reward_zone_log) reward_zone_writer = csv.writer(reward_zone_log) if Profiling: execution_log = stack.enter_context(open(os.path.join(log_directory, 'execution.csv'), 'w', newline='')) execution_writer = csv.writer(execution_log) # ------------------- Webcam Video Recording. ------------------------------------------------------------------
else: for cameraname, camera in Config['Cameras'].items(): if camera['RecordVideo']: print('Over-riding camera configuration to not record video or timestamps!!!') camera['RecordVideo'] = False for cameraname, camera in Config['Cameras'].items(): shared_termination_flag = RunCameraInterface(camera) # this starts a bunch of processes # ------------------- Webcam Video Recording. ------------------------------------------------------------------ if 'GigE-Cameras' in Config: from treadmillio.camera.gigecam import RunCameraInterface if DoLogCommands: for cameraname, camera in Config['GigE-Cameras'].items(): camera['LogDirectory'] = log_directory else: for cameraname,
if 'Cameras' in Config: from treadmillio.uvccam.uvccam import RunCameraInterface if DoLogCommands: for cameraname, camera in Config['Cameras'].items(): camera['LogDirectory'] = log_directory
random_line_split
RunExperiment.py
os.path.join(os.getcwd(), log_directory) orig_log_directory = log_directory k=1 while os.path.exists(log_directory): k = k + 1 log_directory = orig_log_directory + '_' + str(k) if log_directory != orig_log_directory: warnings.warn('Specified experiment logs directory {} exists, using {}'.format(orig_log_directory, log_directory)) print('Creating log directory: {}\n'.format(log_directory)) os.makedirs(log_directory) # Check for available space! if not args.no_check_space: disk_total, disk_used, disk_free = shutil.disk_usage(log_directory) if disk_free < 10*1024.0**3: # if less than 10 GB is available, exit print("\n!!!! Only {} MB available, exiting. Use the '--no-check-space' " "command line option to override. !!!!".format(disk_free/(1024.0**2))) os.removedirs(log_directory) exit(0) else: print('#'*80, '\n') print('Warning!!! Not logging!!!!') print('#'*80, '\n') log_directory = None EnableSound = Config['Preferences'].get('EnableSound', False) # Check for random seed on command line or in preferences if args.random_seed is not None: np.random.seed(int(args.random_seed)) print(f'Setting random seed to {args.random_seed}.') if 'RandomSeed' in Config['Preferences']: Config['Preferences']['RandomSeed'] = int(args.random_seed) print('Overwriting random seed in preferences file (true value will be logged).') elif 'RandomSeed' in Config['Preferences']: np.random.seed(Config['Preferences']['RandomSeed']) print(f"Setting random seed to {Config['Preferences']['RandomSeed']}.") with ExitStack() as stack: # -------------- Initialize Serial IO - Won't actually do anything until we call connect()! -------------------------- from treadmillio.serialinterface import SerialInterface gpio_config = Config.get('GPIO', None) if not gpio_config: warnings.warn("No GPIOs specified in config file. All IOs will be inputs.", RuntimeWarning) maze_config = Config.get('Maze', None) if 'Preferences' in Config: zmq_streaming = Config['Preferences'].get('DataStreamingPort', None) Interface = stack.enter_context(SerialInterface(SerialPort=args.serial_port, gpio_config=gpio_config, maze_config=maze_config, zmq_streaming=zmq_streaming)) #----------------------- Sound stimuli -------------- if 'AuditoryStimuli' in Config and EnableSound: from treadmillio.soundstimulus import SoundStimulusController SoundController = stack.enter_context(SoundStimulusController(Config['AuditoryStimuli'], Interface.virtual_track_length, Interface.maze_topology, log_directory)) else: SoundController = None if 'AuditoryStimuli' in Config: warnings.warn("Config file specified AuditoryStimuli, but EnableSound is False.", RuntimeWarning) # ------------------- Read in State Machine States. ------------------------------------------------------------------ if 'StateMachine' in Config: from treadmillio.taskstatemachine import TaskStateMachine # BUG: Should check to make sure states are all connected properly? StateMachine = stack.enter_context(TaskStateMachine(Config['StateMachine'], Interface, SoundController)) else: StateMachine = None # ------------------- Read in VR Reward Zones. ------------------------------------------------------------------ if 'RewardZones' in Config: from treadmillio.rewardzone import RewardZoneController RewardZones = RewardZoneController(Config['RewardZones'], Interface, SoundController) else: RewardZones = None if DoLogCommands: # -------------------------- Set up all the different log files ------------------------------------- # Log git diffs for provenance import git # gitpython repo = git.Repo(search_parent_directories=True) GitCommit = repo.head.object.hexsha GitChangedFiles = [fn.a_path for fn in repo.index.diff(None)] GitPatch = [fn.diff for fn in repo.index.diff(None, create_patch=True)] with open(os.path.join(log_directory, 'ExperimentCodeDiffs.txt'), 'w') as git_file: print(f' Git Commit: {GitCommit}',file=git_file) if GitChangedFiles: print(f' ChangedFiles: {GitChangedFiles}',file=git_file) print(f'Patch:\n{GitPatch}',file=git_file) # Log config file used with open(os.path.join(log_directory, 'ParsedConfig.yaml'), 'w') as yaml_file: yaml.dump(Config, yaml_file, indent=4) # Create data log file and write header log_file = stack.enter_context(open(os.path.join(log_directory, 'DataLog.csv'), 'w', newline='')) print(f'Experiment Data File.\n Version {NamedVersion}',file=log_file) log_writer = csv.writer(log_file) # logging is actually CSV format if StateMachine and DoLogCommands: # Create state machine log file and write header state_machine_log = stack.enter_context(open(os.path.join(log_directory, 'StatemachineLog.csv'), 'w', newline='')) print(f'State Machine Log File.\n Version {NamedVersion}',file=state_machine_log) state_log_writer = csv.writer(state_machine_log) if RewardZones and DoLogCommands: # Create state machine log file and write header reward_zone_log = stack.enter_context(open(os.path.join(log_directory, 'RewardzoneLog.csv'), 'w', newline='', buffering=1)) print(f'Reward Zone Log File.\n Version {NamedVersion}',file=reward_zone_log) reward_zone_writer = csv.writer(reward_zone_log) if Profiling: execution_log = stack.enter_context(open(os.path.join(log_directory, 'execution.csv'), 'w', newline='')) execution_writer = csv.writer(execution_log) # ------------------- Webcam Video Recording. ------------------------------------------------------------------ if 'Cameras' in Config: from treadmillio.uvccam.uvccam import RunCameraInterface if DoLogCommands: for cameraname, camera in Config['Cameras'].items(): camera['LogDirectory'] = log_directory else: for cameraname, camera in Config['Cameras'].items(): if camera['RecordVideo']: print('Over-riding camera configuration to not record video or timestamps!!!') camera['RecordVideo'] = False for cameraname, camera in Config['Cameras'].items(): shared_termination_flag = RunCameraInterface(camera) # this starts a bunch of processes # ------------------- Webcam Video Recording. ------------------------------------------------------------------ if 'GigE-Cameras' in Config: from treadmillio.camera.gigecam import RunCameraInterface if DoLogCommands: for cameraname, camera in Config['GigE-Cameras'].items(): camera['LogDirectory'] = log_directory else: for cameraname, camera in Config['GigE-Cameras'].items(): if camera['RecordVideo']: print('Over-riding camera configuration to not record video or timestamps!!!') camera['RecordVideo'] = False for cameraname, camera in Config['GigE-Cameras'].items(): shared_termination_flag = RunCameraInterface(camera) # this starts a bunch of processes # TODO: Figure out how to handle errors below. The shared termination flag should work, but it doesn't # ----------------- Initialization ##### Actually connect to IO device. We wait until here so that data doesn't get lost/confused in serial buffer Interface.connect() FlagChar, StructSize, MasterTime, InitialEncoder, InitialUnwrappedEncoder, InitialGPIO, AuxGPIO = Interface.read_data() # This will initialize encoder if SoundController: SoundController.start_capture() # TODO: This doesn't currently do anything if StateMachine: StateMachine.start(MasterTime) first_sample = True while(True): ## every 2 ms happens: FlagChar, StructSize, MasterTime, Encoder, UnwrappedEncoder, GPIO, AuxGPIO = Interface.read_data() last_ts = time.monotonic() # to match with miniscope timestamps (which is written in msec, here is sec) # since read_data() is blocking, this is a farther bound (i.e., ts AFTER) data if DoLogCommands: if not first_sample: log_writer.writerow([MasterTime, GPIO, Encoder, UnwrappedEncoder, last_ts, Interface.pos, Interface.velocity]) # Log data from serial interface else: # for ths first sample, to synchronize to a meaningful clock, we the CLOCK_REALTIME time, in the first row sys_ts = time.time() log_writer.writerow([0, InitialGPIO, InitialEncoder, UnwrappedEncoder, sys_ts, 0, 0]) log_writer.writerow([MasterTime, GPIO, Encoder, UnwrappedEncoder, last_ts, Interface.pos, Interface.velocity]) first_sample = False # -------------------- Updates -------------------- Interface.update_pulses() # lower any outstanding GPIO pulses if SoundController: SoundController.update_beeps(MasterTime) # stop any outstanding beeps if StateMachine:
if DoLogCommands: StateMachine.update_statemachine(state_log_writer.writerow) # update the state machine else: StateMachine.update_statemachine(None) # update the state machine
conditional_block
single_iceberg_model.py
Download timing start start_download = time.time () print ("Downloading Web data to files") #Downloading data and writing to files function def download (url): """ Downloads web data and writes it to a file Parameters ---------- url : String Web address used as the data source. Returns ------- None. """ path, url = url r = requests.get (url, stream = True) content = r.text #print (content) with open (path + '.txt', 'w') as f: f.write (content) #List of url web data urls = [('Lidar', 'https://www.geog.leeds.ac.uk/courses/computing/study/core-python-odl2/assessment2/white1.lidar'), ('Radar', 'https://www.geog.leeds.ac.uk/courses/computing/study/core-python-odl2/assessment2/white1.radar')] #Call download function for x in urls: download (x) """ Stage 3: Reading csv data into lists """ print ("Reading csv data") #Reading csv into lidar #Lidar represents the height of the pixel within the environment lidar = [] with open ('Lidar.txt', newline = '') as f1: reader = csv.reader (f1, quoting=csv.QUOTE_NONNUMERIC) for row in reader: #print (row) rowlist_L = [] for value in row: #print (value) rowlist_L.append (int (value)) lidar.append (rowlist_L) #for row in lidar: #print ("length of row:", len (row)) #List contains empty rows #Remove empty rows lidar_clean = [x for x in lidar if x != []] #print ("lidar2 length =", len(lidar_clean)) # for row in lidar_clean: # print ("length of row:", len (row)) #Data familiaristation # print (lidar_clean) # print (type(lidar_clean)) # print ("Lidar length = ", len(lidar_clean)) #Reading csv into radar #Radar represents ice or water within the environment radar = [] with open ('Radar.txt', newline = '') as f2: reader2 = csv.reader (f2, quoting=csv.QUOTE_NONNUMERIC) for row in reader2: #print (row) rowlist_R = [] for value in row: #print (value) rowlist_R.append (int (value)) radar.append (rowlist_R) #for row in radar: #print ("length of row:", len (row)) #list contains empty rows #remove empty rows radar_clean = [x for x in radar if x != []] #for row in radar_clean: #print ("length of row:", len (row)) #Data familiarisation # print (radar_clean) # print (type(radar_clean)) # print ("Radar length = ", len(radar_clean)) #Download and list creation timing end_download = time.time () download_time = (end_download - start_download) print ("Download to reading in time: " + str(download_time)) # """ # Displaying lidar and radar data # """ # #Code based on https://www.kite.com/python/answers/how-to-show-two-figures-at-once-in-matplotlib-in-python # #Define axes # plt.ylim = (0, 300) # plt.xlim = (0, 300) # #Lidar plot # lidar_plot = plt.figure (1) # #Assign title # plt.title ('Lidar data') # plt.imshow (lidar_clean) # #Radar plot # radar_plot = plt.figure (2) # #Assign title # plt.title ('Radar data') # plt.imshow (radar_clean) # #Show plots # plt.show () #Commented code above has been moved into the GUI so removed here """ Stage 4: Finding ice areas and pulling their heights """ #Calculation timing start_calculation = time.time () #Data familiarisation # print (radar_clean [145][150]) # print (radar_clean [145]) # print (lidar_clean [145][150]) # print (lidar_clean [145]) print ("Locating ice and pulling heights") #Pulling heights from lidar data def pull_heights (): """ Pulls height values from the lidar data from ice locations within the radar data, appending the heights to an ice list Returns ------- None. """ global ice ice = [] for i in range (len (radar_clean)): for j in range (len (radar_clean)):
#print (ice) #print (len (ice)) print ("Ice located and heights pulled") """ Stage 5: Calculating ice mass """ print ("Determining iceberg size") #Pull heights def calculations (): pull_heights () ice_size = (len(ice)) #print ("ice size:", ice_size) #Calculating ice volume above sea level print ("Calculating ice mass") #Convert ice values from cm to m ice_m = ((sum (ice)) * 0.1) #print (ice_m) #Calculating ice volume above surface global ice_volume_positive ice_volume_positive = (ice_size * ice_m) #print (ice_volume_positive) #Calculating sub-surface ice volume global ice_volume_subsurface ice_volume_subsurface = ice_volume_positive * 10 #print (ice_volume_subsurface) #Calculating total ice volume global ice_volume ice_volume = ice_volume_positive + ice_volume_subsurface #print (ice_volume) #Calculating ice mass global ice_mass ice_mass = 900 * ice_volume #print (ice_mass) print ("Ice mass calculated") run = tkinter.Label\ (text= ("Running Program" + "\n")) run.pack () results_btn ['state'] = 'normal' # """ # Stage 6: Calculating towability # """ # print ("Calculating towability") # def towability (): # """ # Determines the towability of the iceberg # Returns # ------- # None. # """ # if ice_mass > 36000000: # print ("Iceberg cannot be towed") # else: # print ("Iceberg can be towed") # print ("Towability calculated") #Commented towability code above moved into GUI so removed here #Calculation timing end_calculation = time.time () calculation_time = (end_calculation - start_calculation) print ("Calculation time: ", calculation_time) # """ # Stage 6: Writing data out to a file # """ # with open ("Data_out.txt", 'w') as FO: # FO.write ("Above surface volume: " + str(ice_volume_positive) + '\n') # FO.write ("Subsurface volume: " + str(ice_volume_subsurface) + '\n') # FO.write ("Total ice volume: " + str(ice_volume) + '\n') # FO.write ("Total mass: " + str(ice_mass) + '\n') # Section moved to after GUI is completed as to access global variables # within calculations function """ Stage 6: Initialise and populate GUI """ #Label code based on https://www.python-course.eu/tkinter_text_widget.php #Create GUI description description = tkinter.Label\ (text="This GUI runs the Iceberg Towability Model and shows its inputs and results"\ + '\n') description.pack () #Create user defined run # menu_bar = tkinter.Menu (root) # root.config (menu = menu_bar) # model_menu = tkinter.Menu (menu_bar) # menu_bar.add_cascade (label= "Model", menu= model_menu) # model_menu.add_command (label= "Run Model", command= calculations) run_btn = tkinter.Button (root, text= 'Run Program', command = calculations) run_btn.pack (side= 'top', fill= 'both') #Show input data graphs def show_graphs (): """ Generates two figures simultaneously displaying lidar and radar input datasets. Figures will appear in individual pop-up windows. Returns ------- None. """ plt.ylim = (0, 300) plt.xlim = (0, 300) #Set up lidar plot to figure 1 lidar_plot = plt.figure (1) #Assign title plt.title ('Lidar data') #Assign data plt.imshow (lidar_clean) #Set up radar plot to figure 2 radar_plot = plt.figure (2) #Assign title plt.title ('Radar data') #Assign data plt.imshow (radar_clean) #Show plots plt.show () #Display the results def show_results ():
radar_clean [i][j] = lidar_clean [i][j] #print (radar_clean [i][j]) if (radar_clean [i][j]) > 100: #print (radar_clean [i][j]) ice.append (lidar_clean [i][j])
conditional_block
single_iceberg_model.py
Download timing start start_download = time.time () print ("Downloading Web data to files") #Downloading data and writing to files function def download (url): """ Downloads web data and writes it to a file Parameters ---------- url : String Web address used as the data source. Returns ------- None. """ path, url = url r = requests.get (url, stream = True) content = r.text #print (content) with open (path + '.txt', 'w') as f: f.write (content) #List of url web data urls = [('Lidar', 'https://www.geog.leeds.ac.uk/courses/computing/study/core-python-odl2/assessment2/white1.lidar'), ('Radar', 'https://www.geog.leeds.ac.uk/courses/computing/study/core-python-odl2/assessment2/white1.radar')] #Call download function for x in urls: download (x) """ Stage 3: Reading csv data into lists """ print ("Reading csv data") #Reading csv into lidar #Lidar represents the height of the pixel within the environment lidar = [] with open ('Lidar.txt', newline = '') as f1: reader = csv.reader (f1, quoting=csv.QUOTE_NONNUMERIC) for row in reader: #print (row) rowlist_L = [] for value in row: #print (value) rowlist_L.append (int (value)) lidar.append (rowlist_L) #for row in lidar: #print ("length of row:", len (row)) #List contains empty rows #Remove empty rows lidar_clean = [x for x in lidar if x != []] #print ("lidar2 length =", len(lidar_clean)) # for row in lidar_clean: # print ("length of row:", len (row)) #Data familiaristation # print (lidar_clean) # print (type(lidar_clean)) # print ("Lidar length = ", len(lidar_clean)) #Reading csv into radar #Radar represents ice or water within the environment radar = [] with open ('Radar.txt', newline = '') as f2: reader2 = csv.reader (f2, quoting=csv.QUOTE_NONNUMERIC) for row in reader2: #print (row) rowlist_R = [] for value in row: #print (value) rowlist_R.append (int (value)) radar.append (rowlist_R) #for row in radar: #print ("length of row:", len (row)) #list contains empty rows #remove empty rows radar_clean = [x for x in radar if x != []] #for row in radar_clean: #print ("length of row:", len (row)) #Data familiarisation # print (radar_clean) # print (type(radar_clean)) # print ("Radar length = ", len(radar_clean)) #Download and list creation timing end_download = time.time () download_time = (end_download - start_download) print ("Download to reading in time: " + str(download_time)) # """ # Displaying lidar and radar data # """ # #Code based on https://www.kite.com/python/answers/how-to-show-two-figures-at-once-in-matplotlib-in-python # #Define axes # plt.ylim = (0, 300) # plt.xlim = (0, 300) # #Lidar plot # lidar_plot = plt.figure (1) # #Assign title # plt.title ('Lidar data') # plt.imshow (lidar_clean) # #Radar plot # radar_plot = plt.figure (2) # #Assign title # plt.title ('Radar data') # plt.imshow (radar_clean) # #Show plots # plt.show () #Commented code above has been moved into the GUI so removed here """ Stage 4: Finding ice areas and pulling their heights """ #Calculation timing start_calculation = time.time () #Data familiarisation # print (radar_clean [145][150]) # print (radar_clean [145]) # print (lidar_clean [145][150]) # print (lidar_clean [145]) print ("Locating ice and pulling heights") #Pulling heights from lidar data def pull_heights (): """ Pulls height values from the lidar data from ice locations within the radar data, appending the heights to an ice list Returns ------- None.
for i in range (len (radar_clean)): for j in range (len (radar_clean)): radar_clean [i][j] = lidar_clean [i][j] #print (radar_clean [i][j]) if (radar_clean [i][j]) > 100: #print (radar_clean [i][j]) ice.append (lidar_clean [i][j]) #print (ice) #print (len (ice)) print ("Ice located and heights pulled") """ Stage 5: Calculating ice mass """ print ("Determining iceberg size") #Pull heights def calculations (): pull_heights () ice_size = (len(ice)) #print ("ice size:", ice_size) #Calculating ice volume above sea level print ("Calculating ice mass") #Convert ice values from cm to m ice_m = ((sum (ice)) * 0.1) #print (ice_m) #Calculating ice volume above surface global ice_volume_positive ice_volume_positive = (ice_size * ice_m) #print (ice_volume_positive) #Calculating sub-surface ice volume global ice_volume_subsurface ice_volume_subsurface = ice_volume_positive * 10 #print (ice_volume_subsurface) #Calculating total ice volume global ice_volume ice_volume = ice_volume_positive + ice_volume_subsurface #print (ice_volume) #Calculating ice mass global ice_mass ice_mass = 900 * ice_volume #print (ice_mass) print ("Ice mass calculated") run = tkinter.Label\ (text= ("Running Program" + "\n")) run.pack () results_btn ['state'] = 'normal' # """ # Stage 6: Calculating towability # """ # print ("Calculating towability") # def towability (): # """ # Determines the towability of the iceberg # Returns # ------- # None. # """ # if ice_mass > 36000000: # print ("Iceberg cannot be towed") # else: # print ("Iceberg can be towed") # print ("Towability calculated") #Commented towability code above moved into GUI so removed here #Calculation timing end_calculation = time.time () calculation_time = (end_calculation - start_calculation) print ("Calculation time: ", calculation_time) # """ # Stage 6: Writing data out to a file # """ # with open ("Data_out.txt", 'w') as FO: # FO.write ("Above surface volume: " + str(ice_volume_positive) + '\n') # FO.write ("Subsurface volume: " + str(ice_volume_subsurface) + '\n') # FO.write ("Total ice volume: " + str(ice_volume) + '\n') # FO.write ("Total mass: " + str(ice_mass) + '\n') # Section moved to after GUI is completed as to access global variables # within calculations function """ Stage 6: Initialise and populate GUI """ #Label code based on https://www.python-course.eu/tkinter_text_widget.php #Create GUI description description = tkinter.Label\ (text="This GUI runs the Iceberg Towability Model and shows its inputs and results"\ + '\n') description.pack () #Create user defined run # menu_bar = tkinter.Menu (root) # root.config (menu = menu_bar) # model_menu = tkinter.Menu (menu_bar) # menu_bar.add_cascade (label= "Model", menu= model_menu) # model_menu.add_command (label= "Run Model", command= calculations) run_btn = tkinter.Button (root, text= 'Run Program', command = calculations) run_btn.pack (side= 'top', fill= 'both') #Show input data graphs def show_graphs (): """ Generates two figures simultaneously displaying lidar and radar input datasets. Figures will appear in individual pop-up windows. Returns ------- None. """ plt.ylim = (0, 300) plt.xlim = (0, 300) #Set up lidar plot to figure 1 lidar_plot = plt.figure (1) #Assign title plt.title ('Lidar data') #Assign data plt.imshow (lidar_clean) #Set up radar plot to figure 2 radar_plot = plt.figure (2) #Assign title plt.title ('Radar data') #Assign data plt.imshow (radar_clean) #Show plots plt.show () #Display the results def show_results ():
""" global ice ice = []
random_line_split
single_iceberg_model.py
Download timing start start_download = time.time () print ("Downloading Web data to files") #Downloading data and writing to files function def download (url):
#List of url web data urls = [('Lidar', 'https://www.geog.leeds.ac.uk/courses/computing/study/core-python-odl2/assessment2/white1.lidar'), ('Radar', 'https://www.geog.leeds.ac.uk/courses/computing/study/core-python-odl2/assessment2/white1.radar')] #Call download function for x in urls: download (x) """ Stage 3: Reading csv data into lists """ print ("Reading csv data") #Reading csv into lidar #Lidar represents the height of the pixel within the environment lidar = [] with open ('Lidar.txt', newline = '') as f1: reader = csv.reader (f1, quoting=csv.QUOTE_NONNUMERIC) for row in reader: #print (row) rowlist_L = [] for value in row: #print (value) rowlist_L.append (int (value)) lidar.append (rowlist_L) #for row in lidar: #print ("length of row:", len (row)) #List contains empty rows #Remove empty rows lidar_clean = [x for x in lidar if x != []] #print ("lidar2 length =", len(lidar_clean)) # for row in lidar_clean: # print ("length of row:", len (row)) #Data familiaristation # print (lidar_clean) # print (type(lidar_clean)) # print ("Lidar length = ", len(lidar_clean)) #Reading csv into radar #Radar represents ice or water within the environment radar = [] with open ('Radar.txt', newline = '') as f2: reader2 = csv.reader (f2, quoting=csv.QUOTE_NONNUMERIC) for row in reader2: #print (row) rowlist_R = [] for value in row: #print (value) rowlist_R.append (int (value)) radar.append (rowlist_R) #for row in radar: #print ("length of row:", len (row)) #list contains empty rows #remove empty rows radar_clean = [x for x in radar if x != []] #for row in radar_clean: #print ("length of row:", len (row)) #Data familiarisation # print (radar_clean) # print (type(radar_clean)) # print ("Radar length = ", len(radar_clean)) #Download and list creation timing end_download = time.time () download_time = (end_download - start_download) print ("Download to reading in time: " + str(download_time)) # """ # Displaying lidar and radar data # """ # #Code based on https://www.kite.com/python/answers/how-to-show-two-figures-at-once-in-matplotlib-in-python # #Define axes # plt.ylim = (0, 300) # plt.xlim = (0, 300) # #Lidar plot # lidar_plot = plt.figure (1) # #Assign title # plt.title ('Lidar data') # plt.imshow (lidar_clean) # #Radar plot # radar_plot = plt.figure (2) # #Assign title # plt.title ('Radar data') # plt.imshow (radar_clean) # #Show plots # plt.show () #Commented code above has been moved into the GUI so removed here """ Stage 4: Finding ice areas and pulling their heights """ #Calculation timing start_calculation = time.time () #Data familiarisation # print (radar_clean [145][150]) # print (radar_clean [145]) # print (lidar_clean [145][150]) # print (lidar_clean [145]) print ("Locating ice and pulling heights") #Pulling heights from lidar data def pull_heights (): """ Pulls height values from the lidar data from ice locations within the radar data, appending the heights to an ice list Returns ------- None. """ global ice ice = [] for i in range (len (radar_clean)): for j in range (len (radar_clean)): radar_clean [i][j] = lidar_clean [i][j] #print (radar_clean [i][j]) if (radar_clean [i][j]) > 100: #print (radar_clean [i][j]) ice.append (lidar_clean [i][j]) #print (ice) #print (len (ice)) print ("Ice located and heights pulled") """ Stage 5: Calculating ice mass """ print ("Determining iceberg size") #Pull heights def calculations (): pull_heights () ice_size = (len(ice)) #print ("ice size:", ice_size) #Calculating ice volume above sea level print ("Calculating ice mass") #Convert ice values from cm to m ice_m = ((sum (ice)) * 0.1) #print (ice_m) #Calculating ice volume above surface global ice_volume_positive ice_volume_positive = (ice_size * ice_m) #print (ice_volume_positive) #Calculating sub-surface ice volume global ice_volume_subsurface ice_volume_subsurface = ice_volume_positive * 10 #print (ice_volume_subsurface) #Calculating total ice volume global ice_volume ice_volume = ice_volume_positive + ice_volume_subsurface #print (ice_volume) #Calculating ice mass global ice_mass ice_mass = 900 * ice_volume #print (ice_mass) print ("Ice mass calculated") run = tkinter.Label\ (text= ("Running Program" + "\n")) run.pack () results_btn ['state'] = 'normal' # """ # Stage 6: Calculating towability # """ # print ("Calculating towability") # def towability (): # """ # Determines the towability of the iceberg # Returns # ------- # None. # """ # if ice_mass > 36000000: # print ("Iceberg cannot be towed") # else: # print ("Iceberg can be towed") # print ("Towability calculated") #Commented towability code above moved into GUI so removed here #Calculation timing end_calculation = time.time () calculation_time = (end_calculation - start_calculation) print ("Calculation time: ", calculation_time) # """ # Stage 6: Writing data out to a file # """ # with open ("Data_out.txt", 'w') as FO: # FO.write ("Above surface volume: " + str(ice_volume_positive) + '\n') # FO.write ("Subsurface volume: " + str(ice_volume_subsurface) + '\n') # FO.write ("Total ice volume: " + str(ice_volume) + '\n') # FO.write ("Total mass: " + str(ice_mass) + '\n') # Section moved to after GUI is completed as to access global variables # within calculations function """ Stage 6: Initialise and populate GUI """ #Label code based on https://www.python-course.eu/tkinter_text_widget.php #Create GUI description description = tkinter.Label\ (text="This GUI runs the Iceberg Towability Model and shows its inputs and results"\ + '\n') description.pack () #Create user defined run # menu_bar = tkinter.Menu (root) # root.config (menu = menu_bar) # model_menu = tkinter.Menu (menu_bar) # menu_bar.add_cascade (label= "Model", menu= model_menu) # model_menu.add_command (label= "Run Model", command= calculations) run_btn = tkinter.Button (root, text= 'Run Program', command = calculations) run_btn.pack (side= 'top', fill= 'both') #Show input data graphs def show_graphs (): """ Generates two figures simultaneously displaying lidar and radar input datasets. Figures will appear in individual pop-up windows. Returns ------- None. """ plt.ylim = (0, 300) plt.xlim = (0, 300) #Set up lidar plot to figure 1 lidar_plot = plt.figure (1) #Assign title plt.title ('Lidar data') #Assign data plt.imshow (lidar_clean) #Set up radar plot to figure 2 radar_plot = plt.figure (2) #Assign title plt.title ('Radar data') #Assign data plt.imshow (radar_clean) #Show plots plt.show () #Display the results def show_results ():
""" Downloads web data and writes it to a file Parameters ---------- url : String Web address used as the data source. Returns ------- None. """ path, url = url r = requests.get (url, stream = True) content = r.text #print (content) with open (path + '.txt', 'w') as f: f.write (content)
identifier_body
single_iceberg_model.py
Download timing start start_download = time.time () print ("Downloading Web data to files") #Downloading data and writing to files function def download (url): """ Downloads web data and writes it to a file Parameters ---------- url : String Web address used as the data source. Returns ------- None. """ path, url = url r = requests.get (url, stream = True) content = r.text #print (content) with open (path + '.txt', 'w') as f: f.write (content) #List of url web data urls = [('Lidar', 'https://www.geog.leeds.ac.uk/courses/computing/study/core-python-odl2/assessment2/white1.lidar'), ('Radar', 'https://www.geog.leeds.ac.uk/courses/computing/study/core-python-odl2/assessment2/white1.radar')] #Call download function for x in urls: download (x) """ Stage 3: Reading csv data into lists """ print ("Reading csv data") #Reading csv into lidar #Lidar represents the height of the pixel within the environment lidar = [] with open ('Lidar.txt', newline = '') as f1: reader = csv.reader (f1, quoting=csv.QUOTE_NONNUMERIC) for row in reader: #print (row) rowlist_L = [] for value in row: #print (value) rowlist_L.append (int (value)) lidar.append (rowlist_L) #for row in lidar: #print ("length of row:", len (row)) #List contains empty rows #Remove empty rows lidar_clean = [x for x in lidar if x != []] #print ("lidar2 length =", len(lidar_clean)) # for row in lidar_clean: # print ("length of row:", len (row)) #Data familiaristation # print (lidar_clean) # print (type(lidar_clean)) # print ("Lidar length = ", len(lidar_clean)) #Reading csv into radar #Radar represents ice or water within the environment radar = [] with open ('Radar.txt', newline = '') as f2: reader2 = csv.reader (f2, quoting=csv.QUOTE_NONNUMERIC) for row in reader2: #print (row) rowlist_R = [] for value in row: #print (value) rowlist_R.append (int (value)) radar.append (rowlist_R) #for row in radar: #print ("length of row:", len (row)) #list contains empty rows #remove empty rows radar_clean = [x for x in radar if x != []] #for row in radar_clean: #print ("length of row:", len (row)) #Data familiarisation # print (radar_clean) # print (type(radar_clean)) # print ("Radar length = ", len(radar_clean)) #Download and list creation timing end_download = time.time () download_time = (end_download - start_download) print ("Download to reading in time: " + str(download_time)) # """ # Displaying lidar and radar data # """ # #Code based on https://www.kite.com/python/answers/how-to-show-two-figures-at-once-in-matplotlib-in-python # #Define axes # plt.ylim = (0, 300) # plt.xlim = (0, 300) # #Lidar plot # lidar_plot = plt.figure (1) # #Assign title # plt.title ('Lidar data') # plt.imshow (lidar_clean) # #Radar plot # radar_plot = plt.figure (2) # #Assign title # plt.title ('Radar data') # plt.imshow (radar_clean) # #Show plots # plt.show () #Commented code above has been moved into the GUI so removed here """ Stage 4: Finding ice areas and pulling their heights """ #Calculation timing start_calculation = time.time () #Data familiarisation # print (radar_clean [145][150]) # print (radar_clean [145]) # print (lidar_clean [145][150]) # print (lidar_clean [145]) print ("Locating ice and pulling heights") #Pulling heights from lidar data def pull_heights (): """ Pulls height values from the lidar data from ice locations within the radar data, appending the heights to an ice list Returns ------- None. """ global ice ice = [] for i in range (len (radar_clean)): for j in range (len (radar_clean)): radar_clean [i][j] = lidar_clean [i][j] #print (radar_clean [i][j]) if (radar_clean [i][j]) > 100: #print (radar_clean [i][j]) ice.append (lidar_clean [i][j]) #print (ice) #print (len (ice)) print ("Ice located and heights pulled") """ Stage 5: Calculating ice mass """ print ("Determining iceberg size") #Pull heights def calculations (): pull_heights () ice_size = (len(ice)) #print ("ice size:", ice_size) #Calculating ice volume above sea level print ("Calculating ice mass") #Convert ice values from cm to m ice_m = ((sum (ice)) * 0.1) #print (ice_m) #Calculating ice volume above surface global ice_volume_positive ice_volume_positive = (ice_size * ice_m) #print (ice_volume_positive) #Calculating sub-surface ice volume global ice_volume_subsurface ice_volume_subsurface = ice_volume_positive * 10 #print (ice_volume_subsurface) #Calculating total ice volume global ice_volume ice_volume = ice_volume_positive + ice_volume_subsurface #print (ice_volume) #Calculating ice mass global ice_mass ice_mass = 900 * ice_volume #print (ice_mass) print ("Ice mass calculated") run = tkinter.Label\ (text= ("Running Program" + "\n")) run.pack () results_btn ['state'] = 'normal' # """ # Stage 6: Calculating towability # """ # print ("Calculating towability") # def towability (): # """ # Determines the towability of the iceberg # Returns # ------- # None. # """ # if ice_mass > 36000000: # print ("Iceberg cannot be towed") # else: # print ("Iceberg can be towed") # print ("Towability calculated") #Commented towability code above moved into GUI so removed here #Calculation timing end_calculation = time.time () calculation_time = (end_calculation - start_calculation) print ("Calculation time: ", calculation_time) # """ # Stage 6: Writing data out to a file # """ # with open ("Data_out.txt", 'w') as FO: # FO.write ("Above surface volume: " + str(ice_volume_positive) + '\n') # FO.write ("Subsurface volume: " + str(ice_volume_subsurface) + '\n') # FO.write ("Total ice volume: " + str(ice_volume) + '\n') # FO.write ("Total mass: " + str(ice_mass) + '\n') # Section moved to after GUI is completed as to access global variables # within calculations function """ Stage 6: Initialise and populate GUI """ #Label code based on https://www.python-course.eu/tkinter_text_widget.php #Create GUI description description = tkinter.Label\ (text="This GUI runs the Iceberg Towability Model and shows its inputs and results"\ + '\n') description.pack () #Create user defined run # menu_bar = tkinter.Menu (root) # root.config (menu = menu_bar) # model_menu = tkinter.Menu (menu_bar) # menu_bar.add_cascade (label= "Model", menu= model_menu) # model_menu.add_command (label= "Run Model", command= calculations) run_btn = tkinter.Button (root, text= 'Run Program', command = calculations) run_btn.pack (side= 'top', fill= 'both') #Show input data graphs def
(): """ Generates two figures simultaneously displaying lidar and radar input datasets. Figures will appear in individual pop-up windows. Returns ------- None. """ plt.ylim = (0, 300) plt.xlim = (0, 300) #Set up lidar plot to figure 1 lidar_plot = plt.figure (1) #Assign title plt.title ('Lidar data') #Assign data plt.imshow (lidar_clean) #Set up radar plot to figure 2 radar_plot = plt.figure (2) #Assign title plt.title ('Radar data') #Assign data plt.imshow (radar_clean) #Show plots plt.show () #Display the results def show_results ():
show_graphs
identifier_name
buffio_test.go
()) // 4 br.UnreadByte() // 撤销吐出,即栈中弹出的a元素又放回来了 fmt.Println(br.Buffered()) // 5 tmp, _ = br.ReadByte() fmt.Printf("%q\n", tmp) // 'a' } func TestReadRune1(t *testing.T) { // ReadRune读取一个utf-8编码的unicode码值 chinese := "中国人" s := strings.NewReader(chinese) br := bufio.NewReader(s) tmp, _, err := br.ReadRune() if err != nil { panic(err) } fmt.Printf("%q\n", tmp) // '中' } func TestUnReadRune1(t *testing.T) { chinese := "中国人" s := strings.NewReader(chinese) br := bufio.NewReader(s) tmp, _, err := br.ReadRune() if err != nil { panic(err) } fmt.Printf("%q\n", tmp) // '中' br.UnreadRune() tmp, _, err = br.ReadRune() if err != nil { panic(err) } fmt.Printf("%q\n", tmp) // '中' } func TestReadLine1(t *testing.T) { // ReadLine 是一个低水平的行读取原语,大多数情况下,应该使用 // ReadBytes('\n') 或 ReadString('\n'),或者使用一个 Scanner。 // // ReadLine 通过调用 ReadSlice 方法实现,返回的也是缓存的切片。用于 // 读取一行数据,不包括行尾标记(\n 或 \r\n)。 // // 只要能读出数据,err 就为 nil。如果没有数据可读,则 isPrefix 返回 // false,err 返回 io.EOF。 // // 如果找到行尾标记,则返回查找结果,isPrefix 返回 false。 // 如果未找到行尾标记,则: // 1、缓存不满,则将缓存填满后再次查找。 // 2、缓存是满的,则返回整个缓存,isPrefix 返回 true。 // // 整个数据尾部“有一个换行标记”和“没有换行标记”的读取结果是一样。 // // 如果 ReadLine 读取到换行标记,则调用 UnreadByte 撤销的是换行标记, // 而不是返回的数据。 // func (b *Reader) ReadLine() (line []byte, isPrefix bool, err error) s := strings.NewReader("123\nzzz") br := bufio.NewReader(s) for line, isPrefix, err := []byte{0}, false, error(nil); len(line) > 0 && err == nil; { line, isPrefix, err = br.ReadLine() fmt.Printf("%q %t %v\n", line, isPrefix, err) /* "123" false <nil> "zzz" false <nil> "" false EOF */ } } func TestReadSlice1(t *testing.T) { // ReadSlice 在 b 中查找 delim 并返回 delim 及其之前的所有数据。 // 该操作会读出数据,返回的切片是已读出的数据的引用,切片中的数据 // 在下一次读取操作之前是有效的。 // // 如果找到 delim,则返回查找结果,err 返回 nil。 // 如果未找到 delim,则: // 1、缓存不满,则将缓存填满后再次查找。 // 2、缓存是满的,则返回整个缓存,err 返回 ErrBufferFull。 // // 如果未找到 delim 且遇到错误(通常是 io.EOF),则返回缓存中的所 // 有数据和遇到的错误。 // // 因为返回的数据有可能被下一次的读写操作修改,所以大多数操作应该 // 使用 ReadBytes 或 ReadString,它们返回的是数据的拷贝。 // func (b *Reader) ReadSlice(delim byte) (line []byte, err error) s := strings.NewReader("ABC DEF GHI") br := bufio.NewReader(s) w, err := br.ReadSlice(' ') if err != nil { panic(err) } fmt.Printf("%q\n", w) // "ABC " w, err = br.ReadSlice(' ') if err != nil { panic(err) } fmt.Printf("%q\n", w) // "DEF " w, err = br.ReadSlice(' ') if err != nil { panic(err) } fmt.Printf("%q\n", w) // panic: EOF } func TestReadBytes1(t *testing.T) { // ReadBytes 功能同 ReadSlice,只不过返回的是缓存的拷贝。 // func (b *Reader) ReadBytes(delim byte) (line []byte, err error) s := strings.NewReader("ABC,EFG,HIJ") br := bufio.NewReader(s) line, err := br.ReadBytes(',') if err != nil { panic(err) } fmt.Printf("%q\n", line) // "ABC," line, err = br.ReadBytes(',') if err != nil { panic(err) } fmt.Printf("%q\n", line) // "EFG," line, err = br.ReadBytes(',') if err != nil { panic(err) // panic: EOF } fmt.Printf("%q\n", line) } func TestReadString1(t *testing.T) { // ReadString 功能同 ReadBytes,只不过返回的是字符串。 // func (b *Reader) ReadString(delim byte) (line string, err error) s := strings.NewReader("你好,我是卡牌") br := bufio.NewReader(s) line, err := br.ReadString(',') if err != nil { panic(err) } fmt.Printf("%s\n", line) // 你好, line, err = br.ReadString(',') if err != nil { panic(err) // panic: EOF } fmt.Printf("%s\n", line) } // bufio.Reader的WriteTo()和 bufio.Writer的ReadFrom() // WriteTo方法实现了io.WriterTo接口。 // func (b *Reader) WriteTo(w io.Writer) (n int64, err error) func TestWriteTo1(t *testing.T) { b := bytes.NewBuffer(make([]byte, 0)) s := strings.NewReader("ABCDEFG") br := bufio.NewReader(s) br.WriteTo(b) fmt.Printf("%s\n", b.String()) // ABCDEFG fmt.Printf("%q\n", b) // "ABCDEFG" fmt.Println(string(65)) // A } // ReadFrom实现了io.ReaderFrom接口。 // func (b *Writer) ReadFrom(r io.Reader) (n int64, err error) // ReadFrom无需使用Flush func TestReadFrom1(t *testing.T) { s := strings.NewReader("hello world") b := bytes.NewBuffer(make([]byte, 0)) bw := bufio.NewWriter(b) bw.ReadFrom(s) fmt.Println(b) } //go test -run=buffio_test.go -bench="BenchmarkReadFrom1" -benchtime="3s" -cpuprofile cpu.out // linux 是 -bench=. func BenchmarkReadFrom1(b *testing.B) { b.ReportAllocs() b.ResetTimer() for idx := 0; idx < b.N; idx++ { s := strings.NewReader("hello world") b := bytes.NewBuffer(make([]byte, 0)) bw := bufio.NewWriter(b) bw.ReadFrom(s) } b.StopTimer() } // Writer实现了为io.Writer接口对象提供缓冲。 // 如果在向一个Writer类型值写入时遇到了错误, // 该对象将不再接受任何数据,返回该错误 // 数据都写入后,调用者有义务调用Flush方法, // 保证所有的数据都交给了下层的io.Writer。 func TestNewWriter1(t *testing.T) { // NewWriter创建一个具有默认大小缓冲、写入w的*Writer。 相当于 NewWriterSize(wr, 4096) // func NewWriter(w io.Writer) *Writer // Buffered()返回缓冲中已使用的字节数。 // func (b *Writer) Buffered() int // Available()返回缓冲中还有多少字节未使用。 // func (b *Writer) Available() int // Reset丢弃缓冲中的数据,清除任何错误,将b重设为将其输出写入w。 // func (b *Writer) Reset(w io.Writer) b := bytes.NewBuffer(make([]byte, 0)) bw := bufio.NewWriter(b) fmt.Println(bw.Available(), bw.Buffered()) // 4096 0 bw.WriteString("card") fmt.Println(bw.Available(), bw.Buffered()) // 4092 4 bw.Reset(b) fmt.Println(bw.Available(), bw.Buffered()) // 4096 0 } func TestWrite1(t *testing.T) { // Write 将 p 中的数据写入 b 中,返回写入的字节数 // 如果写入的字节数小于 p 的长度,则返回一个错误信息 // func (b *Writer) Write(p []byte) (nn int, err error) // Flush 将缓存中的数据提交到底层的 io.Writer 中 // func (b *Writer) Flush() error p := [...]byte{'a', 'b', 'c'}
b := bytes.N
identifier_name
buffio_test.go
s %v %v\n", b[:n], n, err) // 0 EOF } func TestBuffered1(t *testing.T) { // 返回可以从缓存中读取的字节数 // func (b *Reader) Buffered() int { return b.w - b.r } s := strings.NewReader("123456789") br := bufio.NewReader(s) b := make([]byte, 3) br.Read(b) fmt.Println(br.Buffered()) // 6 br.Read(b) fmt.Println(br.Buffered()) // 3 } func TestReadByte1(t *testing.T) { // ReadByte读取并返回一个字节。如果没有可用的数据,会返回错误。 // func (b *Reader) ReadByte() (c byte, err error) origin := "abcd" s := strings.NewReader(origin) br := bufio.NewReader(s) // 第一次读取 tmp, err := br.ReadByte() if err != nil { panic(err) } fmt.Printf("%q\n", tmp) // 'a' // Buffered returns the number of bytes that can be read from the current buffer. fmt.Println(br.Buffered()) // 3 for i := 0; i < len(origin); i++ { tmp, err = br.ReadByte() if err != nil { // panic: EOF 因为已经读取了1个字符 缓存中只剩下3个 // 所以在读取第4个字符报错EOF panic(err) } } } func TestUnreadByte1(t *testing.T) { // 撤消最后读出的字节 s := strings.NewReader("abcde") br := bufio.NewReader(s) tmp, _ := br.ReadByte() fmt.Printf("%q\n", tmp) // 'a' fmt.Println(br.Buffered()) // 4 br.UnreadByte() // 撤销吐出,即栈中弹出的a元素又放回来了 fmt.Println(br.Buffered()) // 5 tmp, _ = br.ReadByte() fmt.Printf("%q\n", tmp) // 'a' } func TestReadRune1(t *testing.T) { // ReadRune读取一个utf-8编码的unicode码值 chinese := "中国人" s := strings.NewReader(chinese) br := bufio.NewReader(s) tmp, _, err := br.ReadRune() if err != nil { panic(err) } fmt.Printf("%q\n", tmp) // '中' } func TestUnReadRune1(t *testing.T) { chinese := "中国人" s := strings.NewReader(chinese) br := bufio.NewReader(s) tmp, _, err := br.ReadRune() if err != nil { panic(err) } fmt.Printf("%q\n", tmp) // '中' br.UnreadRune() tmp, _, err = br.ReadRune() if err != nil { panic(err) } fmt.Printf("%q\n", tmp) // '中' } func TestReadLine1(t *testing.T) { // ReadLine 是一个低水平的行读取原语,大多数情况下,应该使用 // ReadBytes('\n') 或 ReadString('\n'),或者使用一个 Scanner。 // // ReadLine 通过调用 ReadSlice 方法实现,返回的也是缓存的切片。用于 // 读取一行数据,不包括行尾标记(\n 或 \r\n)。 // // 只要能读出数据,err 就为 nil。如果没有数据可读,则 isPrefix 返回 // false,err 返回 io.EOF。 // // 如果找到行尾标记,则返回查找结果,isPrefix 返回 false。 // 如果未找到行尾标记,则:
// 1、缓存不满,则将缓存填满后再次查找。 // 2、缓存是满的,则返回整个缓存,isPrefix 返回 true。 // // 整个数据尾部“有一个换行标记”和“没有换行标记”的读取结果是一样。 // // 如果 ReadLine 读取到换行标记,则调用 UnreadByte 撤销的是换行标记, // 而不是返回的数据。 // func (b *Reader) ReadLine() (line []byte, isPrefix bool, err error) s := strings.NewReader("123\nzzz") br := bufio.NewReader(s) for line, isPrefix, err := []byte{0}, false, error(nil); len(line) > 0 && err == nil; { line, isPrefix, err = br.ReadLine() fmt.Printf("%q %t %v\n", line, isPrefix, err) /* "123" false <nil> "zzz" false <nil> "" false EOF */ } } func TestReadSlice1(t *testing.T) { // ReadSlice 在 b 中查找 delim 并返回 delim 及其之前的所有数据。 // 该操作会读出数据,返回的切片是已读出的数据的引用,切片中的数据 // 在下一次读取操作之前是有效的。 // // 如果找到 delim,则返回查找结果,err 返回 nil。 // 如果未找到 delim,则: // 1、缓存不满,则将缓存填满后再次查找。 // 2、缓存是满的,则返回整个缓存,err 返回 ErrBufferFull。 // // 如果未找到 delim 且遇到错误(通常是 io.EOF),则返回缓存中的所 // 有数据和遇到的错误。 // // 因为返回的数据有可能被下一次的读写操作修改,所以大多数操作应该 // 使用 ReadBytes 或 ReadString,它们返回的是数据的拷贝。 // func (b *Reader) ReadSlice(delim byte) (line []byte, err error) s := strings.NewReader("ABC DEF GHI") br := bufio.NewReader(s) w, err := br.ReadSlice(' ') if err != nil { panic(err) } fmt.Printf("%q\n", w) // "ABC " w, err = br.ReadSlice(' ') if err != nil { panic(err) } fmt.Printf("%q\n", w) // "DEF " w, err = br.ReadSlice(' ') if err != nil { panic(err) } fmt.Printf("%q\n", w) // panic: EOF } func TestReadBytes1(t *testing.T) { // ReadBytes 功能同 ReadSlice,只不过返回的是缓存的拷贝。 // func (b *Reader) ReadBytes(delim byte) (line []byte, err error) s := strings.NewReader("ABC,EFG,HIJ") br := bufio.NewReader(s) line, err := br.ReadBytes(',') if err != nil { panic(err) } fmt.Printf("%q\n", line) // "ABC," line, err = br.ReadBytes(',') if err != nil { panic(err) } fmt.Printf("%q\n", line) // "EFG," line, err = br.ReadBytes(',') if err != nil { panic(err) // panic: EOF } fmt.Printf("%q\n", line) } func TestReadString1(t *testing.T) { // ReadString 功能同 ReadBytes,只不过返回的是字符串。 // func (b *Reader) ReadString(delim byte) (line string, err error) s := strings.NewReader("你好,我是卡牌") br := bufio.NewReader(s) line, err := br.ReadString(',') if err != nil { panic(err) } fmt.Printf("%s\n", line) // 你好, line, err = br.ReadString(',') if err != nil { panic(err) // panic: EOF } fmt.Printf("%s\n", line) } // bufio.Reader的WriteTo()和 bufio.Writer的ReadFrom() // WriteTo方法实现了io.WriterTo接口。 // func (b *Reader) WriteTo(w io.Writer) (n int64, err error) func TestWriteTo1(t *testing.T) { b := bytes.NewBuffer(make([]byte, 0)) s := strings.NewReader("ABCDEFG") br := bufio.NewReader(s) br.WriteTo(b) fmt.Printf("%s\n", b.String()) // ABCDEFG fmt.Printf("%q\n", b) // "ABCDEFG" fmt.Println(string(65)) // A } // ReadFrom实现了io.ReaderFrom接口。 // func (b *Writer) ReadFrom(r io.Reader) (n int64, err error) // ReadFrom无需使用Flush func TestReadFrom1(t *testing.T) { s := strings.NewReader("hello world") b := bytes.NewBuffer(make([]byte, 0)) bw := bufio.NewWriter(b) bw.ReadFrom(s) fmt.Println(b) } //go test -run=buffio_test.go -bench="BenchmarkReadFrom1" -benchtime="3s" -cpuprofile cpu.out // linux 是 -bench=. func BenchmarkReadFrom1(b *testing.B) { b.ReportAllocs() b.ResetTimer() for idx := 0; idx < b.N; idx++ { s := strings.NewReader("hello world") b := bytes.NewBuffer(make([]byte, 0)) bw := bufio.NewWriter(b) bw.ReadFrom(s) } b.StopTimer() } // Writer实现了为io.Writer接口对象提供缓冲。 // 如果在向一个Writer类型值写入时遇到了错误, // 该对象
random_line_split
buffio_test.go
([]byte, 3) br.Read(b) fmt.Println(br.Buffered()) // 6 br.Read(b) fmt.Println(br.Buffered()) // 3 } func TestReadByte1(t *testing.T) { // ReadByte读取并返回一个字节。如果没有可用的数据,会返回错误。 // func (b *Reader) ReadByte() (c byte, err error) origin := "abcd" s := strings.NewReader(origin) br := bufio.NewReader(s) // 第一次读取 tmp, err := br.ReadByte() if err != nil { panic(err) } fmt.Printf("%q\n", tmp) // 'a' // Buffered returns the number of bytes that can be read from the current buffer. fmt.Println(br.Buffered()) // 3 for i := 0; i < len(origin); i++ { tmp, err = br.ReadByte() if err != nil { // panic: EOF 因为已经读取了1个字符 缓存中只剩下3个 // 所以在读取第4个字符报错EOF panic(err) } } } func TestUnreadByte1(t *testing.T) { // 撤消最后读出的字节 s := strings.NewReader("abcde") br := bufio.NewReader(s) tmp, _ := br.ReadByte() fmt.Printf("%q\n", tmp) // 'a' fmt.Println(br.Buffered()) // 4 br.UnreadByte() // 撤销吐出,即栈中弹出的a元素又放回来了 fmt.Println(br.Buffered()) // 5 tmp, _ = br.ReadByte() fmt.Printf("%q\n", tmp) // 'a' } func TestReadRune1(t *testing.T) { // ReadRune读取一个utf-8编码的unicode码值 chinese := "中国人" s := strings.NewReader(chinese) br := bufio.NewReader(s) tmp, _, err := br.ReadRune() if err != nil { panic(err) } fmt.Printf("%q\n", tmp) // '中' } func TestUnReadRune1(t *testing.T) { chinese := "中国人" s := strings.NewReader(chinese) br := bufio.NewReader(s) tmp, _, err := br.ReadRune() if err != nil { panic(err) } fmt.Printf("%q\n", tmp) // '中' br.UnreadRune() tmp, _, err = br.ReadRune() if err != nil { panic(err) } fmt.Printf("%q\n", tmp) // '中' } func TestReadLine1(t *testing.T) { // ReadLine 是一个低水平的行读取原语,大多数情况下,应该使用 // ReadBytes('\n') 或 ReadString('\n'),或者使用一个 Scanner。 // // ReadLine 通过调用 ReadSlice 方法实现,返回的也是缓存的切片。用于 // 读取一行数据,不包括行尾标记(\n 或 \r\n)。 // // 只要能读出数据,err 就为 nil。如果没有数据可读,则 isPrefix 返回 // false,err 返回 io.EOF。 // // 如果找到行尾标记,则返回查找结果,isPrefix 返回 false。 // 如果未找到行尾标记,则: // 1、缓存不满,则将缓存填满后再次查找。 // 2、缓存是满的,则返回整个缓存,isPrefix 返回 true。 // // 整个数据尾部“有一个换行标记”和“没有换行标记”的读取结果是一样。 // // 如果 ReadLine 读取到换行标记,则调用 UnreadByte 撤销的是换行标记, // 而不是返回的数据。 // func (b *Reader) ReadLine() (line []byte, isPrefix bool, err error) s := strings.NewReader("123\nzzz") br := bufio.NewReader(s) for line, isPrefix, err := []byte{0}, false, error(nil); len(line) > 0 && err == nil; { line, isPrefix, err = br.ReadLine() fmt.Printf("%q %t %v\n", line, isPrefix, err) /* "123" false <nil> "zzz" false <nil> "" false EOF */ } } func TestReadSlice1(t *testing.T) { // ReadSlice 在 b 中查找 delim 并返回 delim 及其之前的所有数据。 // 该操作会读出数据,返回的切片是已读出的数据的引用,切片中的数据 // 在下一次读取操作之前是有效的。 // // 如果找到 delim,则返回查找结果,err 返回 nil。 // 如果未找到 delim,则: // 1、缓存不满,则将缓存填满后再次查找。 // 2、缓存是满的,则返回整个缓存,err 返回 ErrBufferFull。 // // 如果未找到 delim 且遇到错误(通常是 io.EOF),则返回缓存中的所 // 有数据和遇到的错误。 // // 因为返回的数据有可能被下一次的读写操作修改,所以大多数操作应该 // 使用 ReadBytes 或 ReadString,它们返回的是数据的拷贝。 // func (b *Reader) ReadSlice(delim byte) (line []byte, err error) s := strings.NewReader("ABC DEF GHI") br := bufio.NewReader(s) w, err := br.ReadSlice(' ') if err != nil { panic(err) } fmt.Printf("%q\n", w) // "ABC " w, err = br.ReadSlice(' ') if err != nil { panic(err) } fmt.Printf("%q\n", w) // "DEF " w, err = br.ReadSlice(' ') if err != nil { panic(err) } fmt.Printf("%q\n", w) // panic: EOF } func TestReadBytes1(t *testing.T) { // ReadBytes 功能同 ReadSlice,只不过返回的是缓存的拷贝。 // func (b *Reader) ReadBytes(delim byte) (line []byte, err error) s := strings.NewReader("ABC,EFG,HIJ") br := bufio.NewReader(s) line, err := br.ReadBytes(',') if err != nil { panic(err) } fmt.Printf("%q\n", line) // "ABC," line, err = br.ReadBytes(',') if err != nil { panic(err) } fmt.Printf("%q\n", line) // "EFG," line, err = br.ReadBytes(',') if err != nil { panic(err) // panic: EOF } fmt.Printf("%q\n", line) } func TestReadString1(t *testing.T) { // ReadString 功能同 ReadBytes,只不过返回的是字符串。 // func (b *Reader) ReadString(delim byte) (line string, err error) s := strings.NewReader("你好,我是卡牌") br := bufio.NewReader(s) line, err := br.ReadString(',') if err != nil { panic(err) } fmt.Printf("%s\n", line) // 你好, line, err = br.ReadString(',') if err != nil { panic(err) // panic: EOF } fmt.Printf("%s\n", line) } // bufio.Reader的WriteTo()和 bufio.Writer的ReadFrom() // WriteTo方法实现了io.WriterTo接口。 // func (b *Reader) WriteTo(w io.Writer) (n int64, err error) func TestWriteTo1(t *testing.T) { b := bytes.NewBuffer(make([]byte, 0)) s := strings.NewReader("ABCDEFG") br := bufio.NewReader(s) br.WriteTo(b) fmt.Printf("%s\n", b.String()) // ABCDEFG fmt.Printf("%q\n", b) // "ABCDEFG" fmt.Println(string(65)) // A } // ReadFrom实现了io.ReaderFrom接口。 // func (b *Writer) ReadFrom(r io.Reader) (n int64, err error) // ReadFrom无需使用Flush func TestReadFrom1(t *testing.T) { s := strings.NewReader("hello world") b := bytes.NewBuffer(make([]byte, 0)) bw := bufio.NewWriter(b) bw.ReadFrom(s) fmt.Println(b) } //go test -run=buffio_test.go -bench="BenchmarkReadFrom1" -benchtime="3s" -cpuprofile cpu.out // linux 是 -bench=. func BenchmarkReadFrom1(b *testing.B) { b.ReportAllocs() b.ResetTimer() for idx := 0; idx < b.N; idx++ { s := strings.NewReader("hello world") b := bytes.NewBuffer(make([]byte, 0)) bw := bufio.NewWriter(b) bw.ReadFrom(s) } b.StopTimer() } // Writer实现了为io.Writer接口对象提供缓冲。 // 如果在向一个Writer类型值写入时遇到了错误, // 该对象将不再接受任何数据,返回该错误 // 数据都写入后,调用者有义务调用Flush方法, // 保证所有的数据都交给了下层的io.Writer。 func TestNewWriter1(t *testing.T) { // NewWriter创建一个具有默认大小缓冲、写入w的*Writer。 相当于 NewWriterSize(w
r, 4096) // func
conditional_block
buffio_test.go
err := []byte{0}, false, error(nil); len(line) > 0 && err == nil; { line, isPrefix, err = br.ReadLine() fmt.Printf("%q %t %v\n", line, isPrefix, err) /* "123" false <nil> "zzz" false <nil> "" false EOF */ } } func TestReadSlice1(t *testing.T) { // ReadSlice 在 b 中查找 delim 并返回 delim 及其之前的所有数据。 // 该操作会读出数据,返回的切片是已读出的数据的引用,切片中的数据 // 在下一次读取操作之前是有效的。 // // 如果找到 delim,则返回查找结果,err 返回 nil。 // 如果未找到 delim,则: // 1、缓存不满,则将缓存填满后再次查找。 // 2、缓存是满的,则返回整个缓存,err 返回 ErrBufferFull。 // // 如果未找到 delim 且遇到错误(通常是 io.EOF),则返回缓存中的所 // 有数据和遇到的错误。 // // 因为返回的数据有可能被下一次的读写操作修改,所以大多数操作应该 // 使用 ReadBytes 或 ReadString,它们返回的是数据的拷贝。 // func (b *Reader) ReadSlice(delim byte) (line []byte, err error) s := strings.NewReader("ABC DEF GHI") br := bufio.NewReader(s) w, err := br.ReadSlice(' ') if err != nil { panic(err) } fmt.Printf("%q\n", w) // "ABC " w, err = br.ReadSlice(' ') if err != nil { panic(err) } fmt.Printf("%q\n", w) // "DEF " w, err = br.ReadSlice(' ') if err != nil { panic(err) } fmt.Printf("%q\n", w) // panic: EOF } func TestReadBytes1(t *testing.T) { // ReadBytes 功能同 ReadSlice,只不过返回的是缓存的拷贝。 // func (b *Reader) ReadBytes(delim byte) (line []byte, err error) s := strings.NewReader("ABC,EFG,HIJ") br := bufio.NewReader(s) line, err := br.ReadBytes(',') if err != nil { panic(err) } fmt.Printf("%q\n", line) // "ABC," line, err = br.ReadBytes(',') if err != nil { panic(err) } fmt.Printf("%q\n", line) // "EFG," line, err = br.ReadBytes(',') if err != nil { panic(err) // panic: EOF } fmt.Printf("%q\n", line) } func TestReadString1(t *testing.T) { // ReadString 功能同 ReadBytes,只不过返回的是字符串。 // func (b *Reader) ReadString(delim byte) (line string, err error) s := strings.NewReader("你好,我是卡牌") br := bufio.NewReader(s) line, err := br.ReadString(',') if err != nil { panic(err) } fmt.Printf("%s\n", line) // 你好, line, err = br.ReadString(',') if err != nil { panic(err) // panic: EOF } fmt.Printf("%s\n", line) } // bufio.Reader的WriteTo()和 bufio.Writer的ReadFrom() // WriteTo方法实现了io.WriterTo接口。 // func (b *Reader) WriteTo(w io.Writer) (n int64, err error) func TestWriteTo1(t *testing.T) { b := bytes.NewBuffer(make([]byte, 0)) s := strings.NewReader("ABCDEFG") br := bufio.NewReader(s) br.WriteTo(b) fmt.Printf("%s\n", b.String()) // ABCDEFG fmt.Printf("%q\n", b) // "ABCDEFG" fmt.Println(string(65)) // A } // ReadFrom实现了io.ReaderFrom接口。 // func (b *Writer) ReadFrom(r io.Reader) (n int64, err error) // ReadFrom无需使用Flush func TestReadFrom1(t *testing.T) { s := strings.NewReader("hello world") b := bytes.NewBuffer(make([]byte, 0)) bw := bufio.NewWriter(b) bw.ReadFrom(s) fmt.Println(b) } //go test -run=buffio_test.go -bench="BenchmarkReadFrom1" -benchtime="3s" -cpuprofile cpu.out // linux 是 -bench=. func BenchmarkReadFrom1(b *testing.B) { b.ReportAllocs() b.ResetTimer() for idx := 0; idx < b.N; idx++ { s := strings.NewReader("hello world") b := bytes.NewBuffer(make([]byte, 0)) bw := bufio.NewWriter(b) bw.ReadFrom(s) } b.StopTimer() } // Writer实现了为io.Writer接口对象提供缓冲。 // 如果在向一个Writer类型值写入时遇到了错误, // 该对象将不再接受任何数据,返回该错误 // 数据都写入后,调用者有义务调用Flush方法, // 保证所有的数据都交给了下层的io.Writer。 func TestNewWriter1(t *testing.T) { // NewWriter创建一个具有默认大小缓冲、写入w的*Writer。 相当于 NewWriterSize(wr, 4096) // func NewWriter(w io.Writer) *Writer // Buffered()返回缓冲中已使用的字节数。 // func (b *Writer) Buffered() int // Available()返回缓冲中还有多少字节未使用。 // func (b *Writer) Available() int // Reset丢弃缓冲中的数据,清除任何错误,将b重设为将其输出写入w。 // func (b *Writer) Reset(w io.Writer) b := bytes.NewBuffer(make([]byte, 0)) bw := bufio.NewWriter(b) fmt.Println(bw.Available(), bw.Buffered()) // 4096 0 bw.WriteString("card") fmt.Println(bw.Available(), bw.Buffered()) // 4092 4 bw.Reset(b) fmt.Println(bw.Available(), bw.Buffered()) // 4096 0 } func TestWrite1(t *testing.T) { // Write 将 p 中的数据写入 b 中,返回写入的字节数 // 如果写入的字节数小于 p 的长度,则返回一个错误信息 // func (b *Writer) Write(p []byte) (nn int, err error) // Flush 将缓存中的数据提交到底层的 io.Writer 中 // func (b *Writer) Flush() error p := [...]byte{'a', 'b', 'c'} b := bytes.NewBuffer(make([]byte, 0)) bw := bufio.NewWriter(b) bw.Write(p[:]) bw.Flush() fmt.Printf("%q\n", b) //"abc" } func TestWriteString1(t *testing.T) { // WriteString 同 Write,只不过写入的是字符串 // func (b *Writer) WriteString(s string) (int, error) b := bytes.NewBuffer(make([]byte, 0)) bw := bufio.NewWriter(b) bw.WriteString("hello world") bw.Flush() fmt.Printf("%s\n", b) //hello world } func TestWriteByte1(t *testing.T) { // WriteByte写入单个字节。 // func (b *Writer) WriteByte(c byte) error b := bytes.NewBuffer(make([]byte, 0)) bw := bufio.NewWriter(b) bw.WriteByte('c') bw.Flush() fmt.Println(b) } func TestWriteRune1(t *testing.T) { // WriteRune写入一个unicode码值(的utf-8编码),返回写入的字节数和可能的错误。 // func (b *Writer) WriteRune(r rune) (size int, err error) b := bytes.NewBuffer(make([]byte, 0)) bw := bufio.NewWriter(b) size, err := bw.WriteRune('周') if err != nil { panic(err) } fmt.Println(size) // 3 bw.Flush() fmt.Println(b) // 周 } func TestReadWriter1(t *testing.T) { // ReadWriter类型保管了指向Reader和Writer类型的指针 // 实现了io.ReadWriter接口。 // NewReadWriter 生成bufio.ReadWriter对象 // func NewReadWriter(r *Reader, w *Writer) *ReadWriter b := bytes.NewBuffer(make([]byte, 0)) bw := bufio.NewWriter(b) s := strings.NewReader("hello world") br := bufio.NewReade
r(s) rw := bufio.NewReadWriter(br, bw) word, err := rw.ReadString(' ') if err != nil { panic(err) } fmt.Printf("%s\n", word) // hello _, err = rw.WriteString(",I'm coming") if err != nil { panic(err) } rw.Flush() fmt.Println(b) } func TestNewScanner1(t *testing.T) { // Scanner 提供了一个方便的接口来读取数据,例如遍历多行文本中的行。Scan 方法会通过 // 一个“匹配函数”读取数据中符合要求的部分,跳过不符合要求的部分。“匹配函数”由调
identifier_body
particle.rs
]; } #[rustfmt::skip] const PARTICLE_TEXTURE_PIXELS: [u8; 64] = [ 0, 0, 1, 1, 1, 1, 0, 0, 0, 1, 1, 1, 1, 1, 1, 0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 1, 1, 1, 1, 1, 1, 0, 0, 0, 1, 1, 1, 1, 0, 0, ]; pub struct ParticlePipeline { pipeline: wgpu::RenderPipeline, bind_group_layouts: Vec<wgpu::BindGroupLayout>, vertex_buffer: wgpu::Buffer, sampler: wgpu::Sampler, textures: Vec<wgpu::Texture>, texture_views: Vec<wgpu::TextureView>, bind_group: wgpu::BindGroup, } impl ParticlePipeline { pub fn new( device: &wgpu::Device, queue: &wgpu::Queue, compiler: &mut shaderc::Compiler, sample_count: u32, palette: &Palette, ) -> ParticlePipeline { let (pipeline, bind_group_layouts) = ParticlePipeline::create(device, compiler, &[], sample_count); use wgpu::util::DeviceExt as _; let vertex_buffer = device.create_buffer_init(&wgpu::util::BufferInitDescriptor { label: None, contents: unsafe { any_slice_as_bytes(&VERTICES) }, usage: wgpu::BufferUsage::VERTEX, }); let sampler = device.create_sampler(&wgpu::SamplerDescriptor { label: Some("particle sampler"), address_mode_u: wgpu::AddressMode::ClampToEdge, address_mode_v: wgpu::AddressMode::ClampToEdge, address_mode_w: wgpu::AddressMode::ClampToEdge, mag_filter: wgpu::FilterMode::Nearest, min_filter: wgpu::FilterMode::Linear, mipmap_filter: wgpu::FilterMode::Linear, lod_min_clamp: -1000.0, lod_max_clamp: 1000.0, compare: None, anisotropy_clamp: NonZeroU8::new(16), border_color: None, }); let textures: Vec<wgpu::Texture> = (0..256) .map(|i| { let mut pixels = PARTICLE_TEXTURE_PIXELS; // set up palette translation for pix in pixels.iter_mut() { if *pix == 0
else { *pix *= i as u8; } } let (diffuse_data, _) = palette.translate(&pixels); create_texture( device, queue, Some(&format!("particle texture {}", i)), 8, 8, &TextureData::Diffuse(diffuse_data), ) }) .collect(); let texture_views: Vec<wgpu::TextureView> = textures .iter() .map(|t| t.create_view(&Default::default())) .collect(); let texture_view_refs = texture_views.iter().collect::<Vec<_>>(); let bind_group = device.create_bind_group(&wgpu::BindGroupDescriptor { label: Some("particle bind group"), layout: &bind_group_layouts[0], entries: &[ wgpu::BindGroupEntry { binding: 0, resource: wgpu::BindingResource::Sampler(&sampler), }, wgpu::BindGroupEntry { binding: 1, resource: wgpu::BindingResource::TextureViewArray(&texture_view_refs[..]), }, ], }); ParticlePipeline { pipeline, bind_group_layouts, sampler, textures, texture_views, bind_group, vertex_buffer, } } pub fn rebuild( &mut self, device: &wgpu::Device, compiler: &mut shaderc::Compiler, sample_count: u32, ) { let layout_refs: Vec<_> = self.bind_group_layouts.iter().collect(); self.pipeline = ParticlePipeline::recreate(device, compiler, &layout_refs, sample_count); } pub fn pipeline(&self) -> &wgpu::RenderPipeline { &self.pipeline } pub fn bind_group_layouts(&self) -> &[wgpu::BindGroupLayout] { &self.bind_group_layouts } pub fn vertex_buffer(&self) -> &wgpu::Buffer { &self.vertex_buffer } pub fn record_draw<'a, 'b, P>( &'a self, pass: &mut wgpu::RenderPass<'a>, bump: &'a Bump, camera: &Camera, particles: P, ) where P: Iterator<Item = &'b Particle>, { use PushConstantUpdate::*; pass.set_pipeline(self.pipeline()); pass.set_vertex_buffer(0, self.vertex_buffer.slice(..)); pass.set_bind_group(0, &self.bind_group, &[]); // face toward camera let Angles { pitch, yaw, roll } = camera.angles(); let rotation = Angles { pitch: -pitch, yaw: -yaw, roll: -roll, } .mat4_wgpu(); for particle in particles { let q_origin = particle.origin(); let translation = Matrix4::from_translation([-q_origin.y, q_origin.z, -q_origin.x].into()); Self::set_push_constants( pass, Update(bump.alloc(VertexPushConstants { transform: camera.view_projection() * translation * rotation, })), Retain, Update(bump.alloc(FragmentPushConstants { color: particle.color() as u32, })), ); pass.draw(0..6, 0..1); } } } #[derive(Copy, Clone, Debug)] pub struct VertexPushConstants { pub transform: Matrix4<f32>, } #[derive(Copy, Clone, Debug)] pub struct FragmentPushConstants { pub color: u32, } const BIND_GROUP_LAYOUT_ENTRIES: &[wgpu::BindGroupLayoutEntry] = &[ wgpu::BindGroupLayoutEntry { binding: 0, visibility: wgpu::ShaderStage::FRAGMENT, ty: wgpu::BindingType::Sampler { filtering: true, comparison: false, }, count: None, }, // per-index texture array wgpu::BindGroupLayoutEntry { binding: 1, visibility: wgpu::ShaderStage::FRAGMENT, ty: wgpu::BindingType::Texture { view_dimension: wgpu::TextureViewDimension::D2, sample_type: wgpu::TextureSampleType::Float { filterable: true }, multisampled: false, }, count: NonZeroU32::new(256), }, ]; lazy_static! { static ref VERTEX_ATTRIBUTES: [[wgpu::VertexAttribute; 2]; 2] = [ wgpu::vertex_attr_array![ // position 0 => Float32x3, // texcoord 1 => Float32x2, ], wgpu::vertex_attr_array![ // instance position 2 => Float32x3, // color index 3 => Uint32, ] ]; } impl Pipeline for ParticlePipeline { type VertexPushConstants = VertexPushConstants; type SharedPushConstants = (); type FragmentPushConstants = FragmentPushConstants; fn name() -> &'static str { "particle" } fn vertex_shader() -> &'static str { include_str!(concat!( env!("CARGO_MANIFEST_DIR"), "/shaders/particle.vert" )) } fn fragment_shader() -> &'static str { include_str!(concat!( env!("CARGO_MANIFEST_DIR"), "/shaders/particle.frag" )) } // NOTE: if any of the binding indices are changed, they must also be changed in // the corresponding shaders and the BindGroupLayout generation functions. fn bind_group_layout_descriptors() -> Vec<wgpu::BindGroupLayoutDescriptor<'static>> { vec![ // group 0 wgpu::BindGroupLayoutDescriptor { label: Some("particle bind group layout"), entries: BIND_GROUP_LAYOUT_ENTRIES, }, ] } fn primitive_state() -> wgpu::PrimitiveState { WorldPipelineBase::primitive_state() } fn color_target_states() -> Vec<wgpu::ColorTargetState> { WorldPipelineBase::color_target_states() } fn depth_stencil_state() -> Option<wgpu::DepthStencilState> { let mut desc = WorldPipelineBase::depth_stencil_state().unwrap(); desc.depth_write_enabled = false; Some(desc)
{ *pix = 0xFF; }
conditional_block
particle.rs
]; } #[rustfmt::skip] const PARTICLE_TEXTURE_PIXELS: [u8; 64] = [ 0, 0, 1, 1, 1, 1, 0, 0, 0, 1, 1, 1, 1, 1, 1, 0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 1, 1, 1, 1, 1, 1, 0, 0, 0, 1, 1, 1, 1, 0, 0, ]; pub struct ParticlePipeline { pipeline: wgpu::RenderPipeline, bind_group_layouts: Vec<wgpu::BindGroupLayout>, vertex_buffer: wgpu::Buffer, sampler: wgpu::Sampler, textures: Vec<wgpu::Texture>, texture_views: Vec<wgpu::TextureView>, bind_group: wgpu::BindGroup, } impl ParticlePipeline { pub fn new( device: &wgpu::Device, queue: &wgpu::Queue, compiler: &mut shaderc::Compiler, sample_count: u32, palette: &Palette, ) -> ParticlePipeline { let (pipeline, bind_group_layouts) = ParticlePipeline::create(device, compiler, &[], sample_count); use wgpu::util::DeviceExt as _; let vertex_buffer = device.create_buffer_init(&wgpu::util::BufferInitDescriptor { label: None, contents: unsafe { any_slice_as_bytes(&VERTICES) }, usage: wgpu::BufferUsage::VERTEX, }); let sampler = device.create_sampler(&wgpu::SamplerDescriptor { label: Some("particle sampler"), address_mode_u: wgpu::AddressMode::ClampToEdge, address_mode_v: wgpu::AddressMode::ClampToEdge, address_mode_w: wgpu::AddressMode::ClampToEdge, mag_filter: wgpu::FilterMode::Nearest, min_filter: wgpu::FilterMode::Linear, mipmap_filter: wgpu::FilterMode::Linear, lod_min_clamp: -1000.0, lod_max_clamp: 1000.0, compare: None, anisotropy_clamp: NonZeroU8::new(16), border_color: None, }); let textures: Vec<wgpu::Texture> = (0..256) .map(|i| { let mut pixels = PARTICLE_TEXTURE_PIXELS; // set up palette translation for pix in pixels.iter_mut() { if *pix == 0 { *pix = 0xFF; } else { *pix *= i as u8; } } let (diffuse_data, _) = palette.translate(&pixels); create_texture( device, queue, Some(&format!("particle texture {}", i)), 8, 8, &TextureData::Diffuse(diffuse_data), ) }) .collect(); let texture_views: Vec<wgpu::TextureView> = textures .iter() .map(|t| t.create_view(&Default::default())) .collect(); let texture_view_refs = texture_views.iter().collect::<Vec<_>>(); let bind_group = device.create_bind_group(&wgpu::BindGroupDescriptor { label: Some("particle bind group"), layout: &bind_group_layouts[0], entries: &[ wgpu::BindGroupEntry { binding: 0, resource: wgpu::BindingResource::Sampler(&sampler), }, wgpu::BindGroupEntry { binding: 1, resource: wgpu::BindingResource::TextureViewArray(&texture_view_refs[..]), }, ], }); ParticlePipeline { pipeline, bind_group_layouts, sampler, textures, texture_views, bind_group, vertex_buffer, } } pub fn rebuild( &mut self, device: &wgpu::Device, compiler: &mut shaderc::Compiler, sample_count: u32, ) { let layout_refs: Vec<_> = self.bind_group_layouts.iter().collect(); self.pipeline = ParticlePipeline::recreate(device, compiler, &layout_refs, sample_count); } pub fn
(&self) -> &wgpu::RenderPipeline { &self.pipeline } pub fn bind_group_layouts(&self) -> &[wgpu::BindGroupLayout] { &self.bind_group_layouts } pub fn vertex_buffer(&self) -> &wgpu::Buffer { &self.vertex_buffer } pub fn record_draw<'a, 'b, P>( &'a self, pass: &mut wgpu::RenderPass<'a>, bump: &'a Bump, camera: &Camera, particles: P, ) where P: Iterator<Item = &'b Particle>, { use PushConstantUpdate::*; pass.set_pipeline(self.pipeline()); pass.set_vertex_buffer(0, self.vertex_buffer.slice(..)); pass.set_bind_group(0, &self.bind_group, &[]); // face toward camera let Angles { pitch, yaw, roll } = camera.angles(); let rotation = Angles { pitch: -pitch, yaw: -yaw, roll: -roll, } .mat4_wgpu(); for particle in particles { let q_origin = particle.origin(); let translation = Matrix4::from_translation([-q_origin.y, q_origin.z, -q_origin.x].into()); Self::set_push_constants( pass, Update(bump.alloc(VertexPushConstants { transform: camera.view_projection() * translation * rotation, })), Retain, Update(bump.alloc(FragmentPushConstants { color: particle.color() as u32, })), ); pass.draw(0..6, 0..1); } } } #[derive(Copy, Clone, Debug)] pub struct VertexPushConstants { pub transform: Matrix4<f32>, } #[derive(Copy, Clone, Debug)] pub struct FragmentPushConstants { pub color: u32, } const BIND_GROUP_LAYOUT_ENTRIES: &[wgpu::BindGroupLayoutEntry] = &[ wgpu::BindGroupLayoutEntry { binding: 0, visibility: wgpu::ShaderStage::FRAGMENT, ty: wgpu::BindingType::Sampler { filtering: true, comparison: false, }, count: None, }, // per-index texture array wgpu::BindGroupLayoutEntry { binding: 1, visibility: wgpu::ShaderStage::FRAGMENT, ty: wgpu::BindingType::Texture { view_dimension: wgpu::TextureViewDimension::D2, sample_type: wgpu::TextureSampleType::Float { filterable: true }, multisampled: false, }, count: NonZeroU32::new(256), }, ]; lazy_static! { static ref VERTEX_ATTRIBUTES: [[wgpu::VertexAttribute; 2]; 2] = [ wgpu::vertex_attr_array![ // position 0 => Float32x3, // texcoord 1 => Float32x2, ], wgpu::vertex_attr_array![ // instance position 2 => Float32x3, // color index 3 => Uint32, ] ]; } impl Pipeline for ParticlePipeline { type VertexPushConstants = VertexPushConstants; type SharedPushConstants = (); type FragmentPushConstants = FragmentPushConstants; fn name() -> &'static str { "particle" } fn vertex_shader() -> &'static str { include_str!(concat!( env!("CARGO_MANIFEST_DIR"), "/shaders/particle.vert" )) } fn fragment_shader() -> &'static str { include_str!(concat!( env!("CARGO_MANIFEST_DIR"), "/shaders/particle.frag" )) } // NOTE: if any of the binding indices are changed, they must also be changed in // the corresponding shaders and the BindGroupLayout generation functions. fn bind_group_layout_descriptors() -> Vec<wgpu::BindGroupLayoutDescriptor<'static>> { vec![ // group 0 wgpu::BindGroupLayoutDescriptor { label: Some("particle bind group layout"), entries: BIND_GROUP_LAYOUT_ENTRIES, }, ] } fn primitive_state() -> wgpu::PrimitiveState { WorldPipelineBase::primitive_state() } fn color_target_states() -> Vec<wgpu::ColorTargetState> { WorldPipelineBase::color_target_states() } fn depth_stencil_state() -> Option<wgpu::DepthStencilState> { let mut desc = WorldPipelineBase::depth_stencil_state().unwrap(); desc.depth_write_enabled = false; Some(desc)
pipeline
identifier_name
particle.rs
]; } #[rustfmt::skip] const PARTICLE_TEXTURE_PIXELS: [u8; 64] = [ 0, 0, 1, 1, 1, 1, 0, 0, 0, 1, 1, 1, 1, 1, 1, 0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 1, 1, 1, 1, 1, 1, 0, 0, 0, 1, 1, 1, 1, 0, 0, ]; pub struct ParticlePipeline { pipeline: wgpu::RenderPipeline, bind_group_layouts: Vec<wgpu::BindGroupLayout>, vertex_buffer: wgpu::Buffer, sampler: wgpu::Sampler, textures: Vec<wgpu::Texture>, texture_views: Vec<wgpu::TextureView>, bind_group: wgpu::BindGroup, } impl ParticlePipeline { pub fn new( device: &wgpu::Device, queue: &wgpu::Queue, compiler: &mut shaderc::Compiler, sample_count: u32, palette: &Palette, ) -> ParticlePipeline { let (pipeline, bind_group_layouts) = ParticlePipeline::create(device, compiler, &[], sample_count); use wgpu::util::DeviceExt as _; let vertex_buffer = device.create_buffer_init(&wgpu::util::BufferInitDescriptor { label: None, contents: unsafe { any_slice_as_bytes(&VERTICES) }, usage: wgpu::BufferUsage::VERTEX, }); let sampler = device.create_sampler(&wgpu::SamplerDescriptor { label: Some("particle sampler"), address_mode_u: wgpu::AddressMode::ClampToEdge, address_mode_v: wgpu::AddressMode::ClampToEdge, address_mode_w: wgpu::AddressMode::ClampToEdge, mag_filter: wgpu::FilterMode::Nearest, min_filter: wgpu::FilterMode::Linear, mipmap_filter: wgpu::FilterMode::Linear, lod_min_clamp: -1000.0, lod_max_clamp: 1000.0, compare: None, anisotropy_clamp: NonZeroU8::new(16), border_color: None, }); let textures: Vec<wgpu::Texture> = (0..256) .map(|i| { let mut pixels = PARTICLE_TEXTURE_PIXELS; // set up palette translation for pix in pixels.iter_mut() { if *pix == 0 { *pix = 0xFF; } else { *pix *= i as u8; } } let (diffuse_data, _) = palette.translate(&pixels); create_texture( device, queue, Some(&format!("particle texture {}", i)), 8, 8, &TextureData::Diffuse(diffuse_data), ) }) .collect(); let texture_views: Vec<wgpu::TextureView> = textures .iter() .map(|t| t.create_view(&Default::default())) .collect(); let texture_view_refs = texture_views.iter().collect::<Vec<_>>(); let bind_group = device.create_bind_group(&wgpu::BindGroupDescriptor { label: Some("particle bind group"), layout: &bind_group_layouts[0], entries: &[ wgpu::BindGroupEntry { binding: 0, resource: wgpu::BindingResource::Sampler(&sampler), }, wgpu::BindGroupEntry { binding: 1, resource: wgpu::BindingResource::TextureViewArray(&texture_view_refs[..]), }, ], }); ParticlePipeline { pipeline, bind_group_layouts, sampler, textures, texture_views, bind_group, vertex_buffer, } } pub fn rebuild( &mut self, device: &wgpu::Device, compiler: &mut shaderc::Compiler, sample_count: u32, ) { let layout_refs: Vec<_> = self.bind_group_layouts.iter().collect(); self.pipeline = ParticlePipeline::recreate(device, compiler, &layout_refs, sample_count); } pub fn pipeline(&self) -> &wgpu::RenderPipeline { &self.pipeline } pub fn bind_group_layouts(&self) -> &[wgpu::BindGroupLayout] { &self.bind_group_layouts } pub fn vertex_buffer(&self) -> &wgpu::Buffer { &self.vertex_buffer } pub fn record_draw<'a, 'b, P>( &'a self, pass: &mut wgpu::RenderPass<'a>, bump: &'a Bump, camera: &Camera, particles: P, ) where P: Iterator<Item = &'b Particle>, { use PushConstantUpdate::*; pass.set_pipeline(self.pipeline()); pass.set_vertex_buffer(0, self.vertex_buffer.slice(..)); pass.set_bind_group(0, &self.bind_group, &[]); // face toward camera let Angles { pitch, yaw, roll } = camera.angles(); let rotation = Angles { pitch: -pitch, yaw: -yaw, roll: -roll, } .mat4_wgpu(); for particle in particles { let q_origin = particle.origin(); let translation = Matrix4::from_translation([-q_origin.y, q_origin.z, -q_origin.x].into()); Self::set_push_constants( pass, Update(bump.alloc(VertexPushConstants { transform: camera.view_projection() * translation * rotation, })), Retain, Update(bump.alloc(FragmentPushConstants { color: particle.color() as u32, })), ); pass.draw(0..6, 0..1); } } } #[derive(Copy, Clone, Debug)] pub struct VertexPushConstants { pub transform: Matrix4<f32>, } #[derive(Copy, Clone, Debug)] pub struct FragmentPushConstants { pub color: u32,
visibility: wgpu::ShaderStage::FRAGMENT, ty: wgpu::BindingType::Sampler { filtering: true, comparison: false, }, count: None, }, // per-index texture array wgpu::BindGroupLayoutEntry { binding: 1, visibility: wgpu::ShaderStage::FRAGMENT, ty: wgpu::BindingType::Texture { view_dimension: wgpu::TextureViewDimension::D2, sample_type: wgpu::TextureSampleType::Float { filterable: true }, multisampled: false, }, count: NonZeroU32::new(256), }, ]; lazy_static! { static ref VERTEX_ATTRIBUTES: [[wgpu::VertexAttribute; 2]; 2] = [ wgpu::vertex_attr_array![ // position 0 => Float32x3, // texcoord 1 => Float32x2, ], wgpu::vertex_attr_array![ // instance position 2 => Float32x3, // color index 3 => Uint32, ] ]; } impl Pipeline for ParticlePipeline { type VertexPushConstants = VertexPushConstants; type SharedPushConstants = (); type FragmentPushConstants = FragmentPushConstants; fn name() -> &'static str { "particle" } fn vertex_shader() -> &'static str { include_str!(concat!( env!("CARGO_MANIFEST_DIR"), "/shaders/particle.vert" )) } fn fragment_shader() -> &'static str { include_str!(concat!( env!("CARGO_MANIFEST_DIR"), "/shaders/particle.frag" )) } // NOTE: if any of the binding indices are changed, they must also be changed in // the corresponding shaders and the BindGroupLayout generation functions. fn bind_group_layout_descriptors() -> Vec<wgpu::BindGroupLayoutDescriptor<'static>> { vec![ // group 0 wgpu::BindGroupLayoutDescriptor { label: Some("particle bind group layout"), entries: BIND_GROUP_LAYOUT_ENTRIES, }, ] } fn primitive_state() -> wgpu::PrimitiveState { WorldPipelineBase::primitive_state() } fn color_target_states() -> Vec<wgpu::ColorTargetState> { WorldPipelineBase::color_target_states() } fn depth_stencil_state() -> Option<wgpu::DepthStencilState> { let mut desc = WorldPipelineBase::depth_stencil_state().unwrap(); desc.depth_write_enabled = false; Some(desc)
} const BIND_GROUP_LAYOUT_ENTRIES: &[wgpu::BindGroupLayoutEntry] = &[ wgpu::BindGroupLayoutEntry { binding: 0,
random_line_split
particle.rs
]; } #[rustfmt::skip] const PARTICLE_TEXTURE_PIXELS: [u8; 64] = [ 0, 0, 1, 1, 1, 1, 0, 0, 0, 1, 1, 1, 1, 1, 1, 0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 1, 1, 1, 1, 1, 1, 0, 0, 0, 1, 1, 1, 1, 0, 0, ]; pub struct ParticlePipeline { pipeline: wgpu::RenderPipeline, bind_group_layouts: Vec<wgpu::BindGroupLayout>, vertex_buffer: wgpu::Buffer, sampler: wgpu::Sampler, textures: Vec<wgpu::Texture>, texture_views: Vec<wgpu::TextureView>, bind_group: wgpu::BindGroup, } impl ParticlePipeline { pub fn new( device: &wgpu::Device, queue: &wgpu::Queue, compiler: &mut shaderc::Compiler, sample_count: u32, palette: &Palette, ) -> ParticlePipeline { let (pipeline, bind_group_layouts) = ParticlePipeline::create(device, compiler, &[], sample_count); use wgpu::util::DeviceExt as _; let vertex_buffer = device.create_buffer_init(&wgpu::util::BufferInitDescriptor { label: None, contents: unsafe { any_slice_as_bytes(&VERTICES) }, usage: wgpu::BufferUsage::VERTEX, }); let sampler = device.create_sampler(&wgpu::SamplerDescriptor { label: Some("particle sampler"), address_mode_u: wgpu::AddressMode::ClampToEdge, address_mode_v: wgpu::AddressMode::ClampToEdge, address_mode_w: wgpu::AddressMode::ClampToEdge, mag_filter: wgpu::FilterMode::Nearest, min_filter: wgpu::FilterMode::Linear, mipmap_filter: wgpu::FilterMode::Linear, lod_min_clamp: -1000.0, lod_max_clamp: 1000.0, compare: None, anisotropy_clamp: NonZeroU8::new(16), border_color: None, }); let textures: Vec<wgpu::Texture> = (0..256) .map(|i| { let mut pixels = PARTICLE_TEXTURE_PIXELS; // set up palette translation for pix in pixels.iter_mut() { if *pix == 0 { *pix = 0xFF; } else { *pix *= i as u8; } } let (diffuse_data, _) = palette.translate(&pixels); create_texture( device, queue, Some(&format!("particle texture {}", i)), 8, 8, &TextureData::Diffuse(diffuse_data), ) }) .collect(); let texture_views: Vec<wgpu::TextureView> = textures .iter() .map(|t| t.create_view(&Default::default())) .collect(); let texture_view_refs = texture_views.iter().collect::<Vec<_>>(); let bind_group = device.create_bind_group(&wgpu::BindGroupDescriptor { label: Some("particle bind group"), layout: &bind_group_layouts[0], entries: &[ wgpu::BindGroupEntry { binding: 0, resource: wgpu::BindingResource::Sampler(&sampler), }, wgpu::BindGroupEntry { binding: 1, resource: wgpu::BindingResource::TextureViewArray(&texture_view_refs[..]), }, ], }); ParticlePipeline { pipeline, bind_group_layouts, sampler, textures, texture_views, bind_group, vertex_buffer, } } pub fn rebuild( &mut self, device: &wgpu::Device, compiler: &mut shaderc::Compiler, sample_count: u32, ) { let layout_refs: Vec<_> = self.bind_group_layouts.iter().collect(); self.pipeline = ParticlePipeline::recreate(device, compiler, &layout_refs, sample_count); } pub fn pipeline(&self) -> &wgpu::RenderPipeline { &self.pipeline } pub fn bind_group_layouts(&self) -> &[wgpu::BindGroupLayout] { &self.bind_group_layouts } pub fn vertex_buffer(&self) -> &wgpu::Buffer { &self.vertex_buffer } pub fn record_draw<'a, 'b, P>( &'a self, pass: &mut wgpu::RenderPass<'a>, bump: &'a Bump, camera: &Camera, particles: P, ) where P: Iterator<Item = &'b Particle>,
Self::set_push_constants( pass, Update(bump.alloc(VertexPushConstants { transform: camera.view_projection() * translation * rotation, })), Retain, Update(bump.alloc(FragmentPushConstants { color: particle.color() as u32, })), ); pass.draw(0..6, 0..1); } } } #[derive(Copy, Clone, Debug)] pub struct VertexPushConstants { pub transform: Matrix4<f32>, } #[derive(Copy, Clone, Debug)] pub struct FragmentPushConstants { pub color: u32, } const BIND_GROUP_LAYOUT_ENTRIES: &[wgpu::BindGroupLayoutEntry] = &[ wgpu::BindGroupLayoutEntry { binding: 0, visibility: wgpu::ShaderStage::FRAGMENT, ty: wgpu::BindingType::Sampler { filtering: true, comparison: false, }, count: None, }, // per-index texture array wgpu::BindGroupLayoutEntry { binding: 1, visibility: wgpu::ShaderStage::FRAGMENT, ty: wgpu::BindingType::Texture { view_dimension: wgpu::TextureViewDimension::D2, sample_type: wgpu::TextureSampleType::Float { filterable: true }, multisampled: false, }, count: NonZeroU32::new(256), }, ]; lazy_static! { static ref VERTEX_ATTRIBUTES: [[wgpu::VertexAttribute; 2]; 2] = [ wgpu::vertex_attr_array![ // position 0 => Float32x3, // texcoord 1 => Float32x2, ], wgpu::vertex_attr_array![ // instance position 2 => Float32x3, // color index 3 => Uint32, ] ]; } impl Pipeline for ParticlePipeline { type VertexPushConstants = VertexPushConstants; type SharedPushConstants = (); type FragmentPushConstants = FragmentPushConstants; fn name() -> &'static str { "particle" } fn vertex_shader() -> &'static str { include_str!(concat!( env!("CARGO_MANIFEST_DIR"), "/shaders/particle.vert" )) } fn fragment_shader() -> &'static str { include_str!(concat!( env!("CARGO_MANIFEST_DIR"), "/shaders/particle.frag" )) } // NOTE: if any of the binding indices are changed, they must also be changed in // the corresponding shaders and the BindGroupLayout generation functions. fn bind_group_layout_descriptors() -> Vec<wgpu::BindGroupLayoutDescriptor<'static>> { vec![ // group 0 wgpu::BindGroupLayoutDescriptor { label: Some("particle bind group layout"), entries: BIND_GROUP_LAYOUT_ENTRIES, }, ] } fn primitive_state() -> wgpu::PrimitiveState { WorldPipelineBase::primitive_state() } fn color_target_states() -> Vec<wgpu::ColorTargetState> { WorldPipelineBase::color_target_states() } fn depth_stencil_state() -> Option<wgpu::DepthStencilState> { let mut desc = WorldPipelineBase::depth_stencil_state().unwrap(); desc.depth_write_enabled = false; Some(desc)
{ use PushConstantUpdate::*; pass.set_pipeline(self.pipeline()); pass.set_vertex_buffer(0, self.vertex_buffer.slice(..)); pass.set_bind_group(0, &self.bind_group, &[]); // face toward camera let Angles { pitch, yaw, roll } = camera.angles(); let rotation = Angles { pitch: -pitch, yaw: -yaw, roll: -roll, } .mat4_wgpu(); for particle in particles { let q_origin = particle.origin(); let translation = Matrix4::from_translation([-q_origin.y, q_origin.z, -q_origin.x].into());
identifier_body
train_transformer.py
global_valid_step += 1 fout.write(' '.join(convert_ids_to_seq(output_seqs[0], vocab_bulider)) + '\n') if (i+1) % opt.logstep == 0: show_gen_seq(src, output_seqs, out_lens, tgt_gold, vocab_bulider, global_valid_step, mode='valid') # summary avg_loss = total_loss / i avg_bleu = bleu_score / i avg_distinct_1 = distinct_1_score / i avg_distinct_2 = distinct_2_score / i writer.log_loss(avg_loss, mode='valid') mylogger.log( i, epoch, model, value=avg_bleu, is_train=False, info=f'loss: {avg_loss:.4f} | ppl: {math.exp(avg_loss):.4f} | BLEU: {avg_bleu:.5f} | d1: {avg_distinct_1:.3f} | d2: {avg_distinct_2:.3f}') fout.close() def run_model(model, train_loader, eval_loader, niter, criterion, optimizer, scheduler): mylogger.log_info('Running Model') for i in range(niter): mylogger.log_info(f'EPOCH: {i}, lr: {optimizer.state_dict()["param_groups"][0]["lr"]}') train(i, model, train_loader, criterion, optimizer, scheduler) eval(i, model, eval_loader, criterion, beam_size=opt.beam) def convert_ids_to_seq(id_seq, vocab_bulider): return [vocab_bulider.id_to_word(idx) for idx in id_seq] def show_gen_seq(batch_in_seqs, batch_out_seqs, batch_out_lens, groud_truth, vocab_bulider, step, mode='train'): for in_id, out_id, out_len, gold_id in zip(batch_in_seqs, batch_out_seqs, batch_out_lens, groud_truth): in_seq = convert_ids_to_seq(in_id, vocab_bulider) out_seq = convert_ids_to_seq(out_id[:out_len] if out_len > 0 else out_id, vocab_bulider) gold_seq = convert_ids_to_seq(gold_id, vocab_bulider) writer.add_text(tag=mode + '_post', sentence=' '.join(in_seq[:get_index(in_seq, '<pad>')]), global_step=step) writer.add_text(tag=mode + '_pred', sentence=' '.join(out_seq), global_step=step) writer.add_text(tag=mode + '_reps', sentence=' '.join(gold_seq[:get_index(in_seq, '<pad>')]), global_step=step) if __name__ == '__main__': begin_time = time.strftime("%H%M%S", time.localtime()) model_name = 'transformer' + begin_time opt = parse_args() device = "cuda" if torch.cuda.is_available() else "cpu" torch.cuda.set_device(opt.gpuid) init_seed(opt.manualSeed) ACCUMULATION = opt.batchsize // opt.realbatch mylogger = LogManager(checkpoint_step=10, save_dir='./save', model_name=model_name, log_file_name=model_name + '.log', mode='max', device=device) mylogger.save_args(opt) writer = SummaryHelper(save_dir='./save', model_name=model_name) train_data_dir = './data/opensubtitles' # train_data_dir = './data/wmt15en-de' vocab_file_list = ['dialogue_length3_6.post'] # vocab_file_list = ['all_de-en.bpe.post', 'all_de-en.bpe.response'] vocab_bulider = VocabBulider( train_data_dir, src_files=vocab_file_list, ignore_unk_error=True, vocab_file='vocab.txt', min_count=opt.mincount, update=opt.update) print('most common 50:', vocab_bulider.most_common(50)) mylogger.log_info('vocab size: %d' % len(vocab_bulider)) # metircs bleu_metirc = BLEUMetric(vocab_bulider.id2vocab, ignore_smoothing_error=True) distinct_1 = DistinctNGram(ngram=1) distinct_2 = DistinctNGram(ngram=2) # train dataset and dataloader if opt.cotk: # use dataset in paper 'cotk' # opensub_file_name_list = ['all_de-en.bpe'] opensub_file_name_list = ['opensub_pair_dev', 'opensub_pair_test', 'opensub_pair_train'] unk_token = None else: # use dataset in paper 'Non-Autoregressive Neural Dialogue Generation' opensub_file_name_list = ['dialogue_length3_6'] unk_token = 'UNknown' opensub_dataset = OpenSubDataset( data_dir=train_data_dir, vocab_bulider=vocab_bulider, file_name_list=opensub_file_name_list, unk_token='UNknown', save_process=False, samples=opt.trainsamples, add_bos=True, add_eos=True) print(opensub_dataset.sample()) opensub_dataloader = DataLoader( opensub_dataset, batch_size=opt.realbatch, collate_fn=PadCollate(dim=0, pad_id=vocab_bulider.padid, device=device), shuffle=True, num_workers=opt.workers, drop_last=True) # dev set dev_data_dir = './data/imsdb' imsdb_file_name_list = ['imsdb_lower'] # dev_data_dir = './data/wmt15en-de' # imsdb_file_name_list = ['newstest'] imsdb_dataset = IMSDBDataset( data_dir=dev_data_dir, vocab_bulider=vocab_bulider, file_name_list=imsdb_file_name_list, save_process=False, samples=opt.validsamples, add_bos=True, add_eos=True) print(imsdb_dataset.sample()) imsdb_dataloader = DataLoader( imsdb_dataset, batch_size=1, collate_fn=PadCollate(dim=0, pad_id=vocab_bulider.padid, device=device), shuffle=False, num_workers=opt.workers, drop_last=True) # model definition if opt.mine: model = Transformer( ntoken=len(vocab_bulider), d_model=opt.embedsize, nhead=opt.nhead, num_encoder_layers=opt.encoderlayer, num_decoder_layers=opt.decoderlayer, dim_feedforward=opt.feedforward, postnorm=True, dropout=opt.dropout, gumbels=opt.gumbels, use_src_mask=False, use_tgt_mask=True, use_memory_mask=False, activation='relu', use_vocab_attn=False, use_pos_attn=False, relative_clip=0, highway=False, device=device, max_sent_length=32, share_input_output_embedding=False, share_encoder_decoder_embedding=True, share_vocab_embedding=True, fix_pos_encoding=opt.fix).to(device) else: model = TransformerTorch( ntoken=len(vocab_bulider), d_model=opt.embedsize, nhead=opt.nhead, num_encoder_layers=opt.encoderlayer, num_decoder_layers=opt.decoderlayer, dim_feedforward=opt.feedforward, postnorm=True, dropout=opt.dropout, gumbels=opt.gumbels, use_src_mask=False, use_tgt_mask=False, use_memory_mask=False, activation='relu', use_vocab_attn=False, use_pos_attn=False, relative_clip=0, highway=False, device=device, max_sent_length=32, share_input_output_embedding=False, share_encoder_decoder_embedding=True, share_vocab_embedding=True, fix_pos_encoding=opt.fix).to(device) model.show_graph() if opt.half: model = model.half() if opt.ft: model = restore_best_state(model, opt.ckpt, save_dir='./save', device=model.device) # optimizer and scheduler if opt.warmup: optimizer = RAdam( filter(lambda p: p.requires_grad, model.parameters()), lr=1., betas=(opt.beta1, opt.beta2), eps=opt.eps) rate_ratio = 1. / math.sqrt(opt.embedsize) # top_lr = 1 / sqrt(d_model * warmup_step) at step == warmup_step scheduler = optim.lr_scheduler.LambdaLR( optimizer, lr_lambda=lambda step: rate_ratio * min(1. / math.sqrt(step+1), step*(opt.warmup_step**(-1.5)))) else: optimizer = RAdam( filter(lambda p: p.requires_grad, model.parameters()), lr=opt.lr, betas=(opt.beta1, opt.beta2), eps=opt.eps, weight_decay=opt.weight_decay) scheduler = optim.lr_scheduler.StepLR( optimizer, step_size=opt.schedulerstep, gamma=opt.gamma) # loss function # criterion = nn.CrossEntropyLoss(ignore_index=vocab_bulider.padid) # for Transformer criterion = LabelSmoothedCrossEntropyLoss(eps=0.1, ignore_index=vocab_bulider.padid) # run model global_train_step, global_valid_step = 0, 0 run_model(
model, opensub_dataloader, imsdb_dataloader, opt.niter, criterion, optimizer, scheduler) writer.close()
random_line_split
train_transformer.py
distinct_2_score += distinct_2(out_seqs, tgt_lens) # summary writer global_train_step += 1 writer.log_loss(loss.item()*ACCUMULATION, mode='train') if (i+1) % ACCUMULATION == 0: # clip_grad_norm_(model.parameters(), max_norm=5) optimizer.step() optimizer.zero_grad() scheduler.step() if (i+1) % opt.logstep == 0: avg_loss = (total_loss / opt.logstep) * ACCUMULATION avg_bleu = bleu_score / opt.logstep avg_distinct_1 = distinct_1_score / opt.logstep avg_distinct_2 = distinct_2_score / opt.logstep mylogger.log( i, epoch, model, value=avg_loss, is_train=True, info=f'loss: {avg_loss:.4f} | ppl: {math.exp(avg_loss):.4f} | BLEU: {avg_bleu:.5f} | d1: {avg_distinct_1:.3f} | d2: {avg_distinct_2:.3f}') total_loss = 0. bleu_score = 0. distinct_1_score, distinct_2_score = 0., 0. show_gen_seq(src[:2], out_seqs[:2], tgt_lens[:2], tgt_gold[:2], vocab_bulider, global_train_step, mode='train') def eval(epoch, model, dataloader, criterion, beam_size=2): global global_valid_step model.eval() criterion.eval() total_loss = 0. bleu_score = 0. distinct_1_score, distinct_2_score = 0., 0. fout = open(os.path.join('./save/' + model_name + '/', model_name + '_' + str(epoch)), 'w', encoding='utf-8') with torch.no_grad(): for i, (src, tgt, src_lens, tgt_lens) in tqdm(enumerate(dataloader, 0), desc='eval', total=len(imsdb_dataset)): tgt_begin = torch.LongTensor([[vocab_bulider['<bos>']]]).to(device) tgt_gold = tgt[:, 1:] if beam_size > 1: output_seqs, output_probs = model.beam_search( src=src, tgt_begin=tgt_begin, src_length=src_lens, eos_token_id=vocab_bulider['<eos>'], beam_size=beam_size, max_length=tgt_lens.item()) else: output_seqs, output_probs = model.greedy( src=src, tgt_begin=tgt_begin, src_length=src_lens, eos_token_id=vocab_bulider['<eos>'], max_length=tgt_lens.item()) min_len = min(tgt_gold.shape[1], output_seqs.shape[1]) # loss loss = criterion(output_probs[:, :min_len, :].permute(0, 2, 1), tgt_gold[:, :min_len]) total_loss += loss.item() # calculate metrics out_lens = [min_len] bleu_score += bleu_metirc(tgt_gold, output_seqs, out_lens) distinct_1_score += distinct_1(output_seqs, out_lens) distinct_2_score += distinct_2(output_seqs, out_lens) # show sequence global_valid_step += 1 fout.write(' '.join(convert_ids_to_seq(output_seqs[0], vocab_bulider)) + '\n') if (i+1) % opt.logstep == 0: show_gen_seq(src, output_seqs, out_lens, tgt_gold, vocab_bulider, global_valid_step, mode='valid') # summary avg_loss = total_loss / i avg_bleu = bleu_score / i avg_distinct_1 = distinct_1_score / i avg_distinct_2 = distinct_2_score / i writer.log_loss(avg_loss, mode='valid') mylogger.log( i, epoch, model, value=avg_bleu, is_train=False, info=f'loss: {avg_loss:.4f} | ppl: {math.exp(avg_loss):.4f} | BLEU: {avg_bleu:.5f} | d1: {avg_distinct_1:.3f} | d2: {avg_distinct_2:.3f}') fout.close() def run_model(model, train_loader, eval_loader, niter, criterion, optimizer, scheduler): mylogger.log_info('Running Model') for i in range(niter): mylogger.log_info(f'EPOCH: {i}, lr: {optimizer.state_dict()["param_groups"][0]["lr"]}') train(i, model, train_loader, criterion, optimizer, scheduler) eval(i, model, eval_loader, criterion, beam_size=opt.beam) def convert_ids_to_seq(id_seq, vocab_bulider): return [vocab_bulider.id_to_word(idx) for idx in id_seq] def show_gen_seq(batch_in_seqs, batch_out_seqs, batch_out_lens, groud_truth, vocab_bulider, step, mode='train'): for in_id, out_id, out_len, gold_id in zip(batch_in_seqs, batch_out_seqs, batch_out_lens, groud_truth): in_seq = convert_ids_to_seq(in_id, vocab_bulider) out_seq = convert_ids_to_seq(out_id[:out_len] if out_len > 0 else out_id, vocab_bulider) gold_seq = convert_ids_to_seq(gold_id, vocab_bulider) writer.add_text(tag=mode + '_post', sentence=' '.join(in_seq[:get_index(in_seq, '<pad>')]), global_step=step) writer.add_text(tag=mode + '_pred', sentence=' '.join(out_seq), global_step=step) writer.add_text(tag=mode + '_reps', sentence=' '.join(gold_seq[:get_index(in_seq, '<pad>')]), global_step=step) if __name__ == '__main__': begin_time = time.strftime("%H%M%S", time.localtime()) model_name = 'transformer' + begin_time opt = parse_args() device = "cuda" if torch.cuda.is_available() else "cpu" torch.cuda.set_device(opt.gpuid) init_seed(opt.manualSeed) ACCUMULATION = opt.batchsize // opt.realbatch mylogger = LogManager(checkpoint_step=10, save_dir='./save', model_name=model_name, log_file_name=model_name + '.log', mode='max', device=device) mylogger.save_args(opt) writer = SummaryHelper(save_dir='./save', model_name=model_name) train_data_dir = './data/opensubtitles' # train_data_dir = './data/wmt15en-de' vocab_file_list = ['dialogue_length3_6.post'] # vocab_file_list = ['all_de-en.bpe.post', 'all_de-en.bpe.response'] vocab_bulider = VocabBulider( train_data_dir, src_files=vocab_file_list, ignore_unk_error=True, vocab_file='vocab.txt', min_count=opt.mincount, update=opt.update) print('most common 50:', vocab_bulider.most_common(50)) mylogger.log_info('vocab size: %d' % len(vocab_bulider)) # metircs bleu_metirc = BLEUMetric(vocab_bulider.id2vocab, ignore_smoothing_error=True) distinct_1 = DistinctNGram(ngram=1) distinct_2 = DistinctNGram(ngram=2) # train dataset and dataloader if opt.cotk: # use dataset in paper 'cotk' # opensub_file_name_list = ['all_de-en.bpe'] opensub_file_name_list = ['opensub_pair_dev', 'opensub_pair_test', 'opensub_pair_train'] unk_token = None else: # use dataset in paper 'Non-Autoregressive Neural Dialogue Generation' opensub_file_name_list = ['dialogue_length3_6'] unk_token = 'UNknown' opensub_dataset = OpenSubDataset( data_dir=train_data_dir, vocab_bulider
global global_train_step model.train() total_loss = 0. bleu_score = 0. distinct_1_score, distinct_2_score = 0., 0. for i, (src, tgt, src_lens, tgt_lens) in tqdm(enumerate(dataloader, 0), desc='train', total=len(opensub_dataset)//opt.realbatch): tgt_input = tgt[:, :-1] tgt_gold = tgt[:, 1:] tgt_lens = tgt_lens - 1 decoder_output_probs, _ = model( src=src, tgt=tgt_input, src_lengths=src_lens, tgt_lengths=tgt_lens) decoder_output_probs_T = decoder_output_probs.permute(0, 2, 1) out_seqs = torch.argmax(decoder_output_probs, dim=2) # loss loss = criterion(decoder_output_probs_T, tgt_gold) / ACCUMULATION loss.backward() total_loss += loss.item() # calculate metrics bleu_score += bleu_metirc(tgt_gold, out_seqs, tgt_lens) distinct_1_score += distinct_1(out_seqs, tgt_lens)
identifier_body
train_transformer.py
=2) # loss loss = criterion(decoder_output_probs_T, tgt_gold) / ACCUMULATION loss.backward() total_loss += loss.item() # calculate metrics bleu_score += bleu_metirc(tgt_gold, out_seqs, tgt_lens) distinct_1_score += distinct_1(out_seqs, tgt_lens) distinct_2_score += distinct_2(out_seqs, tgt_lens) # summary writer global_train_step += 1 writer.log_loss(loss.item()*ACCUMULATION, mode='train') if (i+1) % ACCUMULATION == 0: # clip_grad_norm_(model.parameters(), max_norm=5) optimizer.step() optimizer.zero_grad() scheduler.step() if (i+1) % opt.logstep == 0: avg_loss = (total_loss / opt.logstep) * ACCUMULATION avg_bleu = bleu_score / opt.logstep avg_distinct_1 = distinct_1_score / opt.logstep avg_distinct_2 = distinct_2_score / opt.logstep mylogger.log( i, epoch, model, value=avg_loss, is_train=True, info=f'loss: {avg_loss:.4f} | ppl: {math.exp(avg_loss):.4f} | BLEU: {avg_bleu:.5f} | d1: {avg_distinct_1:.3f} | d2: {avg_distinct_2:.3f}') total_loss = 0. bleu_score = 0. distinct_1_score, distinct_2_score = 0., 0. show_gen_seq(src[:2], out_seqs[:2], tgt_lens[:2], tgt_gold[:2], vocab_bulider, global_train_step, mode='train') def eval(epoch, model, dataloader, criterion, beam_size=2): global global_valid_step model.eval() criterion.eval() total_loss = 0. bleu_score = 0. distinct_1_score, distinct_2_score = 0., 0. fout = open(os.path.join('./save/' + model_name + '/', model_name + '_' + str(epoch)), 'w', encoding='utf-8') with torch.no_grad(): for i, (src, tgt, src_lens, tgt_lens) in tqdm(enumerate(dataloader, 0), desc='eval', total=len(imsdb_dataset)): tgt_begin = torch.LongTensor([[vocab_bulider['<bos>']]]).to(device) tgt_gold = tgt[:, 1:] if beam_size > 1: output_seqs, output_probs = model.beam_search( src=src, tgt_begin=tgt_begin, src_length=src_lens, eos_token_id=vocab_bulider['<eos>'], beam_size=beam_size, max_length=tgt_lens.item()) else: output_seqs, output_probs = model.greedy( src=src, tgt_begin=tgt_begin, src_length=src_lens, eos_token_id=vocab_bulider['<eos>'], max_length=tgt_lens.item()) min_len = min(tgt_gold.shape[1], output_seqs.shape[1]) # loss loss = criterion(output_probs[:, :min_len, :].permute(0, 2, 1), tgt_gold[:, :min_len]) total_loss += loss.item() # calculate metrics out_lens = [min_len] bleu_score += bleu_metirc(tgt_gold, output_seqs, out_lens) distinct_1_score += distinct_1(output_seqs, out_lens) distinct_2_score += distinct_2(output_seqs, out_lens) # show sequence global_valid_step += 1 fout.write(' '.join(convert_ids_to_seq(output_seqs[0], vocab_bulider)) + '\n') if (i+1) % opt.logstep == 0: show_gen_seq(src, output_seqs, out_lens, tgt_gold, vocab_bulider, global_valid_step, mode='valid') # summary avg_loss = total_loss / i avg_bleu = bleu_score / i avg_distinct_1 = distinct_1_score / i avg_distinct_2 = distinct_2_score / i writer.log_loss(avg_loss, mode='valid') mylogger.log( i, epoch, model, value=avg_bleu, is_train=False, info=f'loss: {avg_loss:.4f} | ppl: {math.exp(avg_loss):.4f} | BLEU: {avg_bleu:.5f} | d1: {avg_distinct_1:.3f} | d2: {avg_distinct_2:.3f}') fout.close() def
(model, train_loader, eval_loader, niter, criterion, optimizer, scheduler): mylogger.log_info('Running Model') for i in range(niter): mylogger.log_info(f'EPOCH: {i}, lr: {optimizer.state_dict()["param_groups"][0]["lr"]}') train(i, model, train_loader, criterion, optimizer, scheduler) eval(i, model, eval_loader, criterion, beam_size=opt.beam) def convert_ids_to_seq(id_seq, vocab_bulider): return [vocab_bulider.id_to_word(idx) for idx in id_seq] def show_gen_seq(batch_in_seqs, batch_out_seqs, batch_out_lens, groud_truth, vocab_bulider, step, mode='train'): for in_id, out_id, out_len, gold_id in zip(batch_in_seqs, batch_out_seqs, batch_out_lens, groud_truth): in_seq = convert_ids_to_seq(in_id, vocab_bulider) out_seq = convert_ids_to_seq(out_id[:out_len] if out_len > 0 else out_id, vocab_bulider) gold_seq = convert_ids_to_seq(gold_id, vocab_bulider) writer.add_text(tag=mode + '_post', sentence=' '.join(in_seq[:get_index(in_seq, '<pad>')]), global_step=step) writer.add_text(tag=mode + '_pred', sentence=' '.join(out_seq), global_step=step) writer.add_text(tag=mode + '_reps', sentence=' '.join(gold_seq[:get_index(in_seq, '<pad>')]), global_step=step) if __name__ == '__main__': begin_time = time.strftime("%H%M%S", time.localtime()) model_name = 'transformer' + begin_time opt = parse_args() device = "cuda" if torch.cuda.is_available() else "cpu" torch.cuda.set_device(opt.gpuid) init_seed(opt.manualSeed) ACCUMULATION = opt.batchsize // opt.realbatch mylogger = LogManager(checkpoint_step=10, save_dir='./save', model_name=model_name, log_file_name=model_name + '.log', mode='max', device=device) mylogger.save_args(opt) writer = SummaryHelper(save_dir='./save', model_name=model_name) train_data_dir = './data/opensubtitles' # train_data_dir = './data/wmt15en-de' vocab_file_list = ['dialogue_length3_6.post'] # vocab_file_list = ['all_de-en.bpe.post', 'all_de-en.bpe.response'] vocab_bulider = VocabBulider( train_data_dir, src_files=vocab_file_list, ignore_unk_error=True, vocab_file='vocab.txt', min_count=opt.mincount, update=opt.update) print('most common 50:', vocab_bulider.most_common(50)) mylogger.log_info('vocab size: %d' % len(vocab_bulider)) # metircs bleu_metirc = BLEUMetric(vocab_bulider.id2vocab, ignore_smoothing_error=True) distinct_1 = DistinctNGram(ngram=1) distinct_2 = DistinctNGram(ngram=2) # train dataset and dataloader if opt.cotk: # use dataset in paper 'cotk' # opensub_file_name_list = ['all_de-en.bpe'] opensub_file_name_list = ['opensub_pair_dev', 'opensub_pair_test', 'opensub_pair_train'] unk_token = None else: # use dataset in paper 'Non-Autoregressive Neural Dialogue Generation' opensub_file_name_list = ['dialogue_length3_6'] unk_token = 'UNknown' opensub_dataset = OpenSubDataset( data_dir=train_data_dir, vocab_bulider=vocab_bulider, file_name_list=opensub_file_name_list, unk_token='UNknown', save_process=False, samples=opt.trainsamples, add_bos=True, add_eos=True) print(opensub_dataset.sample()) opensub_dataloader = DataLoader( opensub_dataset, batch_size=opt.realbatch, collate_fn=PadCollate(dim=0, pad_id=vocab_bulider.padid, device=device), shuffle=True, num_workers=opt.workers, drop_last=True) # dev set dev_data_dir = './data/imsdb' imsdb_file_name_list = ['imsdb_lower'] # dev_data_dir = './data/wmt15en-de' # imsdb_file_name_list = ['newstest'] imsdb_dataset
run_model
identifier_name
train_transformer.py
=2) # loss loss = criterion(decoder_output_probs_T, tgt_gold) / ACCUMULATION loss.backward() total_loss += loss.item() # calculate metrics bleu_score += bleu_metirc(tgt_gold, out_seqs, tgt_lens) distinct_1_score += distinct_1(out_seqs, tgt_lens) distinct_2_score += distinct_2(out_seqs, tgt_lens) # summary writer global_train_step += 1 writer.log_loss(loss.item()*ACCUMULATION, mode='train') if (i+1) % ACCUMULATION == 0: # clip_grad_norm_(model.parameters(), max_norm=5) optimizer.step() optimizer.zero_grad() scheduler.step() if (i+1) % opt.logstep == 0:
def eval(epoch, model, dataloader, criterion, beam_size=2): global global_valid_step model.eval() criterion.eval() total_loss = 0. bleu_score = 0. distinct_1_score, distinct_2_score = 0., 0. fout = open(os.path.join('./save/' + model_name + '/', model_name + '_' + str(epoch)), 'w', encoding='utf-8') with torch.no_grad(): for i, (src, tgt, src_lens, tgt_lens) in tqdm(enumerate(dataloader, 0), desc='eval', total=len(imsdb_dataset)): tgt_begin = torch.LongTensor([[vocab_bulider['<bos>']]]).to(device) tgt_gold = tgt[:, 1:] if beam_size > 1: output_seqs, output_probs = model.beam_search( src=src, tgt_begin=tgt_begin, src_length=src_lens, eos_token_id=vocab_bulider['<eos>'], beam_size=beam_size, max_length=tgt_lens.item()) else: output_seqs, output_probs = model.greedy( src=src, tgt_begin=tgt_begin, src_length=src_lens, eos_token_id=vocab_bulider['<eos>'], max_length=tgt_lens.item()) min_len = min(tgt_gold.shape[1], output_seqs.shape[1]) # loss loss = criterion(output_probs[:, :min_len, :].permute(0, 2, 1), tgt_gold[:, :min_len]) total_loss += loss.item() # calculate metrics out_lens = [min_len] bleu_score += bleu_metirc(tgt_gold, output_seqs, out_lens) distinct_1_score += distinct_1(output_seqs, out_lens) distinct_2_score += distinct_2(output_seqs, out_lens) # show sequence global_valid_step += 1 fout.write(' '.join(convert_ids_to_seq(output_seqs[0], vocab_bulider)) + '\n') if (i+1) % opt.logstep == 0: show_gen_seq(src, output_seqs, out_lens, tgt_gold, vocab_bulider, global_valid_step, mode='valid') # summary avg_loss = total_loss / i avg_bleu = bleu_score / i avg_distinct_1 = distinct_1_score / i avg_distinct_2 = distinct_2_score / i writer.log_loss(avg_loss, mode='valid') mylogger.log( i, epoch, model, value=avg_bleu, is_train=False, info=f'loss: {avg_loss:.4f} | ppl: {math.exp(avg_loss):.4f} | BLEU: {avg_bleu:.5f} | d1: {avg_distinct_1:.3f} | d2: {avg_distinct_2:.3f}') fout.close() def run_model(model, train_loader, eval_loader, niter, criterion, optimizer, scheduler): mylogger.log_info('Running Model') for i in range(niter): mylogger.log_info(f'EPOCH: {i}, lr: {optimizer.state_dict()["param_groups"][0]["lr"]}') train(i, model, train_loader, criterion, optimizer, scheduler) eval(i, model, eval_loader, criterion, beam_size=opt.beam) def convert_ids_to_seq(id_seq, vocab_bulider): return [vocab_bulider.id_to_word(idx) for idx in id_seq] def show_gen_seq(batch_in_seqs, batch_out_seqs, batch_out_lens, groud_truth, vocab_bulider, step, mode='train'): for in_id, out_id, out_len, gold_id in zip(batch_in_seqs, batch_out_seqs, batch_out_lens, groud_truth): in_seq = convert_ids_to_seq(in_id, vocab_bulider) out_seq = convert_ids_to_seq(out_id[:out_len] if out_len > 0 else out_id, vocab_bulider) gold_seq = convert_ids_to_seq(gold_id, vocab_bulider) writer.add_text(tag=mode + '_post', sentence=' '.join(in_seq[:get_index(in_seq, '<pad>')]), global_step=step) writer.add_text(tag=mode + '_pred', sentence=' '.join(out_seq), global_step=step) writer.add_text(tag=mode + '_reps', sentence=' '.join(gold_seq[:get_index(in_seq, '<pad>')]), global_step=step) if __name__ == '__main__': begin_time = time.strftime("%H%M%S", time.localtime()) model_name = 'transformer' + begin_time opt = parse_args() device = "cuda" if torch.cuda.is_available() else "cpu" torch.cuda.set_device(opt.gpuid) init_seed(opt.manualSeed) ACCUMULATION = opt.batchsize // opt.realbatch mylogger = LogManager(checkpoint_step=10, save_dir='./save', model_name=model_name, log_file_name=model_name + '.log', mode='max', device=device) mylogger.save_args(opt) writer = SummaryHelper(save_dir='./save', model_name=model_name) train_data_dir = './data/opensubtitles' # train_data_dir = './data/wmt15en-de' vocab_file_list = ['dialogue_length3_6.post'] # vocab_file_list = ['all_de-en.bpe.post', 'all_de-en.bpe.response'] vocab_bulider = VocabBulider( train_data_dir, src_files=vocab_file_list, ignore_unk_error=True, vocab_file='vocab.txt', min_count=opt.mincount, update=opt.update) print('most common 50:', vocab_bulider.most_common(50)) mylogger.log_info('vocab size: %d' % len(vocab_bulider)) # metircs bleu_metirc = BLEUMetric(vocab_bulider.id2vocab, ignore_smoothing_error=True) distinct_1 = DistinctNGram(ngram=1) distinct_2 = DistinctNGram(ngram=2) # train dataset and dataloader if opt.cotk: # use dataset in paper 'cotk' # opensub_file_name_list = ['all_de-en.bpe'] opensub_file_name_list = ['opensub_pair_dev', 'opensub_pair_test', 'opensub_pair_train'] unk_token = None else: # use dataset in paper 'Non-Autoregressive Neural Dialogue Generation' opensub_file_name_list = ['dialogue_length3_6'] unk_token = 'UNknown' opensub_dataset = OpenSubDataset( data_dir=train_data_dir, vocab_bulider=vocab_bulider, file_name_list=opensub_file_name_list, unk_token='UNknown', save_process=False, samples=opt.trainsamples, add_bos=True, add_eos=True) print(opensub_dataset.sample()) opensub_dataloader = DataLoader( opensub_dataset, batch_size=opt.realbatch, collate_fn=PadCollate(dim=0, pad_id=vocab_bulider.padid, device=device), shuffle=True, num_workers=opt.workers, drop_last=True) # dev set dev_data_dir = './data/imsdb' imsdb_file_name_list = ['imsdb_lower'] # dev_data_dir = './data/wmt15en-de' # imsdb_file_name_list = ['newstest'] imsdb_dataset =
avg_loss = (total_loss / opt.logstep) * ACCUMULATION avg_bleu = bleu_score / opt.logstep avg_distinct_1 = distinct_1_score / opt.logstep avg_distinct_2 = distinct_2_score / opt.logstep mylogger.log( i, epoch, model, value=avg_loss, is_train=True, info=f'loss: {avg_loss:.4f} | ppl: {math.exp(avg_loss):.4f} | BLEU: {avg_bleu:.5f} | d1: {avg_distinct_1:.3f} | d2: {avg_distinct_2:.3f}') total_loss = 0. bleu_score = 0. distinct_1_score, distinct_2_score = 0., 0. show_gen_seq(src[:2], out_seqs[:2], tgt_lens[:2], tgt_gold[:2], vocab_bulider, global_train_step, mode='train')
conditional_block
p21_sim.py
693 #DUMMY omega_MDM2 = 8.318 #DUMMY omega_CD = 0.023 #DUMMY omega_CDInk4 = 1 #DUMMY omega_CE = 1 #DUMMY omega_CA = 1 #DUMMY omega_CACDC20 = 1 #DUMMY omega_CB = 1 #DUMMY omega_CBCDC20 = 1 #DUMMY beta_E2FRb = 2 #DUMMY beta_E2FRbMP = beta_E2FRb * 0.3 #DUMMY, but should be smaller than beta_E2FRb beta_mdm2p14 = 1 #DUMMY beta_cp21 = 1 #DUMMY beta_E7p21 = 1 #DUMMY beta_E7pRB = 1 #DUMMY beta_E7pRBP = 1 #DUMMY beta_E7pRBPP = 1 #DUMMY delta_E7p21 = 1 #DUMMY delta_E7pRB = 1 #DUMMY epsilon_RbCD = 0.4 #DUMMY epsilon_RbCE = 0.7 #DUMMY epsilon_E2F = 20 #DUMMY epsilon_CDC20 = 1 #DUMMY sigma_Rb = 0.3 #DUMMY sigma_RbMP = 0.1 #DUMMY sigma_E2F = 0.7 #DUMMY sigma_CDC20 = 1 #DUMMY kappa_CECA = 1 #DUMMY kappa_CBCA = 1 #DUMMY #Knockdown terms. These should all be equal to 1 unless we're testing the #effects of a knockdown. theta_E2F = 1 theta_p53 = 1 theta_CE = 1 theta_MDM2 = 1 theta_CD = 1 theta_CA = 1 theta_Ink4 = 1 theta_CB = 1 theta_CDC20 = 1 k_p21 = 1 k_p53 = 1 k_RbMP = 1 k_RbPP = 1 k_RbCD = 1 k_RbCE = 1 k_E2FCA = 1 k_E2F = 1 k_CD = 1 k_CA = 1 k_CB = 1 k_CDC20CB = 1 k_CDC20 = 1 E2F_tot = 10 #DUMMY CDC20_tot = 0.285 Rb_tot = 10 #DUMMY #Potentially override parameters for infile in infiles: reader = open(infile) for line in reader.readlines(): exec(line) #Dummy initial conditions y0 = [0.1,0.1,0.1,0.1,Rb_tot,0.0,0.0,0.1,0.1,0.1,0.1,0.1,0.1,0.1,0.1,0,0] #Abort early if the output directory already exists. if dirname != "": if os.path.exists(dirname): sys.stderr.write("Output dir " + dirname + " already exists. Aborting.\n") sys.exit(2) #Functions to be called from the derivative functions. def
(t): return 0 #dummy def E7(t): return 0 #dummy #Fractions for inhibition/inactive complex formation. #Each has a sanity check in case chief input is near zero. sanity_threshold = 0.00001 def f(e2f, rb, rbmp): if e2f < sanity_threshold: return 0 return e2f**2 / (e2f + beta_E2FRb * rb + beta_E2FRbMP * rbmp) def g(mdm2, p14): if mdm2 < sanity_threshold: return 0 return mdm2**2 / (mdm2 + beta_mdm2p14 * p14) def h(c, p21, cd, ce, ca, cb): if c < sanity_threshold: return 0 if ca+cb+cd+ce < sanity_threshold: return 0 return c**2 / (c + beta_cp21 * p21 * c / (ca+cb+cd+ce)) def j(e7, e7prb, e7p21): return e7 - e7prb - e7p21 #Variable key #y[0] = p14 #y[1] = Ink4 #y[2] = p21 #y[3] = p53 #y[4] = Rb #y[5] = pRb-P #y[6] = pRb-PP #y[7] = E2F #y[8] = mdm2 #y[9] = pMDM2 #y[10] = CD (Cyclin D/CDK4-6 complex) #y[11] = CE (Cyclin E/CDK2 complex) #y[12] = CA (Cyclin A/CDK2 complex) #y[13] = CB (Cyclin B/CDK1 complex) #y[14] = CDC20 #y[15] = E7-pRB #y[16] = E7-p21 names = [] names.append("p14") names.append("Ink4") names.append("p21") names.append("p53") names.append("pRb") names.append("pRb-P") names.append("pRb-PP") names.append("E2F") names.append("mdm2") names.append("pMDM2") names.append("CD") names.append("CE") names.append("CA") names.append("CB") names.append("CDC20") names.append("E7-pRB") names.append("E7-p21") #The derivative function for the differential equation system. def func(y,t): return [ #We have p14 being produced by E2F after inhibition from Rb #is accounted for, and degraded at a constant rate. alpha_p14 * theta_E2F * f(y[7], y[4], y[5]) - omega_p14 * y[0], #It's just like the p14 equation, but with Ink4 instead! alpha_Ink4 * theta_E2F * f(y[7], y[4], y[5]) - omega_Ink4 * y[1], #Form p21 at a rate proportional to p53 presence; degrade it #"naturally" or with help from Cyclin E/CDK2. E7 sequesters p21. alpha_p21 * theta_p53 * y[3] - omega_p21 * y[2] - omega_p21CE * theta_CE * y[11] * y[2]/(y[2]+k_p21) - beta_E7p21 * j(E7(t),y[15],y[16]) * y[2] + delta_E7p21 * y[16], #P53 is generated naturally at a constant rate, and degrades #both on its own and with help from MDM2. alpha_p53 - omega_p53 * y[3] - (omega_p53MDM2 * theta_MDM2 * g(y[9], y[0]) + omega_p53E6 * E6(t)) * y[3]/(y[3]+k_p53), #Rb gets monophosphorylated by Cyclin D/CDK4-6. Rb-monophosphate #gets its phosphate cleaved at a constant rate. Rb of all sorts #gets sequestered by E7. -epsilon_RbCD * theta_CD * y[4]/(y[4]+k_RbCD) * y[10] + sigma_Rb * y[5]/(y[5]+k_RbMP) - beta_E7pRB * j(E7(t),y[15],y[16]) * y[4] + delta_E7pRB * y[15] * y[4]/Rb_tot, #Rb-monophosphate can be formed by phosphorylation of Rb or cleavage #of Rb-polyphosphate. It can be lost by Cyclin E/CDK2 or #phosphatase activity. Rb of all sorts gets sequestered by E7. epsilon_RbCD * theta_CD * y[4]/(y[4]+k_RbCD) * y[10] - sigma_Rb * y[5]/(y[5]+k_RbMP) - epsilon_RbCE * theta_CE * y[5]/(y[5]+k_RbCE) * y[11] + sigma_Rb
E6
identifier_name
p21_sim.py
693 #DUMMY omega_MDM2 = 8.318 #DUMMY omega_CD = 0.023 #DUMMY omega_CDInk4 = 1 #DUMMY omega_CE = 1 #DUMMY omega_CA = 1 #DUMMY omega_CACDC20 = 1 #DUMMY omega_CB = 1 #DUMMY omega_CBCDC20 = 1 #DUMMY beta_E2FRb = 2 #DUMMY beta_E2FRbMP = beta_E2FRb * 0.3 #DUMMY, but should be smaller than beta_E2FRb beta_mdm2p14 = 1 #DUMMY beta_cp21 = 1 #DUMMY beta_E7p21 = 1 #DUMMY beta_E7pRB = 1 #DUMMY beta_E7pRBP = 1 #DUMMY beta_E7pRBPP = 1 #DUMMY delta_E7p21 = 1 #DUMMY delta_E7pRB = 1 #DUMMY epsilon_RbCD = 0.4 #DUMMY epsilon_RbCE = 0.7 #DUMMY epsilon_E2F = 20 #DUMMY epsilon_CDC20 = 1 #DUMMY sigma_Rb = 0.3 #DUMMY sigma_RbMP = 0.1 #DUMMY sigma_E2F = 0.7 #DUMMY sigma_CDC20 = 1 #DUMMY kappa_CECA = 1 #DUMMY kappa_CBCA = 1 #DUMMY #Knockdown terms. These should all be equal to 1 unless we're testing the #effects of a knockdown. theta_E2F = 1 theta_p53 = 1 theta_CE = 1 theta_MDM2 = 1 theta_CD = 1 theta_CA = 1 theta_Ink4 = 1 theta_CB = 1 theta_CDC20 = 1 k_p21 = 1 k_p53 = 1 k_RbMP = 1 k_RbPP = 1 k_RbCD = 1 k_RbCE = 1 k_E2FCA = 1 k_E2F = 1 k_CD = 1 k_CA = 1 k_CB = 1 k_CDC20CB = 1 k_CDC20 = 1 E2F_tot = 10 #DUMMY CDC20_tot = 0.285 Rb_tot = 10 #DUMMY #Potentially override parameters for infile in infiles: reader = open(infile) for line in reader.readlines(): exec(line) #Dummy initial conditions y0 = [0.1,0.1,0.1,0.1,Rb_tot,0.0,0.0,0.1,0.1,0.1,0.1,0.1,0.1,0.1,0.1,0,0] #Abort early if the output directory already exists. if dirname != "": if os.path.exists(dirname): sys.stderr.write("Output dir " + dirname + " already exists. Aborting.\n") sys.exit(2) #Functions to be called from the derivative functions. def E6(t): return 0 #dummy def E7(t): return 0 #dummy #Fractions for inhibition/inactive complex formation. #Each has a sanity check in case chief input is near zero. sanity_threshold = 0.00001 def f(e2f, rb, rbmp): if e2f < sanity_threshold: return 0 return e2f**2 / (e2f + beta_E2FRb * rb + beta_E2FRbMP * rbmp) def g(mdm2, p14): if mdm2 < sanity_threshold: return 0 return mdm2**2 / (mdm2 + beta_mdm2p14 * p14) def h(c, p21, cd, ce, ca, cb): if c < sanity_threshold: return 0 if ca+cb+cd+ce < sanity_threshold: return 0 return c**2 / (c + beta_cp21 * p21 * c / (ca+cb+cd+ce))
def j(e7, e7prb, e7p21): return e7 - e7prb - e7p21 #Variable key #y[0] = p14 #y[1] = Ink4 #y[2] = p21 #y[3] = p53 #y[4] = Rb #y[5] = pRb-P #y[6] = pRb-PP #y[7] = E2F #y[8] = mdm2 #y[9] = pMDM2 #y[10] = CD (Cyclin D/CDK4-6 complex) #y[11] = CE (Cyclin E/CDK2 complex) #y[12] = CA (Cyclin A/CDK2 complex) #y[13] = CB (Cyclin B/CDK1 complex) #y[14] = CDC20 #y[15] = E7-pRB #y[16] = E7-p21 names = [] names.append("p14") names.append("Ink4") names.append("p21") names.append("p53") names.append("pRb") names.append("pRb-P") names.append("pRb-PP") names.append("E2F") names.append("mdm2") names.append("pMDM2") names.append("CD") names.append("CE") names.append("CA") names.append("CB") names.append("CDC20") names.append("E7-pRB") names.append("E7-p21") #The derivative function for the differential equation system. def func(y,t): return [ #We have p14 being produced by E2F after inhibition from Rb #is accounted for, and degraded at a constant rate. alpha_p14 * theta_E2F * f(y[7], y[4], y[5]) - omega_p14 * y[0], #It's just like the p14 equation, but with Ink4 instead! alpha_Ink4 * theta_E2F * f(y[7], y[4], y[5]) - omega_Ink4 * y[1], #Form p21 at a rate proportional to p53 presence; degrade it #"naturally" or with help from Cyclin E/CDK2. E7 sequesters p21. alpha_p21 * theta_p53 * y[3] - omega_p21 * y[2] - omega_p21CE * theta_CE * y[11] * y[2]/(y[2]+k_p21) - beta_E7p21 * j(E7(t),y[15],y[16]) * y[2] + delta_E7p21 * y[16], #P53 is generated naturally at a constant rate, and degrades #both on its own and with help from MDM2. alpha_p53 - omega_p53 * y[3] - (omega_p53MDM2 * theta_MDM2 * g(y[9], y[0]) + omega_p53E6 * E6(t)) * y[3]/(y[3]+k_p53), #Rb gets monophosphorylated by Cyclin D/CDK4-6. Rb-monophosphate #gets its phosphate cleaved at a constant rate. Rb of all sorts #gets sequestered by E7. -epsilon_RbCD * theta_CD * y[4]/(y[4]+k_RbCD) * y[10] + sigma_Rb * y[5]/(y[5]+k_RbMP) - beta_E7pRB * j(E7(t),y[15],y[16]) * y[4] + delta_E7pRB * y[15] * y[4]/Rb_tot, #Rb-monophosphate can be formed by phosphorylation of Rb or cleavage #of Rb-polyphosphate. It can be lost by Cyclin E/CDK2 or #phosphatase activity. Rb of all sorts gets sequestered by E7. epsilon_RbCD * theta_CD * y[4]/(y[4]+k_RbCD) * y[10] - sigma_Rb * y[5]/(y[5]+k_RbMP) - epsilon_RbCE * theta_CE * y[5]/(y[5]+k_RbCE) * y[11] + sigma_Rb
random_line_split
p21_sim.py
.693 #DUMMY omega_MDM2 = 8.318 #DUMMY omega_CD = 0.023 #DUMMY omega_CDInk4 = 1 #DUMMY omega_CE = 1 #DUMMY omega_CA = 1 #DUMMY omega_CACDC20 = 1 #DUMMY omega_CB = 1 #DUMMY omega_CBCDC20 = 1 #DUMMY beta_E2FRb = 2 #DUMMY beta_E2FRbMP = beta_E2FRb * 0.3 #DUMMY, but should be smaller than beta_E2FRb beta_mdm2p14 = 1 #DUMMY beta_cp21 = 1 #DUMMY beta_E7p21 = 1 #DUMMY beta_E7pRB = 1 #DUMMY beta_E7pRBP = 1 #DUMMY beta_E7pRBPP = 1 #DUMMY delta_E7p21 = 1 #DUMMY delta_E7pRB = 1 #DUMMY epsilon_RbCD = 0.4 #DUMMY epsilon_RbCE = 0.7 #DUMMY epsilon_E2F = 20 #DUMMY epsilon_CDC20 = 1 #DUMMY sigma_Rb = 0.3 #DUMMY sigma_RbMP = 0.1 #DUMMY sigma_E2F = 0.7 #DUMMY sigma_CDC20 = 1 #DUMMY kappa_CECA = 1 #DUMMY kappa_CBCA = 1 #DUMMY #Knockdown terms. These should all be equal to 1 unless we're testing the #effects of a knockdown. theta_E2F = 1 theta_p53 = 1 theta_CE = 1 theta_MDM2 = 1 theta_CD = 1 theta_CA = 1 theta_Ink4 = 1 theta_CB = 1 theta_CDC20 = 1 k_p21 = 1 k_p53 = 1 k_RbMP = 1 k_RbPP = 1 k_RbCD = 1 k_RbCE = 1 k_E2FCA = 1 k_E2F = 1 k_CD = 1 k_CA = 1 k_CB = 1 k_CDC20CB = 1 k_CDC20 = 1 E2F_tot = 10 #DUMMY CDC20_tot = 0.285 Rb_tot = 10 #DUMMY #Potentially override parameters for infile in infiles: reader = open(infile) for line in reader.readlines(): exec(line) #Dummy initial conditions y0 = [0.1,0.1,0.1,0.1,Rb_tot,0.0,0.0,0.1,0.1,0.1,0.1,0.1,0.1,0.1,0.1,0,0] #Abort early if the output directory already exists. if dirname != "": if os.path.exists(dirname): sys.stderr.write("Output dir " + dirname + " already exists. Aborting.\n") sys.exit(2) #Functions to be called from the derivative functions. def E6(t): return 0 #dummy def E7(t): return 0 #dummy #Fractions for inhibition/inactive complex formation. #Each has a sanity check in case chief input is near zero. sanity_threshold = 0.00001 def f(e2f, rb, rbmp): if e2f < sanity_threshold:
return e2f**2 / (e2f + beta_E2FRb * rb + beta_E2FRbMP * rbmp) def g(mdm2, p14): if mdm2 < sanity_threshold: return 0 return mdm2**2 / (mdm2 + beta_mdm2p14 * p14) def h(c, p21, cd, ce, ca, cb): if c < sanity_threshold: return 0 if ca+cb+cd+ce < sanity_threshold: return 0 return c**2 / (c + beta_cp21 * p21 * c / (ca+cb+cd+ce)) def j(e7, e7prb, e7p21): return e7 - e7prb - e7p21 #Variable key #y[0] = p14 #y[1] = Ink4 #y[2] = p21 #y[3] = p53 #y[4] = Rb #y[5] = pRb-P #y[6] = pRb-PP #y[7] = E2F #y[8] = mdm2 #y[9] = pMDM2 #y[10] = CD (Cyclin D/CDK4-6 complex) #y[11] = CE (Cyclin E/CDK2 complex) #y[12] = CA (Cyclin A/CDK2 complex) #y[13] = CB (Cyclin B/CDK1 complex) #y[14] = CDC20 #y[15] = E7-pRB #y[16] = E7-p21 names = [] names.append("p14") names.append("Ink4") names.append("p21") names.append("p53") names.append("pRb") names.append("pRb-P") names.append("pRb-PP") names.append("E2F") names.append("mdm2") names.append("pMDM2") names.append("CD") names.append("CE") names.append("CA") names.append("CB") names.append("CDC20") names.append("E7-pRB") names.append("E7-p21") #The derivative function for the differential equation system. def func(y,t): return [ #We have p14 being produced by E2F after inhibition from Rb #is accounted for, and degraded at a constant rate. alpha_p14 * theta_E2F * f(y[7], y[4], y[5]) - omega_p14 * y[0], #It's just like the p14 equation, but with Ink4 instead! alpha_Ink4 * theta_E2F * f(y[7], y[4], y[5]) - omega_Ink4 * y[1], #Form p21 at a rate proportional to p53 presence; degrade it #"naturally" or with help from Cyclin E/CDK2. E7 sequesters p21. alpha_p21 * theta_p53 * y[3] - omega_p21 * y[2] - omega_p21CE * theta_CE * y[11] * y[2]/(y[2]+k_p21) - beta_E7p21 * j(E7(t),y[15],y[16]) * y[2] + delta_E7p21 * y[16], #P53 is generated naturally at a constant rate, and degrades #both on its own and with help from MDM2. alpha_p53 - omega_p53 * y[3] - (omega_p53MDM2 * theta_MDM2 * g(y[9], y[0]) + omega_p53E6 * E6(t)) * y[3]/(y[3]+k_p53), #Rb gets monophosphorylated by Cyclin D/CDK4-6. Rb-monophosphate #gets its phosphate cleaved at a constant rate. Rb of all sorts #gets sequestered by E7. -epsilon_RbCD * theta_CD * y[4]/(y[4]+k_RbCD) * y[10] + sigma_Rb * y[5]/(y[5]+k_RbMP) - beta_E7pRB * j(E7(t),y[15],y[16]) * y[4] + delta_E7pRB * y[15] * y[4]/Rb_tot, #Rb-monophosphate can be formed by phosphorylation of Rb or cleavage #of Rb-polyphosphate. It can be lost by Cyclin E/CDK2 or #phosphatase activity. Rb of all sorts gets sequestered by E7. epsilon_RbCD * theta_CD * y[4]/(y[4]+k_RbCD) * y[10] - sigma_Rb * y[5]/(y[5]+k_RbMP) - epsilon_RbCE * theta_CE * y[5]/(y[5]+k_RbCE) * y[11] + sigma_Rb
return 0
conditional_block
p21_sim.py
693 #DUMMY omega_MDM2 = 8.318 #DUMMY omega_CD = 0.023 #DUMMY omega_CDInk4 = 1 #DUMMY omega_CE = 1 #DUMMY omega_CA = 1 #DUMMY omega_CACDC20 = 1 #DUMMY omega_CB = 1 #DUMMY omega_CBCDC20 = 1 #DUMMY beta_E2FRb = 2 #DUMMY beta_E2FRbMP = beta_E2FRb * 0.3 #DUMMY, but should be smaller than beta_E2FRb beta_mdm2p14 = 1 #DUMMY beta_cp21 = 1 #DUMMY beta_E7p21 = 1 #DUMMY beta_E7pRB = 1 #DUMMY beta_E7pRBP = 1 #DUMMY beta_E7pRBPP = 1 #DUMMY delta_E7p21 = 1 #DUMMY delta_E7pRB = 1 #DUMMY epsilon_RbCD = 0.4 #DUMMY epsilon_RbCE = 0.7 #DUMMY epsilon_E2F = 20 #DUMMY epsilon_CDC20 = 1 #DUMMY sigma_Rb = 0.3 #DUMMY sigma_RbMP = 0.1 #DUMMY sigma_E2F = 0.7 #DUMMY sigma_CDC20 = 1 #DUMMY kappa_CECA = 1 #DUMMY kappa_CBCA = 1 #DUMMY #Knockdown terms. These should all be equal to 1 unless we're testing the #effects of a knockdown. theta_E2F = 1 theta_p53 = 1 theta_CE = 1 theta_MDM2 = 1 theta_CD = 1 theta_CA = 1 theta_Ink4 = 1 theta_CB = 1 theta_CDC20 = 1 k_p21 = 1 k_p53 = 1 k_RbMP = 1 k_RbPP = 1 k_RbCD = 1 k_RbCE = 1 k_E2FCA = 1 k_E2F = 1 k_CD = 1 k_CA = 1 k_CB = 1 k_CDC20CB = 1 k_CDC20 = 1 E2F_tot = 10 #DUMMY CDC20_tot = 0.285 Rb_tot = 10 #DUMMY #Potentially override parameters for infile in infiles: reader = open(infile) for line in reader.readlines(): exec(line) #Dummy initial conditions y0 = [0.1,0.1,0.1,0.1,Rb_tot,0.0,0.0,0.1,0.1,0.1,0.1,0.1,0.1,0.1,0.1,0,0] #Abort early if the output directory already exists. if dirname != "": if os.path.exists(dirname): sys.stderr.write("Output dir " + dirname + " already exists. Aborting.\n") sys.exit(2) #Functions to be called from the derivative functions. def E6(t): return 0 #dummy def E7(t): return 0 #dummy #Fractions for inhibition/inactive complex formation. #Each has a sanity check in case chief input is near zero. sanity_threshold = 0.00001 def f(e2f, rb, rbmp): if e2f < sanity_threshold: return 0 return e2f**2 / (e2f + beta_E2FRb * rb + beta_E2FRbMP * rbmp) def g(mdm2, p14): if mdm2 < sanity_threshold: return 0 return mdm2**2 / (mdm2 + beta_mdm2p14 * p14) def h(c, p21, cd, ce, ca, cb):
def j(e7, e7prb, e7p21): return e7 - e7prb - e7p21 #Variable key #y[0] = p14 #y[1] = Ink4 #y[2] = p21 #y[3] = p53 #y[4] = Rb #y[5] = pRb-P #y[6] = pRb-PP #y[7] = E2F #y[8] = mdm2 #y[9] = pMDM2 #y[10] = CD (Cyclin D/CDK4-6 complex) #y[11] = CE (Cyclin E/CDK2 complex) #y[12] = CA (Cyclin A/CDK2 complex) #y[13] = CB (Cyclin B/CDK1 complex) #y[14] = CDC20 #y[15] = E7-pRB #y[16] = E7-p21 names = [] names.append("p14") names.append("Ink4") names.append("p21") names.append("p53") names.append("pRb") names.append("pRb-P") names.append("pRb-PP") names.append("E2F") names.append("mdm2") names.append("pMDM2") names.append("CD") names.append("CE") names.append("CA") names.append("CB") names.append("CDC20") names.append("E7-pRB") names.append("E7-p21") #The derivative function for the differential equation system. def func(y,t): return [ #We have p14 being produced by E2F after inhibition from Rb #is accounted for, and degraded at a constant rate. alpha_p14 * theta_E2F * f(y[7], y[4], y[5]) - omega_p14 * y[0], #It's just like the p14 equation, but with Ink4 instead! alpha_Ink4 * theta_E2F * f(y[7], y[4], y[5]) - omega_Ink4 * y[1], #Form p21 at a rate proportional to p53 presence; degrade it #"naturally" or with help from Cyclin E/CDK2. E7 sequesters p21. alpha_p21 * theta_p53 * y[3] - omega_p21 * y[2] - omega_p21CE * theta_CE * y[11] * y[2]/(y[2]+k_p21) - beta_E7p21 * j(E7(t),y[15],y[16]) * y[2] + delta_E7p21 * y[16], #P53 is generated naturally at a constant rate, and degrades #both on its own and with help from MDM2. alpha_p53 - omega_p53 * y[3] - (omega_p53MDM2 * theta_MDM2 * g(y[9], y[0]) + omega_p53E6 * E6(t)) * y[3]/(y[3]+k_p53), #Rb gets monophosphorylated by Cyclin D/CDK4-6. Rb-monophosphate #gets its phosphate cleaved at a constant rate. Rb of all sorts #gets sequestered by E7. -epsilon_RbCD * theta_CD * y[4]/(y[4]+k_RbCD) * y[10] + sigma_Rb * y[5]/(y[5]+k_RbMP) - beta_E7pRB * j(E7(t),y[15],y[16]) * y[4] + delta_E7pRB * y[15] * y[4]/Rb_tot, #Rb-monophosphate can be formed by phosphorylation of Rb or cleavage #of Rb-polyphosphate. It can be lost by Cyclin E/CDK2 or #phosphatase activity. Rb of all sorts gets sequestered by E7. epsilon_RbCD * theta_CD * y[4]/(y[4]+k_RbCD) * y[10] - sigma_Rb * y[5]/(y[5]+k_RbMP) - epsilon_RbCE * theta_CE * y[5]/(y[5]+k_RbCE) * y[11] + sigma_Rb
if c < sanity_threshold: return 0 if ca+cb+cd+ce < sanity_threshold: return 0 return c**2 / (c + beta_cp21 * p21 * c / (ca+cb+cd+ce))
identifier_body
yaml.ts
instanceof type)) { this._index = Number.MIN_SAFE_INTEGER; } } return this as unknown as AncestryBuilder<X>; } /** * Move up the path, asserting that the current node was a key of a mapping * pair. The builder skips over the Pair to the parent YAMLMap. */ parentOfKey(): AncestryBuilder<YAMLMap> { const node = this.get(); this.parent(Pair); const pairNode = this.get(); if (pairNode instanceof Pair && pairNode.key === node) { this.parent(YAMLMap); } else { this._index = Number.MIN_SAFE_INTEGER; } return this as unknown as AncestryBuilder<YAMLMap>; } /** * Get node up to which the assertions have led. */ get(): N | null { return (this._path[this._index] as N) || null; } /** * Get the key of the Pair one level down the path. * * The key is returned only if it indeed is a string Scalar. */ // The `this` argument is for generics restriction of this method. getStringKey(this: AncestryBuilder<YAMLMap>): string | null { const node = this._path[this._index + 1]; if ( node instanceof Pair && node.key instanceof Scalar && typeof node.key.value === 'string' ) { return node.key.value; } return null; } /** * Get the value of the Pair one level down the path. */ // The `this` argument is for generics restriction of this method. getValue(this: AncestryBuilder<YAMLMap>): Node | null { const node = this._path[this._index + 1]; if (node instanceof Pair) { return node.value; } return null; } /** * Get the path to which the assertions have led. * * The path will be a subpath of the original path. */ getPath(): Node[] | null { if (this._index < 0) return null; const path = this._path.slice(0, this._index + 1); return path; } /** * Get the path to the key of the Pair one level down the path to which the * assertions have led. * * The path will be a subpath of the original path. */ // The `this` argument is for generics restriction of this method. getKeyPath(this: AncestryBuilder<YAMLMap>): Node[] | null { if (this._index < 0) return null; const path = this._path.slice(0, this._index + 1); const node = this._path[this._index + 1]; if (node instanceof Pair) { path.push(node); path.push(node.key); return path; } return null; } } export function getPathAt( document: TextDocument, position: Position, docs: Document[], inclusive = false ): Node[] | null { const offset = document.offsetAt(position); const doc = _.find(docs, (d) => contains(d.contents, offset, inclusive)); if (doc && doc.contents) { return getPathAtOffset([doc.contents], offset, inclusive); } return null; } export function contains( node: Node | null, offset: number, inclusive: boolean ): boolean { const range = getOrigRange(node); return !!( range && range[0] <= offset && (range[1] > offset || (inclusive && range[1] >= offset)) ); } export function getPathAtOffset( path: Node[], offset: number, inclusive: boolean ): Node[] | null { if (path) { const currentNode = path[path.length - 1]; if (currentNode instanceof YAMLMap) { let pair = _.find(currentNode.items, (p) => contains(p.key, offset, inclusive) ); if (pair) { return getPathAtOffset(path.concat(pair, pair.key), offset, inclusive); } pair = _.find(currentNode.items, (p) => contains(p.value, offset, inclusive) ); if (pair) { return getPathAtOffset( path.concat(pair, pair.value), offset, inclusive ); } pair = _.find(currentNode.items, (p) => { const inBetweenNode = new Node(); const start = getOrigRange(p.key as Node)?.[1]; const end = getOrigRange(p.value as Node)?.[0]; if (start && end) { inBetweenNode.range = [start, end - 1]; return contains(inBetweenNode, offset, inclusive); } else return false; }); if (pair) { return path.concat(pair, new Node()); } } else if (currentNode instanceof YAMLSeq) { const item = _.find(currentNode.items, (n) => contains(n, offset, inclusive) ); if (item) { return getPathAtOffset(path.concat(item), offset, inclusive); } } else if (contains(currentNode, offset, inclusive)) { return path; } return path.concat(new Node()); // empty node as indentation marker } return null; } export const tasksKey = /^(tasks|pre_tasks|post_tasks|block|rescue|always)$/; /** * Determines whether the path points at a parameter key of an Ansible task. */ export function isTaskParam(path: Node[]): boolean { const taskListPath = new AncestryBuilder(path) .parentOfKey() .parent(YAMLSeq) .getPath(); if (taskListPath) { // basic shape of the task list has been found if (isPlayParam(path) || isBlockParam(path) || isRoleParam(path)) return false; if (taskListPath.length === 1) { // case when the task list is at the top level of the document return true; } const taskListKey = new AncestryBuilder(taskListPath) .parent(YAMLMap) .getStringKey(); if (taskListKey && tasksKey.test(taskListKey)) { // case when a task list is defined explicitly by a keyword return true; } } return false; } /** * Tries to find the list of collections declared at the Ansible play/block/task level. */ export function getDeclaredCollections(modulePath: Node[] | null): string[] { const declaredCollections: string[] = []; const taskParamsNode = new AncestryBuilder(modulePath).parent(YAMLMap).get(); declaredCollections.push(...getDeclaredCollectionsForMap(taskParamsNode)); let path: Node[] | null = new AncestryBuilder(modulePath) .parent(YAMLMap) .getPath(); while (true) { // traverse the YAML up through the Ansible blocks const builder = new AncestryBuilder(path).parent(YAMLSeq).parent(YAMLMap); const key = builder.getStringKey(); if (key && /^block|rescue|always$/.test(key)) { declaredCollections.push(...getDeclaredCollectionsForMap(builder.get())); path = builder.getPath(); } else { break; } } // now we should be at the tasks/pre_tasks/post_tasks level const playParamsNode = new AncestryBuilder(path) .parent(YAMLSeq) .parent(YAMLMap) .get(); declaredCollections.push(...getDeclaredCollectionsForMap(playParamsNode)); return [...new Set(declaredCollections)]; // deduplicate } function getDeclaredCollectionsForMap(playNode: YAMLMap | null): string[] { const declaredCollections: string[] = []; const collectionsPair = _.find( playNode?.items, (pair) => pair.key instanceof Scalar && pair.key.value === 'collections' ); if (collectionsPair) { // we've found the collections declaration const collectionsNode = collectionsPair.value; if (collectionsNode instanceof YAMLSeq) { for (const collectionNode of collectionsNode.items) { if (collectionNode instanceof Scalar) { declaredCollections.push(collectionNode.value); } } } } return declaredCollections; } /** * Heuristically determines whether the path points at an Ansible play. The * `fileUri` helps guessing in case the YAML tree doesn't give any clues. * * Returns `undefined` if highly uncertain. */ export function isPlayParam( path: Node[], fileUri?: string ): boolean | undefined { const isAtRoot = new AncestryBuilder(path).parentOfKey().parent(YAMLSeq).getPath() ?.length === 1; if (isAtRoot) { const mapNode = new AncestryBuilder(path).parentOfKey().get() as YAMLMap; const providedKeys = getYamlMapKeys(mapNode); const containsPlayKeyword = providedKeys.some((p) => playExclusiveKeywords.has(p) ); if (containsPlayKeyword) { return true; } if (fileUri) { const isInRole = /\/roles\/[^/]+\/tasks\//.test(fileUri); if (isInRole) { return false; } } } else
{ return false; }
conditional_block
yaml.ts
this._index--; } } if (type) { if (!(this.get() instanceof type)) { this._index = Number.MIN_SAFE_INTEGER; } } return this as unknown as AncestryBuilder<X>; } /** * Move up the path, asserting that the current node was a key of a mapping * pair. The builder skips over the Pair to the parent YAMLMap. */ parentOfKey(): AncestryBuilder<YAMLMap> { const node = this.get(); this.parent(Pair); const pairNode = this.get(); if (pairNode instanceof Pair && pairNode.key === node) { this.parent(YAMLMap); } else { this._index = Number.MIN_SAFE_INTEGER; } return this as unknown as AncestryBuilder<YAMLMap>; } /** * Get node up to which the assertions have led. */ get(): N | null { return (this._path[this._index] as N) || null; } /** * Get the key of the Pair one level down the path. * * The key is returned only if it indeed is a string Scalar. */ // The `this` argument is for generics restriction of this method. getStringKey(this: AncestryBuilder<YAMLMap>): string | null { const node = this._path[this._index + 1]; if ( node instanceof Pair && node.key instanceof Scalar && typeof node.key.value === 'string' ) { return node.key.value; } return null; } /** * Get the value of the Pair one level down the path. */ // The `this` argument is for generics restriction of this method. getValue(this: AncestryBuilder<YAMLMap>): Node | null { const node = this._path[this._index + 1]; if (node instanceof Pair) { return node.value; } return null; } /** * Get the path to which the assertions have led. * * The path will be a subpath of the original path. */ getPath(): Node[] | null
/** * Get the path to the key of the Pair one level down the path to which the * assertions have led. * * The path will be a subpath of the original path. */ // The `this` argument is for generics restriction of this method. getKeyPath(this: AncestryBuilder<YAMLMap>): Node[] | null { if (this._index < 0) return null; const path = this._path.slice(0, this._index + 1); const node = this._path[this._index + 1]; if (node instanceof Pair) { path.push(node); path.push(node.key); return path; } return null; } } export function getPathAt( document: TextDocument, position: Position, docs: Document[], inclusive = false ): Node[] | null { const offset = document.offsetAt(position); const doc = _.find(docs, (d) => contains(d.contents, offset, inclusive)); if (doc && doc.contents) { return getPathAtOffset([doc.contents], offset, inclusive); } return null; } export function contains( node: Node | null, offset: number, inclusive: boolean ): boolean { const range = getOrigRange(node); return !!( range && range[0] <= offset && (range[1] > offset || (inclusive && range[1] >= offset)) ); } export function getPathAtOffset( path: Node[], offset: number, inclusive: boolean ): Node[] | null { if (path) { const currentNode = path[path.length - 1]; if (currentNode instanceof YAMLMap) { let pair = _.find(currentNode.items, (p) => contains(p.key, offset, inclusive) ); if (pair) { return getPathAtOffset(path.concat(pair, pair.key), offset, inclusive); } pair = _.find(currentNode.items, (p) => contains(p.value, offset, inclusive) ); if (pair) { return getPathAtOffset( path.concat(pair, pair.value), offset, inclusive ); } pair = _.find(currentNode.items, (p) => { const inBetweenNode = new Node(); const start = getOrigRange(p.key as Node)?.[1]; const end = getOrigRange(p.value as Node)?.[0]; if (start && end) { inBetweenNode.range = [start, end - 1]; return contains(inBetweenNode, offset, inclusive); } else return false; }); if (pair) { return path.concat(pair, new Node()); } } else if (currentNode instanceof YAMLSeq) { const item = _.find(currentNode.items, (n) => contains(n, offset, inclusive) ); if (item) { return getPathAtOffset(path.concat(item), offset, inclusive); } } else if (contains(currentNode, offset, inclusive)) { return path; } return path.concat(new Node()); // empty node as indentation marker } return null; } export const tasksKey = /^(tasks|pre_tasks|post_tasks|block|rescue|always)$/; /** * Determines whether the path points at a parameter key of an Ansible task. */ export function isTaskParam(path: Node[]): boolean { const taskListPath = new AncestryBuilder(path) .parentOfKey() .parent(YAMLSeq) .getPath(); if (taskListPath) { // basic shape of the task list has been found if (isPlayParam(path) || isBlockParam(path) || isRoleParam(path)) return false; if (taskListPath.length === 1) { // case when the task list is at the top level of the document return true; } const taskListKey = new AncestryBuilder(taskListPath) .parent(YAMLMap) .getStringKey(); if (taskListKey && tasksKey.test(taskListKey)) { // case when a task list is defined explicitly by a keyword return true; } } return false; } /** * Tries to find the list of collections declared at the Ansible play/block/task level. */ export function getDeclaredCollections(modulePath: Node[] | null): string[] { const declaredCollections: string[] = []; const taskParamsNode = new AncestryBuilder(modulePath).parent(YAMLMap).get(); declaredCollections.push(...getDeclaredCollectionsForMap(taskParamsNode)); let path: Node[] | null = new AncestryBuilder(modulePath) .parent(YAMLMap) .getPath(); while (true) { // traverse the YAML up through the Ansible blocks const builder = new AncestryBuilder(path).parent(YAMLSeq).parent(YAMLMap); const key = builder.getStringKey(); if (key && /^block|rescue|always$/.test(key)) { declaredCollections.push(...getDeclaredCollectionsForMap(builder.get())); path = builder.getPath(); } else { break; } } // now we should be at the tasks/pre_tasks/post_tasks level const playParamsNode = new AncestryBuilder(path) .parent(YAMLSeq) .parent(YAMLMap) .get(); declaredCollections.push(...getDeclaredCollectionsForMap(playParamsNode)); return [...new Set(declaredCollections)]; // deduplicate } function getDeclaredCollectionsForMap(playNode: YAMLMap | null): string[] { const declaredCollections: string[] = []; const collectionsPair = _.find( playNode?.items, (pair) => pair.key instanceof Scalar && pair.key.value === 'collections' ); if (collectionsPair) { // we've found the collections declaration const collectionsNode = collectionsPair.value; if (collectionsNode instanceof YAMLSeq) { for (const collectionNode of collectionsNode.items) { if (collectionNode instanceof Scalar) { declaredCollections.push(collectionNode.value); } } } } return declaredCollections; } /** * Heuristically determines whether the path points at an Ansible play. The * `fileUri` helps guessing in case the YAML tree doesn't give any clues. * * Returns `undefined` if highly uncertain. */ export function isPlayParam( path: Node[], fileUri?: string ): boolean | undefined { const isAtRoot = new AncestryBuilder(path).parentOfKey().parent(YAMLSeq).getPath() ?.length === 1; if (isAtRoot) { const mapNode = new AncestryBuilder(path).parentOfKey().get() as YAMLMap; const providedKeys = getYamlMapKeys(mapNode); const containsPlayKeyword = providedKeys.some((p) => playExclusiveKeywords.has(p) ); if (containsPlayKeyword) { return true; } if (fileUri) { const isInRole = /\/roles\/[^/]+\/tasks\//.test(fileUri); if (is
{ if (this._index < 0) return null; const path = this._path.slice(0, this._index + 1); return path; }
identifier_body
yaml.ts
this._index--; } } if (type) { if (!(this.get() instanceof type)) { this._index = Number.MIN_SAFE_INTEGER; } } return this as unknown as AncestryBuilder<X>; } /** * Move up the path, asserting that the current node was a key of a mapping * pair. The builder skips over the Pair to the parent YAMLMap. */ parentOfKey(): AncestryBuilder<YAMLMap> { const node = this.get(); this.parent(Pair); const pairNode = this.get(); if (pairNode instanceof Pair && pairNode.key === node) { this.parent(YAMLMap); } else { this._index = Number.MIN_SAFE_INTEGER; } return this as unknown as AncestryBuilder<YAMLMap>; } /** * Get node up to which the assertions have led. */ get(): N | null { return (this._path[this._index] as N) || null; } /** * Get the key of the Pair one level down the path. * * The key is returned only if it indeed is a string Scalar. */ // The `this` argument is for generics restriction of this method. getStringKey(this: AncestryBuilder<YAMLMap>): string | null { const node = this._path[this._index + 1]; if ( node instanceof Pair && node.key instanceof Scalar && typeof node.key.value === 'string' ) { return node.key.value; } return null; } /** * Get the value of the Pair one level down the path. */ // The `this` argument is for generics restriction of this method. getValue(this: AncestryBuilder<YAMLMap>): Node | null { const node = this._path[this._index + 1]; if (node instanceof Pair) { return node.value; }
/** * Get the path to which the assertions have led. * * The path will be a subpath of the original path. */ getPath(): Node[] | null { if (this._index < 0) return null; const path = this._path.slice(0, this._index + 1); return path; } /** * Get the path to the key of the Pair one level down the path to which the * assertions have led. * * The path will be a subpath of the original path. */ // The `this` argument is for generics restriction of this method. getKeyPath(this: AncestryBuilder<YAMLMap>): Node[] | null { if (this._index < 0) return null; const path = this._path.slice(0, this._index + 1); const node = this._path[this._index + 1]; if (node instanceof Pair) { path.push(node); path.push(node.key); return path; } return null; } } export function getPathAt( document: TextDocument, position: Position, docs: Document[], inclusive = false ): Node[] | null { const offset = document.offsetAt(position); const doc = _.find(docs, (d) => contains(d.contents, offset, inclusive)); if (doc && doc.contents) { return getPathAtOffset([doc.contents], offset, inclusive); } return null; } export function contains( node: Node | null, offset: number, inclusive: boolean ): boolean { const range = getOrigRange(node); return !!( range && range[0] <= offset && (range[1] > offset || (inclusive && range[1] >= offset)) ); } export function getPathAtOffset( path: Node[], offset: number, inclusive: boolean ): Node[] | null { if (path) { const currentNode = path[path.length - 1]; if (currentNode instanceof YAMLMap) { let pair = _.find(currentNode.items, (p) => contains(p.key, offset, inclusive) ); if (pair) { return getPathAtOffset(path.concat(pair, pair.key), offset, inclusive); } pair = _.find(currentNode.items, (p) => contains(p.value, offset, inclusive) ); if (pair) { return getPathAtOffset( path.concat(pair, pair.value), offset, inclusive ); } pair = _.find(currentNode.items, (p) => { const inBetweenNode = new Node(); const start = getOrigRange(p.key as Node)?.[1]; const end = getOrigRange(p.value as Node)?.[0]; if (start && end) { inBetweenNode.range = [start, end - 1]; return contains(inBetweenNode, offset, inclusive); } else return false; }); if (pair) { return path.concat(pair, new Node()); } } else if (currentNode instanceof YAMLSeq) { const item = _.find(currentNode.items, (n) => contains(n, offset, inclusive) ); if (item) { return getPathAtOffset(path.concat(item), offset, inclusive); } } else if (contains(currentNode, offset, inclusive)) { return path; } return path.concat(new Node()); // empty node as indentation marker } return null; } export const tasksKey = /^(tasks|pre_tasks|post_tasks|block|rescue|always)$/; /** * Determines whether the path points at a parameter key of an Ansible task. */ export function isTaskParam(path: Node[]): boolean { const taskListPath = new AncestryBuilder(path) .parentOfKey() .parent(YAMLSeq) .getPath(); if (taskListPath) { // basic shape of the task list has been found if (isPlayParam(path) || isBlockParam(path) || isRoleParam(path)) return false; if (taskListPath.length === 1) { // case when the task list is at the top level of the document return true; } const taskListKey = new AncestryBuilder(taskListPath) .parent(YAMLMap) .getStringKey(); if (taskListKey && tasksKey.test(taskListKey)) { // case when a task list is defined explicitly by a keyword return true; } } return false; } /** * Tries to find the list of collections declared at the Ansible play/block/task level. */ export function getDeclaredCollections(modulePath: Node[] | null): string[] { const declaredCollections: string[] = []; const taskParamsNode = new AncestryBuilder(modulePath).parent(YAMLMap).get(); declaredCollections.push(...getDeclaredCollectionsForMap(taskParamsNode)); let path: Node[] | null = new AncestryBuilder(modulePath) .parent(YAMLMap) .getPath(); while (true) { // traverse the YAML up through the Ansible blocks const builder = new AncestryBuilder(path).parent(YAMLSeq).parent(YAMLMap); const key = builder.getStringKey(); if (key && /^block|rescue|always$/.test(key)) { declaredCollections.push(...getDeclaredCollectionsForMap(builder.get())); path = builder.getPath(); } else { break; } } // now we should be at the tasks/pre_tasks/post_tasks level const playParamsNode = new AncestryBuilder(path) .parent(YAMLSeq) .parent(YAMLMap) .get(); declaredCollections.push(...getDeclaredCollectionsForMap(playParamsNode)); return [...new Set(declaredCollections)]; // deduplicate } function getDeclaredCollectionsForMap(playNode: YAMLMap | null): string[] { const declaredCollections: string[] = []; const collectionsPair = _.find( playNode?.items, (pair) => pair.key instanceof Scalar && pair.key.value === 'collections' ); if (collectionsPair) { // we've found the collections declaration const collectionsNode = collectionsPair.value; if (collectionsNode instanceof YAMLSeq) { for (const collectionNode of collectionsNode.items) { if (collectionNode instanceof Scalar) { declaredCollections.push(collectionNode.value); } } } } return declaredCollections; } /** * Heuristically determines whether the path points at an Ansible play. The * `fileUri` helps guessing in case the YAML tree doesn't give any clues. * * Returns `undefined` if highly uncertain. */ export function isPlayParam( path: Node[], fileUri?: string ): boolean | undefined { const isAtRoot = new AncestryBuilder(path).parentOfKey().parent(YAMLSeq).getPath() ?.length === 1; if (isAtRoot) { const mapNode = new AncestryBuilder(path).parentOfKey().get() as YAMLMap; const providedKeys = getYamlMapKeys(mapNode); const containsPlayKeyword = providedKeys.some((p) => playExclusiveKeywords.has(p) ); if (containsPlayKeyword) { return true; } if (fileUri) { const isInRole = /\/roles\/[^/]+\/tasks\//.test(fileUri); if (isIn
return null; }
random_line_split
yaml.ts
this._index--; } } if (type) { if (!(this.get() instanceof type)) { this._index = Number.MIN_SAFE_INTEGER; } } return this as unknown as AncestryBuilder<X>; } /** * Move up the path, asserting that the current node was a key of a mapping * pair. The builder skips over the Pair to the parent YAMLMap. */ parentOfKey(): AncestryBuilder<YAMLMap> { const node = this.get(); this.parent(Pair); const pairNode = this.get(); if (pairNode instanceof Pair && pairNode.key === node) { this.parent(YAMLMap); } else { this._index = Number.MIN_SAFE_INTEGER; } return this as unknown as AncestryBuilder<YAMLMap>; } /** * Get node up to which the assertions have led. */
(): N | null { return (this._path[this._index] as N) || null; } /** * Get the key of the Pair one level down the path. * * The key is returned only if it indeed is a string Scalar. */ // The `this` argument is for generics restriction of this method. getStringKey(this: AncestryBuilder<YAMLMap>): string | null { const node = this._path[this._index + 1]; if ( node instanceof Pair && node.key instanceof Scalar && typeof node.key.value === 'string' ) { return node.key.value; } return null; } /** * Get the value of the Pair one level down the path. */ // The `this` argument is for generics restriction of this method. getValue(this: AncestryBuilder<YAMLMap>): Node | null { const node = this._path[this._index + 1]; if (node instanceof Pair) { return node.value; } return null; } /** * Get the path to which the assertions have led. * * The path will be a subpath of the original path. */ getPath(): Node[] | null { if (this._index < 0) return null; const path = this._path.slice(0, this._index + 1); return path; } /** * Get the path to the key of the Pair one level down the path to which the * assertions have led. * * The path will be a subpath of the original path. */ // The `this` argument is for generics restriction of this method. getKeyPath(this: AncestryBuilder<YAMLMap>): Node[] | null { if (this._index < 0) return null; const path = this._path.slice(0, this._index + 1); const node = this._path[this._index + 1]; if (node instanceof Pair) { path.push(node); path.push(node.key); return path; } return null; } } export function getPathAt( document: TextDocument, position: Position, docs: Document[], inclusive = false ): Node[] | null { const offset = document.offsetAt(position); const doc = _.find(docs, (d) => contains(d.contents, offset, inclusive)); if (doc && doc.contents) { return getPathAtOffset([doc.contents], offset, inclusive); } return null; } export function contains( node: Node | null, offset: number, inclusive: boolean ): boolean { const range = getOrigRange(node); return !!( range && range[0] <= offset && (range[1] > offset || (inclusive && range[1] >= offset)) ); } export function getPathAtOffset( path: Node[], offset: number, inclusive: boolean ): Node[] | null { if (path) { const currentNode = path[path.length - 1]; if (currentNode instanceof YAMLMap) { let pair = _.find(currentNode.items, (p) => contains(p.key, offset, inclusive) ); if (pair) { return getPathAtOffset(path.concat(pair, pair.key), offset, inclusive); } pair = _.find(currentNode.items, (p) => contains(p.value, offset, inclusive) ); if (pair) { return getPathAtOffset( path.concat(pair, pair.value), offset, inclusive ); } pair = _.find(currentNode.items, (p) => { const inBetweenNode = new Node(); const start = getOrigRange(p.key as Node)?.[1]; const end = getOrigRange(p.value as Node)?.[0]; if (start && end) { inBetweenNode.range = [start, end - 1]; return contains(inBetweenNode, offset, inclusive); } else return false; }); if (pair) { return path.concat(pair, new Node()); } } else if (currentNode instanceof YAMLSeq) { const item = _.find(currentNode.items, (n) => contains(n, offset, inclusive) ); if (item) { return getPathAtOffset(path.concat(item), offset, inclusive); } } else if (contains(currentNode, offset, inclusive)) { return path; } return path.concat(new Node()); // empty node as indentation marker } return null; } export const tasksKey = /^(tasks|pre_tasks|post_tasks|block|rescue|always)$/; /** * Determines whether the path points at a parameter key of an Ansible task. */ export function isTaskParam(path: Node[]): boolean { const taskListPath = new AncestryBuilder(path) .parentOfKey() .parent(YAMLSeq) .getPath(); if (taskListPath) { // basic shape of the task list has been found if (isPlayParam(path) || isBlockParam(path) || isRoleParam(path)) return false; if (taskListPath.length === 1) { // case when the task list is at the top level of the document return true; } const taskListKey = new AncestryBuilder(taskListPath) .parent(YAMLMap) .getStringKey(); if (taskListKey && tasksKey.test(taskListKey)) { // case when a task list is defined explicitly by a keyword return true; } } return false; } /** * Tries to find the list of collections declared at the Ansible play/block/task level. */ export function getDeclaredCollections(modulePath: Node[] | null): string[] { const declaredCollections: string[] = []; const taskParamsNode = new AncestryBuilder(modulePath).parent(YAMLMap).get(); declaredCollections.push(...getDeclaredCollectionsForMap(taskParamsNode)); let path: Node[] | null = new AncestryBuilder(modulePath) .parent(YAMLMap) .getPath(); while (true) { // traverse the YAML up through the Ansible blocks const builder = new AncestryBuilder(path).parent(YAMLSeq).parent(YAMLMap); const key = builder.getStringKey(); if (key && /^block|rescue|always$/.test(key)) { declaredCollections.push(...getDeclaredCollectionsForMap(builder.get())); path = builder.getPath(); } else { break; } } // now we should be at the tasks/pre_tasks/post_tasks level const playParamsNode = new AncestryBuilder(path) .parent(YAMLSeq) .parent(YAMLMap) .get(); declaredCollections.push(...getDeclaredCollectionsForMap(playParamsNode)); return [...new Set(declaredCollections)]; // deduplicate } function getDeclaredCollectionsForMap(playNode: YAMLMap | null): string[] { const declaredCollections: string[] = []; const collectionsPair = _.find( playNode?.items, (pair) => pair.key instanceof Scalar && pair.key.value === 'collections' ); if (collectionsPair) { // we've found the collections declaration const collectionsNode = collectionsPair.value; if (collectionsNode instanceof YAMLSeq) { for (const collectionNode of collectionsNode.items) { if (collectionNode instanceof Scalar) { declaredCollections.push(collectionNode.value); } } } } return declaredCollections; } /** * Heuristically determines whether the path points at an Ansible play. The * `fileUri` helps guessing in case the YAML tree doesn't give any clues. * * Returns `undefined` if highly uncertain. */ export function isPlayParam( path: Node[], fileUri?: string ): boolean | undefined { const isAtRoot = new AncestryBuilder(path).parentOfKey().parent(YAMLSeq).getPath() ?.length === 1; if (isAtRoot) { const mapNode = new AncestryBuilder(path).parentOfKey().get() as YAMLMap; const providedKeys = getYamlMapKeys(mapNode); const containsPlayKeyword = providedKeys.some((p) => playExclusiveKeywords.has(p) ); if (containsPlayKeyword) { return true; } if (fileUri) { const isInRole = /\/roles\/[^/]+\/tasks\//.test(fileUri); if (isInRole
get
identifier_name
main.rs
first.starts_with("== ")); assert!(first.ends_with(" ==")); let name = first .trim_end_matches(" ==") .trim_start_matches("== ") .to_owned(); let message = lines .next() .ok_or_else(|| anyhow!("No second line"))? .trim() .to_owned(); assert!(!message.is_empty(), "Expected non-empty message"); let next = lines.next().ok_or_else(|| anyhow!("No third line"))?.trim(); assert!(next.is_empty(), "Expected third line to be empty"); let next = lines .next() .ok_or_else(|| anyhow!("No fourth line"))? .trim(); assert!( next == "Doors here lead:", "Expected third line to be 'Doors here lead:'" ); let mut directions = Vec::new(); let mut next = lines .next() .ok_or_else(|| anyhow!("No line after doors"))? .trim(); while next.starts_with("- ") { let dir = next.trim_start_matches("- "); let dir = str_to_compass(dir)?; directions.push(dir); next = lines .next() .ok_or_else(|| anyhow!("No line after directions"))? .trim(); } assert!( next.is_empty(), "Expected line after directions to be empty" ); next = lines .next() .ok_or_else(|| anyhow!("No line after directions + empty"))? .trim(); let mut items = BTreeSet::new(); if next == "Items here:" { next = lines .next() .ok_or_else(|| anyhow!("No line after items"))? .trim(); while next.starts_with("- ") { let item = next.trim_start_matches("- "); items.insert(item.to_owned()); next = lines .next()
.ok_or_else(|| anyhow!("No line after items"))? .trim(); } assert!( next.is_empty(), "Expected line after items to be empty, got '{}'", next ); next = lines .next() .ok_or_else(|| anyhow!("No line after items + empty"))? .trim(); } if next.contains("Alert!") { return Err(Ejection(next.into()).into()); } assert_eq!( next, "Command?", "Expected line after items to be 'Command?'" ); assert!(lines.next().is_none()); Ok(Room { name, message, items, directions, }) } } #[derive(Debug, Clone)] pub struct Explorer { comp: IntComp, room: Key, direction: Compass, carrying: BTreeSet<String>, map: Map, } impl Explorer { fn new(mut comp: IntComp) -> AnyResult<Self> { let mut output = OutputVec::new(); comp.run_to_input(&mut output)?; let out = output.as_string()?; let room = Room::from_str(&out)?; let mut map: Map = Default::default(); let key = map.add_room(room); let exp = Explorer { comp, room: key, direction: Compass::North, carrying: Default::default(), map, }; Ok(exp) } fn see_room(&self) -> &Room { self.map.get(self.room) } fn process_input_str(&mut self, output: &mut OutputVec, input: &str) -> anyhow::Result<String> { log::debug!("Process 1: '{}'", input); self.comp .process_ascii(input, output)? .expect(Stopped::Input)?; log::debug!("Process 2: '\\n'"); self.comp .process_ascii("\n", output)? .expect(Stopped::Input)?; log::debug!("Processed: '\\n'"); Ok(output.as_string()?) } fn process_str(&mut self, input: &str) -> anyhow::Result<String> { let mut out = OutputVec::new(); match self.process_input_str(&mut out, input) { Ok(v) => Ok(v), Err(e) => { let output = out.as_string()?; log::warn!("process_str failure on input {}, output: {}", input, output); Err(e) } } } // fn north(&mut self) -> anyhow::Result<String> { // self.process_str("north") // } // fn south(&mut self) -> anyhow::Result<String> { // self.process_str("south") // } // fn east(&mut self) -> anyhow::Result<String> { // self.process_str("east") // } // fn west(&mut self) -> anyhow::Result<String> { // self.process_str("west") // } pub fn step(&mut self, direction: Compass) -> anyhow::Result<()> { let input = match direction { Compass::East => "east", Compass::North => "north", Compass::South => "south", Compass::West => "west", }; log::debug!("Taking step {}", input); let output = self.process_str(input)?; log::debug!("Took step:\n{}\n", output); let room = Room::from_str(&output)?; let new = self.map.add_room(room); self.map.add_door(self.room, direction, new); self.room = new; self.direction = direction; Ok(()) } pub fn take(&mut self, item: &str) -> anyhow::Result<String> { log::debug!("Taking {}", item); let mut s = String::from("take "); s.push_str(item); let result = self.process_str(&s)?; let new = self.carrying.insert(item.to_string()); let room = self.map.rooms.get_mut(self.room).unwrap(); room.items.remove(item); assert!(new, "Expected to add {}", item); log::debug!(" took {}", item); Ok(result) } pub fn drop(&mut self, item: &str) -> anyhow::Result<String> { let found = self.carrying.remove(item); assert!(!found, "Expected to drop {}", item); let mut s = String::from("drop "); s.push_str(item); self.process_str(&s) } pub fn inventory(&mut self) -> anyhow::Result<String> { self.process_str("inv") } fn left_wall_step(&mut self) -> AnyResult<()> { let mut dir = self.direction + Turn::Left; for _ in 0..4 { log::debug!("Checking {} -> {}", self.direction, dir); if self.see_room().directions.contains(&dir) { break; } dir = dir + Turn::Right; } assert!(self.see_room().directions.contains(&dir)); self.step(dir)?; log::debug!("Stepped {}, {}", dir, self.see_room().name); Ok(()) } fn explore_and_take(&mut self, items: &BTreeSet<String>) -> AnyResult<()> { let start = self.room; let mut start_directions = self.see_room().directions.clone(); start_directions.reverse(); loop { let overlap: BTreeSet<String> = items .intersection(&self.see_room().items) .map(|s| s.to_owned()) .collect(); for item in overlap { let _output = self.take(&item)?; // println!("Took {}, output: {}", item, output.trim()); } if self.see_room().name == "Security Checkpoint" { // println!("inv: {}", self.inventory()?); log::info!("Turning around at security checkpoint"); self.step(self.direction + Turn::Reverse)?; continue; } if self.room == start { let dir = match start_directions.pop() { None => return Ok(()), Some(d) => d, }; self.step(dir)?; continue; } self.left_wall_step()?; } } pub fn goto(&mut self, room: &str) -> AnyResult<()> { loop { if self.see_room().name == room { return Ok(()); } self.left_wall_step()?; } } } type Key = slotmap::DefaultKey; #[derive(Default, Debug, Clone)] pub struct Map { rooms_by_name: HashMap<String, Key>, rooms: SlotMap<Key, Room>, doors: HashMap<Key, BTreeMap<Compass, Key>>, unvisited: HashMap<Key, BTreeSet<Compass>>, } impl Map { fn add_room(&mut self, room: Room) -> Key { if let Some(&key) = self.rooms_by_name.get(&room.name) { return key; } let name = room.name.clone(); let directions = room.directions.clone(); let key = self.rooms.insert(room); self.rooms_by_name.insert(name, key); let unvisited = self.unvisited.insert(key, Default::default()); assert!(unvisited.is_none()); let unvisited = self.unvisited.get_mut(&key).unwrap(); for dir in directions { unvisited.insert(dir); } key } fn visit(&mut self, room: Key, direction: Compass) { if let Occupied(mut o) = self.unvisited.entry(room) { o.get
random_line_split
main.rs
_str(&mut self, output: &mut OutputVec, input: &str) -> anyhow::Result<String> { log::debug!("Process 1: '{}'", input); self.comp .process_ascii(input, output)? .expect(Stopped::Input)?; log::debug!("Process 2: '\\n'"); self.comp .process_ascii("\n", output)? .expect(Stopped::Input)?; log::debug!("Processed: '\\n'"); Ok(output.as_string()?) } fn process_str(&mut self, input: &str) -> anyhow::Result<String> { let mut out = OutputVec::new(); match self.process_input_str(&mut out, input) { Ok(v) => Ok(v), Err(e) => { let output = out.as_string()?; log::warn!("process_str failure on input {}, output: {}", input, output); Err(e) } } } // fn north(&mut self) -> anyhow::Result<String> { // self.process_str("north") // } // fn south(&mut self) -> anyhow::Result<String> { // self.process_str("south") // } // fn east(&mut self) -> anyhow::Result<String> { // self.process_str("east") // } // fn west(&mut self) -> anyhow::Result<String> { // self.process_str("west") // } pub fn step(&mut self, direction: Compass) -> anyhow::Result<()> { let input = match direction { Compass::East => "east", Compass::North => "north", Compass::South => "south", Compass::West => "west", }; log::debug!("Taking step {}", input); let output = self.process_str(input)?; log::debug!("Took step:\n{}\n", output); let room = Room::from_str(&output)?; let new = self.map.add_room(room); self.map.add_door(self.room, direction, new); self.room = new; self.direction = direction; Ok(()) } pub fn take(&mut self, item: &str) -> anyhow::Result<String> { log::debug!("Taking {}", item); let mut s = String::from("take "); s.push_str(item); let result = self.process_str(&s)?; let new = self.carrying.insert(item.to_string()); let room = self.map.rooms.get_mut(self.room).unwrap(); room.items.remove(item); assert!(new, "Expected to add {}", item); log::debug!(" took {}", item); Ok(result) } pub fn drop(&mut self, item: &str) -> anyhow::Result<String> { let found = self.carrying.remove(item); assert!(!found, "Expected to drop {}", item); let mut s = String::from("drop "); s.push_str(item); self.process_str(&s) } pub fn inventory(&mut self) -> anyhow::Result<String> { self.process_str("inv") } fn left_wall_step(&mut self) -> AnyResult<()> { let mut dir = self.direction + Turn::Left; for _ in 0..4 { log::debug!("Checking {} -> {}", self.direction, dir); if self.see_room().directions.contains(&dir) { break; } dir = dir + Turn::Right; } assert!(self.see_room().directions.contains(&dir)); self.step(dir)?; log::debug!("Stepped {}, {}", dir, self.see_room().name); Ok(()) } fn explore_and_take(&mut self, items: &BTreeSet<String>) -> AnyResult<()> { let start = self.room; let mut start_directions = self.see_room().directions.clone(); start_directions.reverse(); loop { let overlap: BTreeSet<String> = items .intersection(&self.see_room().items) .map(|s| s.to_owned()) .collect(); for item in overlap { let _output = self.take(&item)?; // println!("Took {}, output: {}", item, output.trim()); } if self.see_room().name == "Security Checkpoint" { // println!("inv: {}", self.inventory()?); log::info!("Turning around at security checkpoint"); self.step(self.direction + Turn::Reverse)?; continue; } if self.room == start { let dir = match start_directions.pop() { None => return Ok(()), Some(d) => d, }; self.step(dir)?; continue; } self.left_wall_step()?; } } pub fn goto(&mut self, room: &str) -> AnyResult<()> { loop { if self.see_room().name == room { return Ok(()); } self.left_wall_step()?; } } } type Key = slotmap::DefaultKey; #[derive(Default, Debug, Clone)] pub struct Map { rooms_by_name: HashMap<String, Key>, rooms: SlotMap<Key, Room>, doors: HashMap<Key, BTreeMap<Compass, Key>>, unvisited: HashMap<Key, BTreeSet<Compass>>, } impl Map { fn add_room(&mut self, room: Room) -> Key { if let Some(&key) = self.rooms_by_name.get(&room.name) { return key; } let name = room.name.clone(); let directions = room.directions.clone(); let key = self.rooms.insert(room); self.rooms_by_name.insert(name, key); let unvisited = self.unvisited.insert(key, Default::default()); assert!(unvisited.is_none()); let unvisited = self.unvisited.get_mut(&key).unwrap(); for dir in directions { unvisited.insert(dir); } key } fn visit(&mut self, room: Key, direction: Compass) { if let Occupied(mut o) = self.unvisited.entry(room) { o.get_mut().remove(&direction); if o.get().is_empty() { o.remove(); } } } fn add_door(&mut self, first: Key, direction: Compass, second: Key) { self.doors .entry(first) .or_default() .insert(direction, second); self.doors .entry(second) .or_default() .insert(direction + Turn::Reverse, first); self.visit(first, direction); self.visit(second, direction + Turn::Reverse); } pub fn len(&self) -> usize { self.rooms.len() } pub fn is_empty(&self) -> bool { self.rooms.is_empty() } pub fn contains(&self, room: &Room) -> bool { self.rooms_by_name.contains_key(&room.name) } fn get(&self, key: Key) -> &Room { self.rooms.get(key).unwrap() } #[allow(dead_code)] fn to_coords(&self, origin: Option<Key>) -> HashMap<Position, Key> { let start = match (origin, self.rooms.iter().next()) { (Some(k), _) => k, (None, None) => return Default::default(), (None, Some((k, _r))) => k, }; let mut queue = vec![(Position(0, 0), start)]; let mut seen = HashSet::new(); let mut coords = HashMap::new(); while let Some((pos, r)) = queue.pop() { match coords.entry(pos) { Occupied(o) => { assert!(seen.contains(&r)); assert!(*o.get() == r); } Vacant(v) => { assert!(!seen.contains(&r)); seen.insert(r); v.insert(r); let neighbors = self.doors.get(&r).unwrap(); for (&d, &r) in neighbors { queue.push((pos + d, r)); } } } } coords } } /* NV || SB KT=GW=PS || || CQ=HD || || OB=ST || || SG HB=EG=WD=AR=SL || || HW=HC=CO || SC AR: Arcade CO: Corridor CQ: Crew Quarters EG: Engineering GW: Gift Wrapping Center HB: Hull Breach HC: Hot Chocolate Fountain HD: Holodeck HW: Hallway KT: Kitchen NV: Navigation OB: Observatory SB: Sick Bay SC: Security Checkpoint SG: Storage SL: Science Lab ST: Stables WD: Warp Drive Maintenance */ fn try_item_combos(initial_explorer: Explorer, items: Vec<String>) -> AnyResult<Explorer>
{ let total = 1 << items.len(); for n in 0..total { let mut explorer = initial_explorer.clone(); let cur_items: BTreeSet<String> = items .iter() .enumerate() .filter_map(|(i, item)| { if (n & (1 << i)) == 0 { None } else { Some(item.clone()) } }) .collect(); log::info!("Items: {:?}", cur_items); explorer.explore_and_take(&cur_items)?; assert_eq!(explorer.carrying, cur_items);
identifier_body
main.rs
line to be empty"); let next = lines .next() .ok_or_else(|| anyhow!("No fourth line"))? .trim(); assert!( next == "Doors here lead:", "Expected third line to be 'Doors here lead:'" ); let mut directions = Vec::new(); let mut next = lines .next() .ok_or_else(|| anyhow!("No line after doors"))? .trim(); while next.starts_with("- ") { let dir = next.trim_start_matches("- "); let dir = str_to_compass(dir)?; directions.push(dir); next = lines .next() .ok_or_else(|| anyhow!("No line after directions"))? .trim(); } assert!( next.is_empty(), "Expected line after directions to be empty" ); next = lines .next() .ok_or_else(|| anyhow!("No line after directions + empty"))? .trim(); let mut items = BTreeSet::new(); if next == "Items here:" { next = lines .next() .ok_or_else(|| anyhow!("No line after items"))? .trim(); while next.starts_with("- ") { let item = next.trim_start_matches("- "); items.insert(item.to_owned()); next = lines .next() .ok_or_else(|| anyhow!("No line after items"))? .trim(); } assert!( next.is_empty(), "Expected line after items to be empty, got '{}'", next ); next = lines .next() .ok_or_else(|| anyhow!("No line after items + empty"))? .trim(); } if next.contains("Alert!") { return Err(Ejection(next.into()).into()); } assert_eq!( next, "Command?", "Expected line after items to be 'Command?'" ); assert!(lines.next().is_none()); Ok(Room { name, message, items, directions, }) } } #[derive(Debug, Clone)] pub struct Explorer { comp: IntComp, room: Key, direction: Compass, carrying: BTreeSet<String>, map: Map, } impl Explorer { fn new(mut comp: IntComp) -> AnyResult<Self> { let mut output = OutputVec::new(); comp.run_to_input(&mut output)?; let out = output.as_string()?; let room = Room::from_str(&out)?; let mut map: Map = Default::default(); let key = map.add_room(room); let exp = Explorer { comp, room: key, direction: Compass::North, carrying: Default::default(), map, }; Ok(exp) } fn see_room(&self) -> &Room { self.map.get(self.room) } fn process_input_str(&mut self, output: &mut OutputVec, input: &str) -> anyhow::Result<String> { log::debug!("Process 1: '{}'", input); self.comp .process_ascii(input, output)? .expect(Stopped::Input)?; log::debug!("Process 2: '\\n'"); self.comp .process_ascii("\n", output)? .expect(Stopped::Input)?; log::debug!("Processed: '\\n'"); Ok(output.as_string()?) } fn process_str(&mut self, input: &str) -> anyhow::Result<String> { let mut out = OutputVec::new(); match self.process_input_str(&mut out, input) { Ok(v) => Ok(v), Err(e) => { let output = out.as_string()?; log::warn!("process_str failure on input {}, output: {}", input, output); Err(e) } } } // fn north(&mut self) -> anyhow::Result<String> { // self.process_str("north") // } // fn south(&mut self) -> anyhow::Result<String> { // self.process_str("south") // } // fn east(&mut self) -> anyhow::Result<String> { // self.process_str("east") // } // fn west(&mut self) -> anyhow::Result<String> { // self.process_str("west") // } pub fn step(&mut self, direction: Compass) -> anyhow::Result<()> { let input = match direction { Compass::East => "east", Compass::North => "north", Compass::South => "south", Compass::West => "west", }; log::debug!("Taking step {}", input); let output = self.process_str(input)?; log::debug!("Took step:\n{}\n", output); let room = Room::from_str(&output)?; let new = self.map.add_room(room); self.map.add_door(self.room, direction, new); self.room = new; self.direction = direction; Ok(()) } pub fn take(&mut self, item: &str) -> anyhow::Result<String> { log::debug!("Taking {}", item); let mut s = String::from("take "); s.push_str(item); let result = self.process_str(&s)?; let new = self.carrying.insert(item.to_string()); let room = self.map.rooms.get_mut(self.room).unwrap(); room.items.remove(item); assert!(new, "Expected to add {}", item); log::debug!(" took {}", item); Ok(result) } pub fn drop(&mut self, item: &str) -> anyhow::Result<String> { let found = self.carrying.remove(item); assert!(!found, "Expected to drop {}", item); let mut s = String::from("drop "); s.push_str(item); self.process_str(&s) } pub fn inventory(&mut self) -> anyhow::Result<String> { self.process_str("inv") } fn left_wall_step(&mut self) -> AnyResult<()> { let mut dir = self.direction + Turn::Left; for _ in 0..4 { log::debug!("Checking {} -> {}", self.direction, dir); if self.see_room().directions.contains(&dir) { break; } dir = dir + Turn::Right; } assert!(self.see_room().directions.contains(&dir)); self.step(dir)?; log::debug!("Stepped {}, {}", dir, self.see_room().name); Ok(()) } fn explore_and_take(&mut self, items: &BTreeSet<String>) -> AnyResult<()> { let start = self.room; let mut start_directions = self.see_room().directions.clone(); start_directions.reverse(); loop { let overlap: BTreeSet<String> = items .intersection(&self.see_room().items) .map(|s| s.to_owned()) .collect(); for item in overlap { let _output = self.take(&item)?; // println!("Took {}, output: {}", item, output.trim()); } if self.see_room().name == "Security Checkpoint" { // println!("inv: {}", self.inventory()?); log::info!("Turning around at security checkpoint"); self.step(self.direction + Turn::Reverse)?; continue; } if self.room == start { let dir = match start_directions.pop() { None => return Ok(()), Some(d) => d, }; self.step(dir)?; continue; } self.left_wall_step()?; } } pub fn goto(&mut self, room: &str) -> AnyResult<()> { loop { if self.see_room().name == room { return Ok(()); } self.left_wall_step()?; } } } type Key = slotmap::DefaultKey; #[derive(Default, Debug, Clone)] pub struct Map { rooms_by_name: HashMap<String, Key>, rooms: SlotMap<Key, Room>, doors: HashMap<Key, BTreeMap<Compass, Key>>, unvisited: HashMap<Key, BTreeSet<Compass>>, } impl Map { fn add_room(&mut self, room: Room) -> Key { if let Some(&key) = self.rooms_by_name.get(&room.name) { return key; } let name = room.name.clone(); let directions = room.directions.clone(); let key = self.rooms.insert(room); self.rooms_by_name.insert(name, key); let unvisited = self.unvisited.insert(key, Default::default()); assert!(unvisited.is_none()); let unvisited = self.unvisited.get_mut(&key).unwrap(); for dir in directions { unvisited.insert(dir); } key } fn visit(&mut self, room: Key, direction: Compass) { if let Occupied(mut o) = self.unvisited.entry(room) { o.get_mut().remove(&direction); if o.get().is_empty() { o.remove(); } } } fn add_door(&mut self, first: Key, direction: Compass, second: Key) { self.doors .entry(first) .or_default() .insert(direction, second); self.doors .entry(second) .or_default() .insert(direction + Turn::Reverse, first); self.visit(first, direction); self.visit(second, direction + Turn::Reverse); } pub fn
len
identifier_name
main.rs
.starts_with("== ")); assert!(first.ends_with(" ==")); let name = first .trim_end_matches(" ==") .trim_start_matches("== ") .to_owned(); let message = lines .next() .ok_or_else(|| anyhow!("No second line"))? .trim() .to_owned(); assert!(!message.is_empty(), "Expected non-empty message"); let next = lines.next().ok_or_else(|| anyhow!("No third line"))?.trim(); assert!(next.is_empty(), "Expected third line to be empty"); let next = lines .next() .ok_or_else(|| anyhow!("No fourth line"))? .trim(); assert!( next == "Doors here lead:", "Expected third line to be 'Doors here lead:'" ); let mut directions = Vec::new(); let mut next = lines .next() .ok_or_else(|| anyhow!("No line after doors"))? .trim(); while next.starts_with("- ") { let dir = next.trim_start_matches("- "); let dir = str_to_compass(dir)?; directions.push(dir); next = lines .next() .ok_or_else(|| anyhow!("No line after directions"))? .trim(); } assert!( next.is_empty(), "Expected line after directions to be empty" ); next = lines .next() .ok_or_else(|| anyhow!("No line after directions + empty"))? .trim(); let mut items = BTreeSet::new(); if next == "Items here:" { next = lines .next() .ok_or_else(|| anyhow!("No line after items"))? .trim(); while next.starts_with("- ") { let item = next.trim_start_matches("- "); items.insert(item.to_owned()); next = lines .next() .ok_or_else(|| anyhow!("No line after items"))? .trim(); } assert!( next.is_empty(), "Expected line after items to be empty, got '{}'", next ); next = lines .next() .ok_or_else(|| anyhow!("No line after items + empty"))? .trim(); } if next.contains("Alert!") { return Err(Ejection(next.into()).into()); } assert_eq!( next, "Command?", "Expected line after items to be 'Command?'" ); assert!(lines.next().is_none()); Ok(Room { name, message, items, directions, }) } } #[derive(Debug, Clone)] pub struct Explorer { comp: IntComp, room: Key, direction: Compass, carrying: BTreeSet<String>, map: Map, } impl Explorer { fn new(mut comp: IntComp) -> AnyResult<Self> { let mut output = OutputVec::new(); comp.run_to_input(&mut output)?; let out = output.as_string()?; let room = Room::from_str(&out)?; let mut map: Map = Default::default(); let key = map.add_room(room); let exp = Explorer { comp, room: key, direction: Compass::North, carrying: Default::default(), map, }; Ok(exp) } fn see_room(&self) -> &Room { self.map.get(self.room) } fn process_input_str(&mut self, output: &mut OutputVec, input: &str) -> anyhow::Result<String> { log::debug!("Process 1: '{}'", input); self.comp .process_ascii(input, output)? .expect(Stopped::Input)?; log::debug!("Process 2: '\\n'"); self.comp .process_ascii("\n", output)? .expect(Stopped::Input)?; log::debug!("Processed: '\\n'"); Ok(output.as_string()?) } fn process_str(&mut self, input: &str) -> anyhow::Result<String> { let mut out = OutputVec::new(); match self.process_input_str(&mut out, input) { Ok(v) => Ok(v), Err(e) =>
} } // fn north(&mut self) -> anyhow::Result<String> { // self.process_str("north") // } // fn south(&mut self) -> anyhow::Result<String> { // self.process_str("south") // } // fn east(&mut self) -> anyhow::Result<String> { // self.process_str("east") // } // fn west(&mut self) -> anyhow::Result<String> { // self.process_str("west") // } pub fn step(&mut self, direction: Compass) -> anyhow::Result<()> { let input = match direction { Compass::East => "east", Compass::North => "north", Compass::South => "south", Compass::West => "west", }; log::debug!("Taking step {}", input); let output = self.process_str(input)?; log::debug!("Took step:\n{}\n", output); let room = Room::from_str(&output)?; let new = self.map.add_room(room); self.map.add_door(self.room, direction, new); self.room = new; self.direction = direction; Ok(()) } pub fn take(&mut self, item: &str) -> anyhow::Result<String> { log::debug!("Taking {}", item); let mut s = String::from("take "); s.push_str(item); let result = self.process_str(&s)?; let new = self.carrying.insert(item.to_string()); let room = self.map.rooms.get_mut(self.room).unwrap(); room.items.remove(item); assert!(new, "Expected to add {}", item); log::debug!(" took {}", item); Ok(result) } pub fn drop(&mut self, item: &str) -> anyhow::Result<String> { let found = self.carrying.remove(item); assert!(!found, "Expected to drop {}", item); let mut s = String::from("drop "); s.push_str(item); self.process_str(&s) } pub fn inventory(&mut self) -> anyhow::Result<String> { self.process_str("inv") } fn left_wall_step(&mut self) -> AnyResult<()> { let mut dir = self.direction + Turn::Left; for _ in 0..4 { log::debug!("Checking {} -> {}", self.direction, dir); if self.see_room().directions.contains(&dir) { break; } dir = dir + Turn::Right; } assert!(self.see_room().directions.contains(&dir)); self.step(dir)?; log::debug!("Stepped {}, {}", dir, self.see_room().name); Ok(()) } fn explore_and_take(&mut self, items: &BTreeSet<String>) -> AnyResult<()> { let start = self.room; let mut start_directions = self.see_room().directions.clone(); start_directions.reverse(); loop { let overlap: BTreeSet<String> = items .intersection(&self.see_room().items) .map(|s| s.to_owned()) .collect(); for item in overlap { let _output = self.take(&item)?; // println!("Took {}, output: {}", item, output.trim()); } if self.see_room().name == "Security Checkpoint" { // println!("inv: {}", self.inventory()?); log::info!("Turning around at security checkpoint"); self.step(self.direction + Turn::Reverse)?; continue; } if self.room == start { let dir = match start_directions.pop() { None => return Ok(()), Some(d) => d, }; self.step(dir)?; continue; } self.left_wall_step()?; } } pub fn goto(&mut self, room: &str) -> AnyResult<()> { loop { if self.see_room().name == room { return Ok(()); } self.left_wall_step()?; } } } type Key = slotmap::DefaultKey; #[derive(Default, Debug, Clone)] pub struct Map { rooms_by_name: HashMap<String, Key>, rooms: SlotMap<Key, Room>, doors: HashMap<Key, BTreeMap<Compass, Key>>, unvisited: HashMap<Key, BTreeSet<Compass>>, } impl Map { fn add_room(&mut self, room: Room) -> Key { if let Some(&key) = self.rooms_by_name.get(&room.name) { return key; } let name = room.name.clone(); let directions = room.directions.clone(); let key = self.rooms.insert(room); self.rooms_by_name.insert(name, key); let unvisited = self.unvisited.insert(key, Default::default()); assert!(unvisited.is_none()); let unvisited = self.unvisited.get_mut(&key).unwrap(); for dir in directions { unvisited.insert(dir); } key } fn visit(&mut self, room: Key, direction: Compass) { if let Occupied(mut o) = self.unvisited.entry(room) { o
{ let output = out.as_string()?; log::warn!("process_str failure on input {}, output: {}", input, output); Err(e) }
conditional_block
client.go
err := syscall.SetsockoptInt(s, syscall.SOL_SOCKET, syscall.SO_RCVBUF, bufferSize); err != nil { return 0, nil, err } } } return s, lsa, nil } // Make syscall asking for all connections. Invoke 'cb' for each connection. func queryAllConnections(bufferSize int, cb func(Conn), inetFamily uint8) error { s, lsa, err := connectNetfilter(bufferSize, 0) if err != nil { return err } defer syscall.Close(s) msg := ConntrackListReq{ Header: syscall.NlMsghdr{ Len: syscall.NLMSG_HDRLEN + sizeofGenmsg, Type: (NFNL_SUBSYS_CTNETLINK << 8) | uint16(IpctnlMsgCtGet), Flags: syscall.NLM_F_REQUEST | syscall.NLM_F_DUMP, Pid: 0, Seq: 0, }, Body: unix.Nfgenmsg{ Nfgen_family: inetFamily, Version: NFNETLINK_V0, Res_id: 0, }, } wb := msg.toWireFormat() // fmt.Printf("msg bytes: %q\n", wb) if err := syscall.Sendto(s, wb, 0, lsa); err != nil { return err } return readMsgs(s, cb) } // Stream all connections instead of query for all of them at once. func StreamAllConnections(inetFamily uint8) chan Conn { ch := make(chan Conn, 1) go func() { queryAllConnections(0, func(c Conn) { ch <- c }, inetFamily) close(ch) }() return ch } // Lists all the connections that conntrack is tracking. func Connections(inetFamily uint8) ([]Conn, error) { return ConnectionsSize(0, inetFamily) } // Lists all the connections that conntrack is tracking, using specified netlink buffer size. func ConnectionsSize(bufferSize int, inetFamily uint8) ([]Conn, error) { var conns []Conn queryAllConnections(bufferSize, func(c Conn) { conns = append(conns, c) }, inetFamily) return conns, nil } // Established lists all established TCP connections. func Established(inetFamily uint8) ([]ConnTCP, error) { var conns []ConnTCP local := localIPs() err := queryAllConnections(0, func(c Conn) { if c.MsgType != NfctMsgUpdate { fmt.Printf("msg isn't an update: %d\n", c.MsgType) return } if c.TCPState != "ESTABLISHED" { // fmt.Printf("state isn't ESTABLISHED: %s\n", c.TCPState) return } if tc := c.ConnTCP(local); tc != nil { conns = append(conns, *tc) } }, inetFamily) if err != nil { return nil, err } return conns, nil } // Follow gives a channel with all changes. func Follow(flags uint32) (<-chan Conn, func(), error) { return FollowSize(0, flags) } // Follow gives a channel with all changes, , using specified netlink buffer size. func FollowSize(bufferSize int, flags uint32) (<-chan Conn, func(), error) { var closing bool s, _, err := connectNetfilter(bufferSize, flags) stop := func() { closing = true syscall.Close(s) } if err != nil { return nil, stop, err } res := make(chan Conn, 1) go func() { defer syscall.Close(s) if err := readMsgs(s, func(c Conn) { // if conn.TCPState != 3 { // // 3 is TCP established. // continue // } res <- c }); err != nil && !closing { panic(err) } }() return res, stop, nil } func readMsgs(s int, cb func(Conn)) error { rb := make([]byte, 2*syscall.Getpagesize()) loop: for { nr, _, err := syscall.Recvfrom(s, rb, 0) if err == syscall.ENOBUFS { // ENOBUF means we miss some events here. No way around it. That's life. cb(Conn{Err: syscall.ENOBUFS}) continue } else if err != nil { return err } msgs, err := syscall.ParseNetlinkMessage(rb[:nr]) if err != nil { return err } for _, msg := range msgs { if msg.Header.Type == unix.NLMSG_ERROR { return errors.New("NLMSG_ERROR") } if msg.Header.Type == unix.NLMSG_DONE { break loop } if nflnSubsysID(msg.Header.Type) != NFNL_SUBSYS_CTNETLINK { return fmt.Errorf( "unexpected subsys_id: %d\n", nflnSubsysID(msg.Header.Type), ) } conn, err := parsePayload(msg.Data[sizeofGenmsg:]) if err != nil { return err } // Taken from conntrack/parse.c:__parse_message_type switch CntlMsgTypes(nflnMsgType(msg.Header.Type)) { case IpctnlMsgCtNew: conn.MsgType = NfctMsgUpdate if msg.Header.Flags&(syscall.NLM_F_CREATE|syscall.NLM_F_EXCL) > 0 { conn.MsgType = NfctMsgNew } case IpctnlMsgCtDelete: conn.MsgType = NfctMsgDestroy } cb(*conn) } } return nil } type Tuple struct { Proto uint8 Src net.IP SrcPort uint16 Dst net.IP DstPort uint16 // Flow stats. Bytes uint64 Packets uint64 // ICMP stuff. IcmpId uint16 IcmpType uint8 IcmpCode uint8 } func (t Tuple) String() string { return fmt.Sprintf("src=%v dst=%v sport=%d dport=%d packets=%d size=%d",t.Src,t.Dst,t.SrcPort,t.DstPort,t.Packets,t.Bytes) } type Conn struct { MsgType NfConntrackMsg TCPState string Status CtStatus Orig Tuple Reply Tuple // ct.mark, used to set permission type of the flow. CtMark uint32 // ct.id, used to identify connections. CtId uint32 // For multitenancy. Zone uint16 // Error, if any. Err error } func (c Conn) String() string{ return fmt.Sprintf("%d %s %v %v mark=%d\n",c.Orig.Proto,c.TCPState,c.Orig,c.Reply,c.CtMark) } // ConnTCP decides which way this connection is going and makes a ConnTCP. func (c Conn) ConnTCP(local map[string]struct{}) *ConnTCP { // conntrack gives us all connections, even things passing through, but it // doesn't tell us what the local IP is. So we use `local` as a guide // what's local. src := c.Orig.Src.String() dst := c.Orig.Dst.String() _, srcLocal := local[src] _, dstLocal := local[dst] // If both are local we must just order things predictably. if srcLocal && dstLocal { srcLocal = c.Orig.SrcPort < c.Orig.DstPort } if srcLocal { return &ConnTCP{ Local: src, LocalPort: strconv.Itoa(int(c.Orig.SrcPort)), Remote: dst, RemotePort: strconv.Itoa(int(c.Orig.DstPort)), } } if dstLocal { return &ConnTCP{ Local: dst, LocalPort: strconv.Itoa(int(c.Orig.DstPort)), Remote: src, RemotePort: strconv.Itoa(int(c.Orig.SrcPort)), } } // Neither is local. conntrack also reports NAT connections. return nil } func parsePayload(b []byte) (*Conn, error) { // Most of this comes from libnetfilter_conntrack/src/conntrack/parse_mnl.c conn := &Conn{} var attrSpace [16]Attr attrs, err := parseAttrs(b, attrSpace[0:0]) if err != nil
for _, attr := range attrs { switch CtattrType(attr.Typ) { case CtaTupleOrig: parseTuple(attr.Msg, &conn.Orig) case CtaTupleReply: parseTuple(attr.Msg, &conn.Reply) case CtaCountersOrig: conn.Orig.Packets, conn.Orig.Bytes, _ = parseCounters(attr.Msg) case CtaCountersReply: conn.Reply.Packets, conn.Reply.Bytes, _ = parseCounters
{ return conn, err }
conditional_block
client.go
err := syscall.SetsockoptInt(s, syscall.SOL_SOCKET, syscall.SO_RCVBUF, bufferSize); err != nil { return 0, nil, err } } } return s, lsa, nil } // Make syscall asking for all connections. Invoke 'cb' for each connection. func queryAllConnections(bufferSize int, cb func(Conn), inetFamily uint8) error { s, lsa, err := connectNetfilter(bufferSize, 0) if err != nil { return err } defer syscall.Close(s) msg := ConntrackListReq{ Header: syscall.NlMsghdr{ Len: syscall.NLMSG_HDRLEN + sizeofGenmsg, Type: (NFNL_SUBSYS_CTNETLINK << 8) | uint16(IpctnlMsgCtGet), Flags: syscall.NLM_F_REQUEST | syscall.NLM_F_DUMP, Pid: 0, Seq: 0, }, Body: unix.Nfgenmsg{ Nfgen_family: inetFamily, Version: NFNETLINK_V0, Res_id: 0, }, } wb := msg.toWireFormat() // fmt.Printf("msg bytes: %q\n", wb) if err := syscall.Sendto(s, wb, 0, lsa); err != nil { return err } return readMsgs(s, cb) } // Stream all connections instead of query for all of them at once. func StreamAllConnections(inetFamily uint8) chan Conn { ch := make(chan Conn, 1) go func() { queryAllConnections(0, func(c Conn) { ch <- c }, inetFamily) close(ch) }() return ch } // Lists all the connections that conntrack is tracking. func Connections(inetFamily uint8) ([]Conn, error) { return ConnectionsSize(0, inetFamily) } // Lists all the connections that conntrack is tracking, using specified netlink buffer size. func ConnectionsSize(bufferSize int, inetFamily uint8) ([]Conn, error) { var conns []Conn queryAllConnections(bufferSize, func(c Conn) { conns = append(conns, c) }, inetFamily) return conns, nil } // Established lists all established TCP connections. func Established(inetFamily uint8) ([]ConnTCP, error) { var conns []ConnTCP local := localIPs() err := queryAllConnections(0, func(c Conn) { if c.MsgType != NfctMsgUpdate { fmt.Printf("msg isn't an update: %d\n", c.MsgType) return } if c.TCPState != "ESTABLISHED" { // fmt.Printf("state isn't ESTABLISHED: %s\n", c.TCPState) return } if tc := c.ConnTCP(local); tc != nil { conns = append(conns, *tc) } }, inetFamily) if err != nil { return nil, err } return conns, nil } // Follow gives a channel with all changes. func Follow(flags uint32) (<-chan Conn, func(), error) { return FollowSize(0, flags) } // Follow gives a channel with all changes, , using specified netlink buffer size. func FollowSize(bufferSize int, flags uint32) (<-chan Conn, func(), error) { var closing bool s, _, err := connectNetfilter(bufferSize, flags) stop := func() { closing = true syscall.Close(s) } if err != nil { return nil, stop, err } res := make(chan Conn, 1) go func() { defer syscall.Close(s) if err := readMsgs(s, func(c Conn) { // if conn.TCPState != 3 { // // 3 is TCP established. // continue // } res <- c }); err != nil && !closing { panic(err) } }() return res, stop, nil } func readMsgs(s int, cb func(Conn)) error { rb := make([]byte, 2*syscall.Getpagesize()) loop: for { nr, _, err := syscall.Recvfrom(s, rb, 0) if err == syscall.ENOBUFS { // ENOBUF means we miss some events here. No way around it. That's life. cb(Conn{Err: syscall.ENOBUFS}) continue } else if err != nil { return err } msgs, err := syscall.ParseNetlinkMessage(rb[:nr]) if err != nil { return err } for _, msg := range msgs { if msg.Header.Type == unix.NLMSG_ERROR { return errors.New("NLMSG_ERROR") } if msg.Header.Type == unix.NLMSG_DONE { break loop } if nflnSubsysID(msg.Header.Type) != NFNL_SUBSYS_CTNETLINK { return fmt.Errorf( "unexpected subsys_id: %d\n", nflnSubsysID(msg.Header.Type), ) } conn, err := parsePayload(msg.Data[sizeofGenmsg:]) if err != nil { return err } // Taken from conntrack/parse.c:__parse_message_type switch CntlMsgTypes(nflnMsgType(msg.Header.Type)) { case IpctnlMsgCtNew: conn.MsgType = NfctMsgUpdate if msg.Header.Flags&(syscall.NLM_F_CREATE|syscall.NLM_F_EXCL) > 0 { conn.MsgType = NfctMsgNew } case IpctnlMsgCtDelete: conn.MsgType = NfctMsgDestroy } cb(*conn) } } return nil } type Tuple struct { Proto uint8 Src net.IP SrcPort uint16 Dst net.IP DstPort uint16 // Flow stats. Bytes uint64 Packets uint64 // ICMP stuff. IcmpId uint16 IcmpType uint8 IcmpCode uint8 } func (t Tuple) String() string { return fmt.Sprintf("src=%v dst=%v sport=%d dport=%d packets=%d size=%d",t.Src,t.Dst,t.SrcPort,t.DstPort,t.Packets,t.Bytes) } type Conn struct { MsgType NfConntrackMsg TCPState string Status CtStatus Orig Tuple Reply Tuple // ct.mark, used to set permission type of the flow. CtMark uint32 // ct.id, used to identify connections. CtId uint32 // For multitenancy. Zone uint16 // Error, if any. Err error } func (c Conn)
() string{ return fmt.Sprintf("%d %s %v %v mark=%d\n",c.Orig.Proto,c.TCPState,c.Orig,c.Reply,c.CtMark) } // ConnTCP decides which way this connection is going and makes a ConnTCP. func (c Conn) ConnTCP(local map[string]struct{}) *ConnTCP { // conntrack gives us all connections, even things passing through, but it // doesn't tell us what the local IP is. So we use `local` as a guide // what's local. src := c.Orig.Src.String() dst := c.Orig.Dst.String() _, srcLocal := local[src] _, dstLocal := local[dst] // If both are local we must just order things predictably. if srcLocal && dstLocal { srcLocal = c.Orig.SrcPort < c.Orig.DstPort } if srcLocal { return &ConnTCP{ Local: src, LocalPort: strconv.Itoa(int(c.Orig.SrcPort)), Remote: dst, RemotePort: strconv.Itoa(int(c.Orig.DstPort)), } } if dstLocal { return &ConnTCP{ Local: dst, LocalPort: strconv.Itoa(int(c.Orig.DstPort)), Remote: src, RemotePort: strconv.Itoa(int(c.Orig.SrcPort)), } } // Neither is local. conntrack also reports NAT connections. return nil } func parsePayload(b []byte) (*Conn, error) { // Most of this comes from libnetfilter_conntrack/src/conntrack/parse_mnl.c conn := &Conn{} var attrSpace [16]Attr attrs, err := parseAttrs(b, attrSpace[0:0]) if err != nil { return conn, err } for _, attr := range attrs { switch CtattrType(attr.Typ) { case CtaTupleOrig: parseTuple(attr.Msg, &conn.Orig) case CtaTupleReply: parseTuple(attr.Msg, &conn.Reply) case CtaCountersOrig: conn.Orig.Packets, conn.Orig.Bytes, _ = parseCounters(attr.Msg) case CtaCountersReply: conn.Reply.Packets, conn.Reply.Bytes, _ = parseCounters(attr
String
identifier_name
client.go
err := syscall.SetsockoptInt(s, syscall.SOL_SOCKET, syscall.SO_RCVBUF, bufferSize); err != nil { return 0, nil, err } } } return s, lsa, nil } // Make syscall asking for all connections. Invoke 'cb' for each connection. func queryAllConnections(bufferSize int, cb func(Conn), inetFamily uint8) error { s, lsa, err := connectNetfilter(bufferSize, 0) if err != nil { return err } defer syscall.Close(s) msg := ConntrackListReq{ Header: syscall.NlMsghdr{ Len: syscall.NLMSG_HDRLEN + sizeofGenmsg, Type: (NFNL_SUBSYS_CTNETLINK << 8) | uint16(IpctnlMsgCtGet), Flags: syscall.NLM_F_REQUEST | syscall.NLM_F_DUMP, Pid: 0, Seq: 0, }, Body: unix.Nfgenmsg{ Nfgen_family: inetFamily, Version: NFNETLINK_V0, Res_id: 0, }, } wb := msg.toWireFormat() // fmt.Printf("msg bytes: %q\n", wb) if err := syscall.Sendto(s, wb, 0, lsa); err != nil { return err } return readMsgs(s, cb) } // Stream all connections instead of query for all of them at once. func StreamAllConnections(inetFamily uint8) chan Conn { ch := make(chan Conn, 1) go func() { queryAllConnections(0, func(c Conn) { ch <- c }, inetFamily) close(ch) }() return ch } // Lists all the connections that conntrack is tracking. func Connections(inetFamily uint8) ([]Conn, error) { return ConnectionsSize(0, inetFamily) } // Lists all the connections that conntrack is tracking, using specified netlink buffer size. func ConnectionsSize(bufferSize int, inetFamily uint8) ([]Conn, error) { var conns []Conn queryAllConnections(bufferSize, func(c Conn) { conns = append(conns, c) }, inetFamily) return conns, nil } // Established lists all established TCP connections. func Established(inetFamily uint8) ([]ConnTCP, error) { var conns []ConnTCP local := localIPs() err := queryAllConnections(0, func(c Conn) { if c.MsgType != NfctMsgUpdate { fmt.Printf("msg isn't an update: %d\n", c.MsgType) return } if c.TCPState != "ESTABLISHED" { // fmt.Printf("state isn't ESTABLISHED: %s\n", c.TCPState) return } if tc := c.ConnTCP(local); tc != nil { conns = append(conns, *tc) } }, inetFamily) if err != nil { return nil, err } return conns, nil } // Follow gives a channel with all changes. func Follow(flags uint32) (<-chan Conn, func(), error) { return FollowSize(0, flags) } // Follow gives a channel with all changes, , using specified netlink buffer size. func FollowSize(bufferSize int, flags uint32) (<-chan Conn, func(), error) { var closing bool s, _, err := connectNetfilter(bufferSize, flags) stop := func() { closing = true syscall.Close(s) } if err != nil { return nil, stop, err } res := make(chan Conn, 1) go func() { defer syscall.Close(s) if err := readMsgs(s, func(c Conn) { // if conn.TCPState != 3 { // // 3 is TCP established. // continue // } res <- c }); err != nil && !closing { panic(err) } }() return res, stop, nil } func readMsgs(s int, cb func(Conn)) error { rb := make([]byte, 2*syscall.Getpagesize()) loop: for { nr, _, err := syscall.Recvfrom(s, rb, 0) if err == syscall.ENOBUFS { // ENOBUF means we miss some events here. No way around it. That's life. cb(Conn{Err: syscall.ENOBUFS}) continue } else if err != nil { return err } msgs, err := syscall.ParseNetlinkMessage(rb[:nr]) if err != nil { return err } for _, msg := range msgs { if msg.Header.Type == unix.NLMSG_ERROR { return errors.New("NLMSG_ERROR") } if msg.Header.Type == unix.NLMSG_DONE { break loop } if nflnSubsysID(msg.Header.Type) != NFNL_SUBSYS_CTNETLINK { return fmt.Errorf( "unexpected subsys_id: %d\n", nflnSubsysID(msg.Header.Type), ) } conn, err := parsePayload(msg.Data[sizeofGenmsg:]) if err != nil { return err }
if msg.Header.Flags&(syscall.NLM_F_CREATE|syscall.NLM_F_EXCL) > 0 { conn.MsgType = NfctMsgNew } case IpctnlMsgCtDelete: conn.MsgType = NfctMsgDestroy } cb(*conn) } } return nil } type Tuple struct { Proto uint8 Src net.IP SrcPort uint16 Dst net.IP DstPort uint16 // Flow stats. Bytes uint64 Packets uint64 // ICMP stuff. IcmpId uint16 IcmpType uint8 IcmpCode uint8 } func (t Tuple) String() string { return fmt.Sprintf("src=%v dst=%v sport=%d dport=%d packets=%d size=%d",t.Src,t.Dst,t.SrcPort,t.DstPort,t.Packets,t.Bytes) } type Conn struct { MsgType NfConntrackMsg TCPState string Status CtStatus Orig Tuple Reply Tuple // ct.mark, used to set permission type of the flow. CtMark uint32 // ct.id, used to identify connections. CtId uint32 // For multitenancy. Zone uint16 // Error, if any. Err error } func (c Conn) String() string{ return fmt.Sprintf("%d %s %v %v mark=%d\n",c.Orig.Proto,c.TCPState,c.Orig,c.Reply,c.CtMark) } // ConnTCP decides which way this connection is going and makes a ConnTCP. func (c Conn) ConnTCP(local map[string]struct{}) *ConnTCP { // conntrack gives us all connections, even things passing through, but it // doesn't tell us what the local IP is. So we use `local` as a guide // what's local. src := c.Orig.Src.String() dst := c.Orig.Dst.String() _, srcLocal := local[src] _, dstLocal := local[dst] // If both are local we must just order things predictably. if srcLocal && dstLocal { srcLocal = c.Orig.SrcPort < c.Orig.DstPort } if srcLocal { return &ConnTCP{ Local: src, LocalPort: strconv.Itoa(int(c.Orig.SrcPort)), Remote: dst, RemotePort: strconv.Itoa(int(c.Orig.DstPort)), } } if dstLocal { return &ConnTCP{ Local: dst, LocalPort: strconv.Itoa(int(c.Orig.DstPort)), Remote: src, RemotePort: strconv.Itoa(int(c.Orig.SrcPort)), } } // Neither is local. conntrack also reports NAT connections. return nil } func parsePayload(b []byte) (*Conn, error) { // Most of this comes from libnetfilter_conntrack/src/conntrack/parse_mnl.c conn := &Conn{} var attrSpace [16]Attr attrs, err := parseAttrs(b, attrSpace[0:0]) if err != nil { return conn, err } for _, attr := range attrs { switch CtattrType(attr.Typ) { case CtaTupleOrig: parseTuple(attr.Msg, &conn.Orig) case CtaTupleReply: parseTuple(attr.Msg, &conn.Reply) case CtaCountersOrig: conn.Orig.Packets, conn.Orig.Bytes, _ = parseCounters(attr.Msg) case CtaCountersReply: conn.Reply.Packets, conn.Reply.Bytes, _ = parseCounters(attr.Msg
// Taken from conntrack/parse.c:__parse_message_type switch CntlMsgTypes(nflnMsgType(msg.Header.Type)) { case IpctnlMsgCtNew: conn.MsgType = NfctMsgUpdate
random_line_split
client.go
err := syscall.SetsockoptInt(s, syscall.SOL_SOCKET, syscall.SO_RCVBUF, bufferSize); err != nil { return 0, nil, err } } } return s, lsa, nil } // Make syscall asking for all connections. Invoke 'cb' for each connection. func queryAllConnections(bufferSize int, cb func(Conn), inetFamily uint8) error { s, lsa, err := connectNetfilter(bufferSize, 0) if err != nil { return err } defer syscall.Close(s) msg := ConntrackListReq{ Header: syscall.NlMsghdr{ Len: syscall.NLMSG_HDRLEN + sizeofGenmsg, Type: (NFNL_SUBSYS_CTNETLINK << 8) | uint16(IpctnlMsgCtGet), Flags: syscall.NLM_F_REQUEST | syscall.NLM_F_DUMP, Pid: 0, Seq: 0, }, Body: unix.Nfgenmsg{ Nfgen_family: inetFamily, Version: NFNETLINK_V0, Res_id: 0, }, } wb := msg.toWireFormat() // fmt.Printf("msg bytes: %q\n", wb) if err := syscall.Sendto(s, wb, 0, lsa); err != nil { return err } return readMsgs(s, cb) } // Stream all connections instead of query for all of them at once. func StreamAllConnections(inetFamily uint8) chan Conn { ch := make(chan Conn, 1) go func() { queryAllConnections(0, func(c Conn) { ch <- c }, inetFamily) close(ch) }() return ch } // Lists all the connections that conntrack is tracking. func Connections(inetFamily uint8) ([]Conn, error) { return ConnectionsSize(0, inetFamily) } // Lists all the connections that conntrack is tracking, using specified netlink buffer size. func ConnectionsSize(bufferSize int, inetFamily uint8) ([]Conn, error) { var conns []Conn queryAllConnections(bufferSize, func(c Conn) { conns = append(conns, c) }, inetFamily) return conns, nil } // Established lists all established TCP connections. func Established(inetFamily uint8) ([]ConnTCP, error) { var conns []ConnTCP local := localIPs() err := queryAllConnections(0, func(c Conn) { if c.MsgType != NfctMsgUpdate { fmt.Printf("msg isn't an update: %d\n", c.MsgType) return } if c.TCPState != "ESTABLISHED" { // fmt.Printf("state isn't ESTABLISHED: %s\n", c.TCPState) return } if tc := c.ConnTCP(local); tc != nil { conns = append(conns, *tc) } }, inetFamily) if err != nil { return nil, err } return conns, nil } // Follow gives a channel with all changes. func Follow(flags uint32) (<-chan Conn, func(), error) { return FollowSize(0, flags) } // Follow gives a channel with all changes, , using specified netlink buffer size. func FollowSize(bufferSize int, flags uint32) (<-chan Conn, func(), error) { var closing bool s, _, err := connectNetfilter(bufferSize, flags) stop := func() { closing = true syscall.Close(s) } if err != nil { return nil, stop, err } res := make(chan Conn, 1) go func() { defer syscall.Close(s) if err := readMsgs(s, func(c Conn) { // if conn.TCPState != 3 { // // 3 is TCP established. // continue // } res <- c }); err != nil && !closing { panic(err) } }() return res, stop, nil } func readMsgs(s int, cb func(Conn)) error { rb := make([]byte, 2*syscall.Getpagesize()) loop: for { nr, _, err := syscall.Recvfrom(s, rb, 0) if err == syscall.ENOBUFS { // ENOBUF means we miss some events here. No way around it. That's life. cb(Conn{Err: syscall.ENOBUFS}) continue } else if err != nil { return err } msgs, err := syscall.ParseNetlinkMessage(rb[:nr]) if err != nil { return err } for _, msg := range msgs { if msg.Header.Type == unix.NLMSG_ERROR { return errors.New("NLMSG_ERROR") } if msg.Header.Type == unix.NLMSG_DONE { break loop } if nflnSubsysID(msg.Header.Type) != NFNL_SUBSYS_CTNETLINK { return fmt.Errorf( "unexpected subsys_id: %d\n", nflnSubsysID(msg.Header.Type), ) } conn, err := parsePayload(msg.Data[sizeofGenmsg:]) if err != nil { return err } // Taken from conntrack/parse.c:__parse_message_type switch CntlMsgTypes(nflnMsgType(msg.Header.Type)) { case IpctnlMsgCtNew: conn.MsgType = NfctMsgUpdate if msg.Header.Flags&(syscall.NLM_F_CREATE|syscall.NLM_F_EXCL) > 0 { conn.MsgType = NfctMsgNew } case IpctnlMsgCtDelete: conn.MsgType = NfctMsgDestroy } cb(*conn) } } return nil } type Tuple struct { Proto uint8 Src net.IP SrcPort uint16 Dst net.IP DstPort uint16 // Flow stats. Bytes uint64 Packets uint64 // ICMP stuff. IcmpId uint16 IcmpType uint8 IcmpCode uint8 } func (t Tuple) String() string { return fmt.Sprintf("src=%v dst=%v sport=%d dport=%d packets=%d size=%d",t.Src,t.Dst,t.SrcPort,t.DstPort,t.Packets,t.Bytes) } type Conn struct { MsgType NfConntrackMsg TCPState string Status CtStatus Orig Tuple Reply Tuple // ct.mark, used to set permission type of the flow. CtMark uint32 // ct.id, used to identify connections. CtId uint32 // For multitenancy. Zone uint16 // Error, if any. Err error } func (c Conn) String() string
// ConnTCP decides which way this connection is going and makes a ConnTCP. func (c Conn) ConnTCP(local map[string]struct{}) *ConnTCP { // conntrack gives us all connections, even things passing through, but it // doesn't tell us what the local IP is. So we use `local` as a guide // what's local. src := c.Orig.Src.String() dst := c.Orig.Dst.String() _, srcLocal := local[src] _, dstLocal := local[dst] // If both are local we must just order things predictably. if srcLocal && dstLocal { srcLocal = c.Orig.SrcPort < c.Orig.DstPort } if srcLocal { return &ConnTCP{ Local: src, LocalPort: strconv.Itoa(int(c.Orig.SrcPort)), Remote: dst, RemotePort: strconv.Itoa(int(c.Orig.DstPort)), } } if dstLocal { return &ConnTCP{ Local: dst, LocalPort: strconv.Itoa(int(c.Orig.DstPort)), Remote: src, RemotePort: strconv.Itoa(int(c.Orig.SrcPort)), } } // Neither is local. conntrack also reports NAT connections. return nil } func parsePayload(b []byte) (*Conn, error) { // Most of this comes from libnetfilter_conntrack/src/conntrack/parse_mnl.c conn := &Conn{} var attrSpace [16]Attr attrs, err := parseAttrs(b, attrSpace[0:0]) if err != nil { return conn, err } for _, attr := range attrs { switch CtattrType(attr.Typ) { case CtaTupleOrig: parseTuple(attr.Msg, &conn.Orig) case CtaTupleReply: parseTuple(attr.Msg, &conn.Reply) case CtaCountersOrig: conn.Orig.Packets, conn.Orig.Bytes, _ = parseCounters(attr.Msg) case CtaCountersReply: conn.Reply.Packets, conn.Reply.Bytes, _ = parseCounters(attr
{ return fmt.Sprintf("%d %s %v %v mark=%d\n",c.Orig.Proto,c.TCPState,c.Orig,c.Reply,c.CtMark) }
identifier_body
mongos.go
-operator", "app.kubernetes.io/part-of": "percona-server-mongodb", } if cr.Spec.Sharding.Mongos.Labels != nil { for k, v := range cr.Spec.Sharding.Mongos.Labels { ls[k] = v } } c, err := mongosContainer(cr, customConf.Type.IsUsable(), cfgInstances) if err != nil { return appsv1.DeploymentSpec{}, fmt.Errorf("failed to create container %v", err) } initContainers := InitContainers(cr, operatorPod) for i := range initContainers { initContainers[i].Resources.Limits = c.Resources.Limits initContainers[i].Resources.Requests = c.Resources.Requests } containers, ok := cr.Spec.Sharding.Mongos.MultiAZ.WithSidecars(c) if !ok { log.Info(fmt.Sprintf("Sidecar container name cannot be %s. It's skipped", c.Name)) } annotations := cr.Spec.Sharding.Mongos.MultiAZ.Annotations if annotations == nil { annotations = make(map[string]string) } if customConf.Type.IsUsable() { annotations["percona.com/configuration-hash"] = customConf.HashHex } zero := intstr.FromInt(0) return appsv1.DeploymentSpec{ Replicas: &cr.Spec.Sharding.Mongos.Size, Selector: &metav1.LabelSelector{ MatchLabels: ls, }, Template: corev1.PodTemplateSpec{ ObjectMeta: metav1.ObjectMeta{ Labels: ls, Annotations: annotations, }, Spec: corev1.PodSpec{ SecurityContext: cr.Spec.Sharding.Mongos.PodSecurityContext, Affinity: PodAffinity(cr, cr.Spec.Sharding.Mongos.MultiAZ.Affinity, ls), NodeSelector: cr.Spec.Sharding.Mongos.MultiAZ.NodeSelector, Tolerations: cr.Spec.Sharding.Mongos.MultiAZ.Tolerations, PriorityClassName: cr.Spec.Sharding.Mongos.MultiAZ.PriorityClassName, RestartPolicy: corev1.RestartPolicyAlways, ImagePullSecrets: cr.Spec.ImagePullSecrets, Containers: containers, InitContainers: initContainers, Volumes: volumes(cr, customConf.Type), SchedulerName: cr.Spec.SchedulerName, RuntimeClassName: cr.Spec.Sharding.Mongos.MultiAZ.RuntimeClassName, }, }, Strategy: appsv1.DeploymentStrategy{ Type: appsv1.RollingUpdateDeploymentStrategyType, RollingUpdate: &appsv1.RollingUpdateDeployment{ MaxSurge: &zero, }, }, }, nil } func InitContainers(cr *api.PerconaServerMongoDB, operatorPod corev1.Pod) []corev1.Container { image := cr.Spec.InitImage if len(image) == 0 { if cr.CompareVersion(version.Version) != 0 { image = strings.Split(operatorPod.Spec.Containers[0].Image, ":")[0] + ":" + cr.Spec.CRVersion } else { image = operatorPod.Spec.Containers[0].Image } } return []corev1.Container{EntrypointInitContainer(image, cr.Spec.ImagePullPolicy)} } func
(cr *api.PerconaServerMongoDB, useConfigFile bool, cfgInstances []string) (corev1.Container, error) { fvar := false resources, err := CreateResources(cr.Spec.Sharding.Mongos.ResourcesSpec) if err != nil { return corev1.Container{}, fmt.Errorf("resource creation: %v", err) } volumes := []corev1.VolumeMount{ { Name: MongodDataVolClaimName, MountPath: MongodContainerDataDir, }, { Name: InternalKey(cr), MountPath: mongodSecretsDir, ReadOnly: true, }, { Name: "ssl", MountPath: sslDir, ReadOnly: true, }, { Name: "ssl-internal", MountPath: sslInternalDir, ReadOnly: true, }, } if useConfigFile { volumes = append(volumes, corev1.VolumeMount{ Name: "config", MountPath: mongosConfigDir, }) } if cr.CompareVersion("1.8.0") >= 0 { volumes = append(volumes, corev1.VolumeMount{ Name: "users-secret-file", MountPath: "/etc/users-secret", ReadOnly: true, }) } container := corev1.Container{ Name: "mongos", Image: cr.Spec.Image, ImagePullPolicy: cr.Spec.ImagePullPolicy, Args: mongosContainerArgs(cr, resources, useConfigFile, cfgInstances), Ports: []corev1.ContainerPort{ { Name: mongosPortName, HostPort: cr.Spec.Sharding.Mongos.HostPort, ContainerPort: cr.Spec.Sharding.Mongos.Port, }, }, Env: []corev1.EnvVar{ { Name: "MONGODB_PORT", Value: strconv.Itoa(int(cr.Spec.Sharding.Mongos.Port)), }, }, EnvFrom: []corev1.EnvFromSource{ { SecretRef: &corev1.SecretEnvSource{ LocalObjectReference: corev1.LocalObjectReference{ Name: cr.Spec.Secrets.Users, }, Optional: &fvar, }, }, { SecretRef: &corev1.SecretEnvSource{ LocalObjectReference: corev1.LocalObjectReference{ Name: api.UserSecretName(cr), }, Optional: &fvar, }, }, }, WorkingDir: MongodContainerDataDir, LivenessProbe: &cr.Spec.Sharding.Mongos.LivenessProbe.Probe, ReadinessProbe: cr.Spec.Sharding.Mongos.ReadinessProbe, SecurityContext: cr.Spec.Sharding.Mongos.ContainerSecurityContext, Resources: resources, VolumeMounts: volumes, Command: []string{"/data/db/ps-entry.sh"}, } return container, nil } func mongosContainerArgs(cr *api.PerconaServerMongoDB, resources corev1.ResourceRequirements, useConfigFile bool, cfgInstances []string) []string { mdSpec := cr.Spec.Mongod msSpec := cr.Spec.Sharding.Mongos cfgRs := cr.Spec.Sharding.ConfigsvrReplSet // sort config instances to prevent unnecessary updates sort.Strings(cfgInstances) configDB := fmt.Sprintf("%s/%s", cfgRs.Name, strings.Join(cfgInstances, ",")) args := []string{ "mongos", "--bind_ip_all", "--port=" + strconv.Itoa(int(msSpec.Port)), "--sslAllowInvalidCertificates", "--configdb", configDB, } if cr.CompareVersion("1.7.0") >= 0 { args = append(args, "--relaxPermChecks", ) } if cr.Spec.UnsafeConf { args = append(args, "--clusterAuthMode=keyFile", "--keyFile="+mongodSecretsDir+"/mongodb-key", ) } else { args = append(args, "--sslMode=preferSSL", "--clusterAuthMode=x509", ) } if mdSpec.Security != nil && mdSpec.Security.RedactClientLogData { args = append(args, "--redactClientLogData") } if msSpec.SetParameter != nil { if msSpec.SetParameter.CursorTimeoutMillis > 0 { args = append(args, "--setParameter", "cursorTimeoutMillis="+strconv.Itoa(msSpec.SetParameter.CursorTimeoutMillis), ) } } if msSpec.AuditLog != nil && msSpec.AuditLog.Destination == api.AuditLogDestinationFile { if msSpec.AuditLog.Filter == "" { msSpec.AuditLog.Filter = "{}" } args = append(args, "--auditDestination=file", "--auditFilter="+msSpec.AuditLog.Filter, "--auditFormat="+string(msSpec.AuditLog.Format), ) switch msSpec.AuditLog.Format { case api.AuditLogFormatBSON: args = append(args, "--auditPath="+MongodContainerDataDir+"/auditLog.bson") default: args = append(args, "--auditPath="+MongodContainerDataDir+"/auditLog.json") } } if useConfigFile { args = append(args, fmt.Sprintf("--config=%s/mongos.conf", mongosConfigDir)) } return args } func volumes(cr *api.PerconaServerMongoDB, configSource VolumeSourceType) []corev1.Volume { fvar, tvar := false, true volumes := []corev1.Volume{ { Name: InternalKey(cr), VolumeSource: corev1.VolumeSource{ Secret: &corev1.SecretVolumeSource{ DefaultMode: &secretFileMode, SecretName:
mongosContainer
identifier_name
mongos.go
", "app.kubernetes.io/part-of": "percona-server-mongodb", } if cr.Spec.Sharding.Mongos.Labels != nil { for k, v := range cr.Spec.Sharding.Mongos.Labels { ls[k] = v } } c, err := mongosContainer(cr, customConf.Type.IsUsable(), cfgInstances) if err != nil { return appsv1.DeploymentSpec{}, fmt.Errorf("failed to create container %v", err) } initContainers := InitContainers(cr, operatorPod) for i := range initContainers { initContainers[i].Resources.Limits = c.Resources.Limits initContainers[i].Resources.Requests = c.Resources.Requests } containers, ok := cr.Spec.Sharding.Mongos.MultiAZ.WithSidecars(c) if !ok { log.Info(fmt.Sprintf("Sidecar container name cannot be %s. It's skipped", c.Name)) } annotations := cr.Spec.Sharding.Mongos.MultiAZ.Annotations if annotations == nil { annotations = make(map[string]string) } if customConf.Type.IsUsable() { annotations["percona.com/configuration-hash"] = customConf.HashHex } zero := intstr.FromInt(0) return appsv1.DeploymentSpec{ Replicas: &cr.Spec.Sharding.Mongos.Size, Selector: &metav1.LabelSelector{ MatchLabels: ls, }, Template: corev1.PodTemplateSpec{ ObjectMeta: metav1.ObjectMeta{ Labels: ls, Annotations: annotations, }, Spec: corev1.PodSpec{ SecurityContext: cr.Spec.Sharding.Mongos.PodSecurityContext, Affinity: PodAffinity(cr, cr.Spec.Sharding.Mongos.MultiAZ.Affinity, ls), NodeSelector: cr.Spec.Sharding.Mongos.MultiAZ.NodeSelector, Tolerations: cr.Spec.Sharding.Mongos.MultiAZ.Tolerations, PriorityClassName: cr.Spec.Sharding.Mongos.MultiAZ.PriorityClassName, RestartPolicy: corev1.RestartPolicyAlways, ImagePullSecrets: cr.Spec.ImagePullSecrets, Containers: containers, InitContainers: initContainers, Volumes: volumes(cr, customConf.Type), SchedulerName: cr.Spec.SchedulerName, RuntimeClassName: cr.Spec.Sharding.Mongos.MultiAZ.RuntimeClassName, }, }, Strategy: appsv1.DeploymentStrategy{ Type: appsv1.RollingUpdateDeploymentStrategyType, RollingUpdate: &appsv1.RollingUpdateDeployment{ MaxSurge: &zero, }, }, }, nil } func InitContainers(cr *api.PerconaServerMongoDB, operatorPod corev1.Pod) []corev1.Container { image := cr.Spec.InitImage if len(image) == 0 { if cr.CompareVersion(version.Version) != 0 { image = strings.Split(operatorPod.Spec.Containers[0].Image, ":")[0] + ":" + cr.Spec.CRVersion } else { image = operatorPod.Spec.Containers[0].Image } } return []corev1.Container{EntrypointInitContainer(image, cr.Spec.ImagePullPolicy)} } func mongosContainer(cr *api.PerconaServerMongoDB, useConfigFile bool, cfgInstances []string) (corev1.Container, error) { fvar := false resources, err := CreateResources(cr.Spec.Sharding.Mongos.ResourcesSpec) if err != nil { return corev1.Container{}, fmt.Errorf("resource creation: %v", err) } volumes := []corev1.VolumeMount{ { Name: MongodDataVolClaimName, MountPath: MongodContainerDataDir, }, { Name: InternalKey(cr), MountPath: mongodSecretsDir, ReadOnly: true, }, { Name: "ssl", MountPath: sslDir, ReadOnly: true, }, { Name: "ssl-internal", MountPath: sslInternalDir, ReadOnly: true, }, } if useConfigFile { volumes = append(volumes, corev1.VolumeMount{ Name: "config", MountPath: mongosConfigDir, }) } if cr.CompareVersion("1.8.0") >= 0 { volumes = append(volumes, corev1.VolumeMount{ Name: "users-secret-file", MountPath: "/etc/users-secret", ReadOnly: true, }) } container := corev1.Container{ Name: "mongos", Image: cr.Spec.Image, ImagePullPolicy: cr.Spec.ImagePullPolicy, Args: mongosContainerArgs(cr, resources, useConfigFile, cfgInstances), Ports: []corev1.ContainerPort{ { Name: mongosPortName, HostPort: cr.Spec.Sharding.Mongos.HostPort, ContainerPort: cr.Spec.Sharding.Mongos.Port, }, }, Env: []corev1.EnvVar{ { Name: "MONGODB_PORT", Value: strconv.Itoa(int(cr.Spec.Sharding.Mongos.Port)), }, }, EnvFrom: []corev1.EnvFromSource{ { SecretRef: &corev1.SecretEnvSource{ LocalObjectReference: corev1.LocalObjectReference{ Name: cr.Spec.Secrets.Users, }, Optional: &fvar, }, }, { SecretRef: &corev1.SecretEnvSource{ LocalObjectReference: corev1.LocalObjectReference{ Name: api.UserSecretName(cr), }, Optional: &fvar, }, }, }, WorkingDir: MongodContainerDataDir, LivenessProbe: &cr.Spec.Sharding.Mongos.LivenessProbe.Probe, ReadinessProbe: cr.Spec.Sharding.Mongos.ReadinessProbe, SecurityContext: cr.Spec.Sharding.Mongos.ContainerSecurityContext, Resources: resources, VolumeMounts: volumes, Command: []string{"/data/db/ps-entry.sh"}, } return container, nil } func mongosContainerArgs(cr *api.PerconaServerMongoDB, resources corev1.ResourceRequirements, useConfigFile bool, cfgInstances []string) []string { mdSpec := cr.Spec.Mongod msSpec := cr.Spec.Sharding.Mongos cfgRs := cr.Spec.Sharding.ConfigsvrReplSet // sort config instances to prevent unnecessary updates sort.Strings(cfgInstances) configDB := fmt.Sprintf("%s/%s", cfgRs.Name, strings.Join(cfgInstances, ",")) args := []string{ "mongos", "--bind_ip_all", "--port=" + strconv.Itoa(int(msSpec.Port)), "--sslAllowInvalidCertificates", "--configdb", configDB, } if cr.CompareVersion("1.7.0") >= 0 { args = append(args, "--relaxPermChecks", ) } if cr.Spec.UnsafeConf { args = append(args, "--clusterAuthMode=keyFile", "--keyFile="+mongodSecretsDir+"/mongodb-key", ) } else { args = append(args, "--sslMode=preferSSL", "--clusterAuthMode=x509", ) } if mdSpec.Security != nil && mdSpec.Security.RedactClientLogData { args = append(args, "--redactClientLogData") } if msSpec.SetParameter != nil { if msSpec.SetParameter.CursorTimeoutMillis > 0 { args = append(args, "--setParameter", "cursorTimeoutMillis="+strconv.Itoa(msSpec.SetParameter.CursorTimeoutMillis), ) } } if msSpec.AuditLog != nil && msSpec.AuditLog.Destination == api.AuditLogDestinationFile { if msSpec.AuditLog.Filter == "" { msSpec.AuditLog.Filter = "{}" } args = append(args, "--auditDestination=file", "--auditFilter="+msSpec.AuditLog.Filter, "--auditFormat="+string(msSpec.AuditLog.Format), ) switch msSpec.AuditLog.Format { case api.AuditLogFormatBSON: args = append(args, "--auditPath="+MongodContainerDataDir+"/auditLog.bson") default: args = append(args, "--auditPath="+MongodContainerDataDir+"/auditLog.json") } } if useConfigFile
return args } func volumes(cr *api.PerconaServerMongoDB, configSource VolumeSourceType) []corev1.Volume { fvar, tvar := false, true volumes := []corev1.Volume{ { Name: InternalKey(cr), VolumeSource: corev1.VolumeSource{ Secret: &corev1.SecretVolumeSource{ DefaultMode: &secretFileMode, SecretName:
{ args = append(args, fmt.Sprintf("--config=%s/mongos.conf", mongosConfigDir)) }
conditional_block
mongos.go
), SchedulerName: cr.Spec.SchedulerName, RuntimeClassName: cr.Spec.Sharding.Mongos.MultiAZ.RuntimeClassName, }, }, Strategy: appsv1.DeploymentStrategy{ Type: appsv1.RollingUpdateDeploymentStrategyType, RollingUpdate: &appsv1.RollingUpdateDeployment{ MaxSurge: &zero, }, }, }, nil } func InitContainers(cr *api.PerconaServerMongoDB, operatorPod corev1.Pod) []corev1.Container { image := cr.Spec.InitImage if len(image) == 0 { if cr.CompareVersion(version.Version) != 0 { image = strings.Split(operatorPod.Spec.Containers[0].Image, ":")[0] + ":" + cr.Spec.CRVersion } else { image = operatorPod.Spec.Containers[0].Image } } return []corev1.Container{EntrypointInitContainer(image, cr.Spec.ImagePullPolicy)} } func mongosContainer(cr *api.PerconaServerMongoDB, useConfigFile bool, cfgInstances []string) (corev1.Container, error) { fvar := false resources, err := CreateResources(cr.Spec.Sharding.Mongos.ResourcesSpec) if err != nil { return corev1.Container{}, fmt.Errorf("resource creation: %v", err) } volumes := []corev1.VolumeMount{ { Name: MongodDataVolClaimName, MountPath: MongodContainerDataDir, }, { Name: InternalKey(cr), MountPath: mongodSecretsDir, ReadOnly: true, }, { Name: "ssl", MountPath: sslDir, ReadOnly: true, }, { Name: "ssl-internal", MountPath: sslInternalDir, ReadOnly: true, }, } if useConfigFile { volumes = append(volumes, corev1.VolumeMount{ Name: "config", MountPath: mongosConfigDir, }) } if cr.CompareVersion("1.8.0") >= 0 { volumes = append(volumes, corev1.VolumeMount{ Name: "users-secret-file", MountPath: "/etc/users-secret", ReadOnly: true, }) } container := corev1.Container{ Name: "mongos", Image: cr.Spec.Image, ImagePullPolicy: cr.Spec.ImagePullPolicy, Args: mongosContainerArgs(cr, resources, useConfigFile, cfgInstances), Ports: []corev1.ContainerPort{ { Name: mongosPortName, HostPort: cr.Spec.Sharding.Mongos.HostPort, ContainerPort: cr.Spec.Sharding.Mongos.Port, }, }, Env: []corev1.EnvVar{ { Name: "MONGODB_PORT", Value: strconv.Itoa(int(cr.Spec.Sharding.Mongos.Port)), }, }, EnvFrom: []corev1.EnvFromSource{ { SecretRef: &corev1.SecretEnvSource{ LocalObjectReference: corev1.LocalObjectReference{ Name: cr.Spec.Secrets.Users, }, Optional: &fvar, }, }, { SecretRef: &corev1.SecretEnvSource{ LocalObjectReference: corev1.LocalObjectReference{ Name: api.UserSecretName(cr), }, Optional: &fvar, }, }, }, WorkingDir: MongodContainerDataDir, LivenessProbe: &cr.Spec.Sharding.Mongos.LivenessProbe.Probe, ReadinessProbe: cr.Spec.Sharding.Mongos.ReadinessProbe, SecurityContext: cr.Spec.Sharding.Mongos.ContainerSecurityContext, Resources: resources, VolumeMounts: volumes, Command: []string{"/data/db/ps-entry.sh"}, } return container, nil } func mongosContainerArgs(cr *api.PerconaServerMongoDB, resources corev1.ResourceRequirements, useConfigFile bool, cfgInstances []string) []string { mdSpec := cr.Spec.Mongod msSpec := cr.Spec.Sharding.Mongos cfgRs := cr.Spec.Sharding.ConfigsvrReplSet // sort config instances to prevent unnecessary updates sort.Strings(cfgInstances) configDB := fmt.Sprintf("%s/%s", cfgRs.Name, strings.Join(cfgInstances, ",")) args := []string{ "mongos", "--bind_ip_all", "--port=" + strconv.Itoa(int(msSpec.Port)), "--sslAllowInvalidCertificates", "--configdb", configDB, } if cr.CompareVersion("1.7.0") >= 0 { args = append(args, "--relaxPermChecks", ) } if cr.Spec.UnsafeConf { args = append(args, "--clusterAuthMode=keyFile", "--keyFile="+mongodSecretsDir+"/mongodb-key", ) } else { args = append(args, "--sslMode=preferSSL", "--clusterAuthMode=x509", ) } if mdSpec.Security != nil && mdSpec.Security.RedactClientLogData { args = append(args, "--redactClientLogData") } if msSpec.SetParameter != nil { if msSpec.SetParameter.CursorTimeoutMillis > 0 { args = append(args, "--setParameter", "cursorTimeoutMillis="+strconv.Itoa(msSpec.SetParameter.CursorTimeoutMillis), ) } } if msSpec.AuditLog != nil && msSpec.AuditLog.Destination == api.AuditLogDestinationFile { if msSpec.AuditLog.Filter == "" { msSpec.AuditLog.Filter = "{}" } args = append(args, "--auditDestination=file", "--auditFilter="+msSpec.AuditLog.Filter, "--auditFormat="+string(msSpec.AuditLog.Format), ) switch msSpec.AuditLog.Format { case api.AuditLogFormatBSON: args = append(args, "--auditPath="+MongodContainerDataDir+"/auditLog.bson") default: args = append(args, "--auditPath="+MongodContainerDataDir+"/auditLog.json") } } if useConfigFile { args = append(args, fmt.Sprintf("--config=%s/mongos.conf", mongosConfigDir)) } return args } func volumes(cr *api.PerconaServerMongoDB, configSource VolumeSourceType) []corev1.Volume { fvar, tvar := false, true volumes := []corev1.Volume{ { Name: InternalKey(cr), VolumeSource: corev1.VolumeSource{ Secret: &corev1.SecretVolumeSource{ DefaultMode: &secretFileMode, SecretName: InternalKey(cr), Optional: &fvar, }, }, }, { Name: "ssl", VolumeSource: corev1.VolumeSource{ Secret: &corev1.SecretVolumeSource{ SecretName: cr.Spec.Secrets.SSL, Optional: &cr.Spec.UnsafeConf, DefaultMode: &secretFileMode, }, }, }, { Name: "ssl-internal", VolumeSource: corev1.VolumeSource{ Secret: &corev1.SecretVolumeSource{ SecretName: cr.Spec.Secrets.SSLInternal, Optional: &tvar, DefaultMode: &secretFileMode, }, }, }, { Name: MongodDataVolClaimName, VolumeSource: corev1.VolumeSource{ EmptyDir: &corev1.EmptyDirVolumeSource{}, }, }, } if cr.CompareVersion("1.8.0") >= 0 { volumes = append(volumes, corev1.Volume{ Name: "users-secret-file", VolumeSource: corev1.VolumeSource{ Secret: &corev1.SecretVolumeSource{ SecretName: api.InternalUserSecretName(cr), }, }, }) } if configSource.IsUsable() { volumes = append(volumes, corev1.Volume{ Name: "config", VolumeSource: configSource.VolumeSource(MongosCustomConfigName(cr.Name)), }) } return volumes } func MongosService(cr *api.PerconaServerMongoDB) corev1.Service { svc := corev1.Service{ TypeMeta: metav1.TypeMeta{ APIVersion: "v1", Kind: "Service", }, ObjectMeta: metav1.ObjectMeta{ Name: cr.Name + "-" + "mongos", Namespace: cr.Namespace, }, } if cr.Spec.Sharding.Mongos != nil { svc.Annotations = cr.Spec.Sharding.Mongos.Expose.ServiceAnnotations } return svc } func MongosServiceSpec(cr *api.PerconaServerMongoDB) corev1.ServiceSpec { ls := map[string]string{
"app.kubernetes.io/name": "percona-server-mongodb",
random_line_split
mongos.go
.IsUsable(), cfgInstances) if err != nil { return appsv1.DeploymentSpec{}, fmt.Errorf("failed to create container %v", err) } initContainers := InitContainers(cr, operatorPod) for i := range initContainers { initContainers[i].Resources.Limits = c.Resources.Limits initContainers[i].Resources.Requests = c.Resources.Requests } containers, ok := cr.Spec.Sharding.Mongos.MultiAZ.WithSidecars(c) if !ok { log.Info(fmt.Sprintf("Sidecar container name cannot be %s. It's skipped", c.Name)) } annotations := cr.Spec.Sharding.Mongos.MultiAZ.Annotations if annotations == nil { annotations = make(map[string]string) } if customConf.Type.IsUsable() { annotations["percona.com/configuration-hash"] = customConf.HashHex } zero := intstr.FromInt(0) return appsv1.DeploymentSpec{ Replicas: &cr.Spec.Sharding.Mongos.Size, Selector: &metav1.LabelSelector{ MatchLabels: ls, }, Template: corev1.PodTemplateSpec{ ObjectMeta: metav1.ObjectMeta{ Labels: ls, Annotations: annotations, }, Spec: corev1.PodSpec{ SecurityContext: cr.Spec.Sharding.Mongos.PodSecurityContext, Affinity: PodAffinity(cr, cr.Spec.Sharding.Mongos.MultiAZ.Affinity, ls), NodeSelector: cr.Spec.Sharding.Mongos.MultiAZ.NodeSelector, Tolerations: cr.Spec.Sharding.Mongos.MultiAZ.Tolerations, PriorityClassName: cr.Spec.Sharding.Mongos.MultiAZ.PriorityClassName, RestartPolicy: corev1.RestartPolicyAlways, ImagePullSecrets: cr.Spec.ImagePullSecrets, Containers: containers, InitContainers: initContainers, Volumes: volumes(cr, customConf.Type), SchedulerName: cr.Spec.SchedulerName, RuntimeClassName: cr.Spec.Sharding.Mongos.MultiAZ.RuntimeClassName, }, }, Strategy: appsv1.DeploymentStrategy{ Type: appsv1.RollingUpdateDeploymentStrategyType, RollingUpdate: &appsv1.RollingUpdateDeployment{ MaxSurge: &zero, }, }, }, nil } func InitContainers(cr *api.PerconaServerMongoDB, operatorPod corev1.Pod) []corev1.Container { image := cr.Spec.InitImage if len(image) == 0 { if cr.CompareVersion(version.Version) != 0 { image = strings.Split(operatorPod.Spec.Containers[0].Image, ":")[0] + ":" + cr.Spec.CRVersion } else { image = operatorPod.Spec.Containers[0].Image } } return []corev1.Container{EntrypointInitContainer(image, cr.Spec.ImagePullPolicy)} } func mongosContainer(cr *api.PerconaServerMongoDB, useConfigFile bool, cfgInstances []string) (corev1.Container, error) { fvar := false resources, err := CreateResources(cr.Spec.Sharding.Mongos.ResourcesSpec) if err != nil { return corev1.Container{}, fmt.Errorf("resource creation: %v", err) } volumes := []corev1.VolumeMount{ { Name: MongodDataVolClaimName, MountPath: MongodContainerDataDir, }, { Name: InternalKey(cr), MountPath: mongodSecretsDir, ReadOnly: true, }, { Name: "ssl", MountPath: sslDir, ReadOnly: true, }, { Name: "ssl-internal", MountPath: sslInternalDir, ReadOnly: true, }, } if useConfigFile { volumes = append(volumes, corev1.VolumeMount{ Name: "config", MountPath: mongosConfigDir, }) } if cr.CompareVersion("1.8.0") >= 0 { volumes = append(volumes, corev1.VolumeMount{ Name: "users-secret-file", MountPath: "/etc/users-secret", ReadOnly: true, }) } container := corev1.Container{ Name: "mongos", Image: cr.Spec.Image, ImagePullPolicy: cr.Spec.ImagePullPolicy, Args: mongosContainerArgs(cr, resources, useConfigFile, cfgInstances), Ports: []corev1.ContainerPort{ { Name: mongosPortName, HostPort: cr.Spec.Sharding.Mongos.HostPort, ContainerPort: cr.Spec.Sharding.Mongos.Port, }, }, Env: []corev1.EnvVar{ { Name: "MONGODB_PORT", Value: strconv.Itoa(int(cr.Spec.Sharding.Mongos.Port)), }, }, EnvFrom: []corev1.EnvFromSource{ { SecretRef: &corev1.SecretEnvSource{ LocalObjectReference: corev1.LocalObjectReference{ Name: cr.Spec.Secrets.Users, }, Optional: &fvar, }, }, { SecretRef: &corev1.SecretEnvSource{ LocalObjectReference: corev1.LocalObjectReference{ Name: api.UserSecretName(cr), }, Optional: &fvar, }, }, }, WorkingDir: MongodContainerDataDir, LivenessProbe: &cr.Spec.Sharding.Mongos.LivenessProbe.Probe, ReadinessProbe: cr.Spec.Sharding.Mongos.ReadinessProbe, SecurityContext: cr.Spec.Sharding.Mongos.ContainerSecurityContext, Resources: resources, VolumeMounts: volumes, Command: []string{"/data/db/ps-entry.sh"}, } return container, nil } func mongosContainerArgs(cr *api.PerconaServerMongoDB, resources corev1.ResourceRequirements, useConfigFile bool, cfgInstances []string) []string { mdSpec := cr.Spec.Mongod msSpec := cr.Spec.Sharding.Mongos cfgRs := cr.Spec.Sharding.ConfigsvrReplSet // sort config instances to prevent unnecessary updates sort.Strings(cfgInstances) configDB := fmt.Sprintf("%s/%s", cfgRs.Name, strings.Join(cfgInstances, ",")) args := []string{ "mongos", "--bind_ip_all", "--port=" + strconv.Itoa(int(msSpec.Port)), "--sslAllowInvalidCertificates", "--configdb", configDB, } if cr.CompareVersion("1.7.0") >= 0 { args = append(args, "--relaxPermChecks", ) } if cr.Spec.UnsafeConf { args = append(args, "--clusterAuthMode=keyFile", "--keyFile="+mongodSecretsDir+"/mongodb-key", ) } else { args = append(args, "--sslMode=preferSSL", "--clusterAuthMode=x509", ) } if mdSpec.Security != nil && mdSpec.Security.RedactClientLogData { args = append(args, "--redactClientLogData") } if msSpec.SetParameter != nil { if msSpec.SetParameter.CursorTimeoutMillis > 0 { args = append(args, "--setParameter", "cursorTimeoutMillis="+strconv.Itoa(msSpec.SetParameter.CursorTimeoutMillis), ) } } if msSpec.AuditLog != nil && msSpec.AuditLog.Destination == api.AuditLogDestinationFile { if msSpec.AuditLog.Filter == "" { msSpec.AuditLog.Filter = "{}" } args = append(args, "--auditDestination=file", "--auditFilter="+msSpec.AuditLog.Filter, "--auditFormat="+string(msSpec.AuditLog.Format), ) switch msSpec.AuditLog.Format { case api.AuditLogFormatBSON: args = append(args, "--auditPath="+MongodContainerDataDir+"/auditLog.bson") default: args = append(args, "--auditPath="+MongodContainerDataDir+"/auditLog.json") } } if useConfigFile { args = append(args, fmt.Sprintf("--config=%s/mongos.conf", mongosConfigDir)) } return args } func volumes(cr *api.PerconaServerMongoDB, configSource VolumeSourceType) []corev1.Volume
{ fvar, tvar := false, true volumes := []corev1.Volume{ { Name: InternalKey(cr), VolumeSource: corev1.VolumeSource{ Secret: &corev1.SecretVolumeSource{ DefaultMode: &secretFileMode, SecretName: InternalKey(cr), Optional: &fvar, }, }, }, { Name: "ssl", VolumeSource: corev1.VolumeSource{ Secret: &corev1.SecretVolumeSource{ SecretName: cr.Spec.Secrets.SSL, Optional: &cr.Spec.UnsafeConf,
identifier_body
gridview.rs
ust_planned].commands_to_finalize.push(cmd) } pub fn rollback(&mut self, new_snapshot: &Snapshot) { let old_planned: Vec<_> = self.planned.drain(..).collect(); self.planned .push_back(new_snapshot.new_with_same_droplets()); assert_eq!(self.planned.len(), 1); for planned_snapshot in old_planned { planned_snapshot.abort(self) } } pub fn perturb(&self, rng: &mut impl Rng, snapshot: &Snapshot) -> Option<Snapshot> { let now = snapshot; let then = self.completed.last()?; let id = { let ids: Vec<_> = now.droplets.keys().collect(); match rng.choose(ids.as_slice()) { Some(&&id) => id, None => return None, } }; let mut now2 = now.new_with_same_droplets(); if let Some(old_droplet) = then.droplets.get(&id) { let was_there = now2.droplets.insert(id, old_droplet.clone()); assert!(was_there.is_some()); } Some(now2) } pub fn add_error_edges(&mut self, planned: &Snapshot, actual: &Snapshot) { let previous = self.completed.last().unwrap(); let edges = previous.get_error_edges(planned, actual); let n_edges = edges.len(); warn!( "Added error {} edges, now there are {}: {:?}", n_edges, self.bad_edges.len() / 2, edges, ); for (loc1, loc2) in edges { // for now, insert edges both ways self.bad_edges.insert((loc1, loc2)); self.bad_edges.insert((loc2, loc1)); } } } pub struct GridSubView<'a> { backing_gridview: &'a mut GridView, mapping: Map<Location, Location>, ids: Set<DropletId>, } impl<'a> GridSubView<'a> { pub fn tick(&mut self) { self.backing_gridview.tick() } #[cfg(feature = "pi")] pub fn with_pi<T>(&mut self, f: impl FnOnce(&mut RaspberryPi) -> T) -> Option<T> { self.backing_gridview.pi.as_mut().map(f) } pub fn get_electrode(&self, loc: &Location) -> Option<&Electrode> { let actual_loc = self.mapping.get(loc)?; self.backing_gridview.grid.get_cell(&actual_loc) } // TODO: translate or somehow hide the untranslated location of this pub fn get(&self, id: &DropletId) -> &Droplet { assert!(self.ids.contains(&id)); &self.backing_gridview.snapshot().droplets[id] } fn get_mut(&mut self, id: &DropletId) -> &mut Droplet { assert!(self.ids.contains(&id)); self.backing_gridview .snapshot_mut() .droplets .get_mut(id) .unwrap() } pub fn insert(&mut self, mut droplet: Droplet) { let new_loc = self.mapping.get(&droplet.location); trace!("Inserting {:#?} at {:?}", droplet, new_loc); droplet.location = *new_loc.unwrap(); let was_not_there = self.ids.insert(droplet.id); assert!(was_not_there); let snapshot = self.backing_gridview.snapshot_mut(); let was_there = snapshot.droplets.insert(droplet.id, droplet); assert!(was_there.is_none()); } pub fn remove(&mut self, id: &DropletId) -> Droplet { let was_there = self.ids.remove(id); assert!(was_there); let snapshot = self.backing_gridview.snapshot_mut(); let mut droplet = snapshot.droplets.remove(id).unwrap(); // FIXME this is pretty dumb let (unmapped_loc, _) = self .mapping .iter() .find(|(_, &v)| v == droplet.location) .unwrap(); droplet.location = *unmapped_loc; droplet } fn check_droplet(&self, id: &DropletId) { // TODO will this have translated or real location?? let droplet = self.get(id); let mapped_to: Set<_> = self.mapping.values().collect(); // TODO this is pretty slow for i in 0..droplet.dimensions.y { for j in 0..droplet.dimensions.x { let loc = Location { y: droplet.location.y + i, x: droplet.location.x + j, }; if !mapped_to.contains(&loc) { panic!("{} was unmapped!, mapping: {:#?}", loc, self.mapping); } } } } fn update(&mut self, id: &DropletId, func: impl FnOnce(&mut Droplet)) { func(self.get_mut(id)); self.check_droplet(id); } pub fn move_west(&mut self, id: DropletId) { trace!("Moving droplet {:?} west", id); self.update(&id, |droplet| { droplet.location = droplet.location.west(); }) } pub fn move_east(&mut self, id: DropletId) { trace!("Moving droplet {:?} east", id); self.update(&id, |droplet| { droplet.location = droplet.location.east(); }) } pub fn move_north(&mut self, id: DropletId) { trace!("Moving droplet {:?} north", id); self.update(&id, |droplet| { droplet.location = droplet.location.north(); }) } pub fn move_south(&mut self, id: DropletId) { trace!("Moving droplet {:?} south", id); self.update(&id, |droplet| { droplet.location = droplet.location.south(); }) } } #[cfg(test)] pub mod tests { use super::*; use grid::parse::tests::parse_strings; pub fn id2c(id: &DropletId) -> char { assert!(id.id < 255); (id.id as u8) as char } pub fn c2id(c: char) -> DropletId { for u in 0x00u8..0xff { let c2 = u as char; if c == c2 { return DropletId { id: u as usize, process_id: 0, }; } } panic!("Can't make {} a u8", c); } pub fn parse_gridview(strs: &[&str]) -> GridView { // same chars are guaranteed to have the same ids let (grid, blobs) = parse_strings(&strs); let mut snapshot = Snapshot::default(); for (ch, blob) in blobs.iter() { let id = c2id(*ch); snapshot.droplets.insert(id, blob.to_droplet(id)); } let mut gv = GridView::new(grid); gv.planned[0] = snapshot; gv } pub fn parse_snapshot(strs: &[&str]) -> Snapshot { let mut gv = parse_gridview(strs); gv.planned.remove(0).unwrap() } fn check_all_matched( snapshot_strs: &[&str], blob_strs: &[&str], ) -> Option<Map<DropletId, SimpleBlob>> { let snapshot = parse_snapshot(&snapshot_strs); let (_, chip_blobs) = parse_strings(&blob_strs); let blobs: Vec<SimpleBlob> = chip_blobs.values().cloned().collect(); let result: Map<DropletId, SimpleBlob> = snapshot.match_with_blobs(&blobs)?; // create the expected map by mapping the ids in the snapshot // to the associated blob which corresponds to the character let mut expected: Map<DropletId, SimpleBlob> = Map::new(); for id in snapshot.droplets.keys() { expected.insert(*id, chip_blobs[&id2c(id)].clone()); } for id in expected.keys() { // we can't compare blobs or droplets, so we get the droplet_info assert_eq!( result.get(id).map(|blob| blob.to_droplet(*id).info()), expected.get(id).map(|blob| blob.to_droplet(*id).info()) ) } Some(result) } #[test] fn test_no_diff() { let strs = vec![ "aa..........c", ".....bb......", ".............", ".............", ]; assert!(check_all_matched(&strs, &strs).is_some()); } #[test] fn test_location_diff() { let exec_strs = vec![ "aa..........c", ".....bb......", ".............", ".............", ]; let chip_strs = vec![ "aa...........", "............c", ".....bb......", ".............",
];
random_line_split
gridview.rs
let mut ids = vec![]; // store corresponding ids to indeces let mut matches = vec![]; // store similarity between blobs/droplets let n = blobs.len(); // store the id of each droplet in its corresponding // index in 'ids', then store the similarity of each // droplet to each blob in 'matches' for (&id, droplet) in &self.droplets { ids.push(id); for blob in blobs { let similarity = blob.get_similarity(&droplet); // must be non-negative for the algorithm to work assert!(similarity >= 0); matches.push(similarity); } } // convert the matches vector to a matrix // input should be [1,2,3,4], where the output // matrix is [[1,2],[3,4]] let m: Matrix<i32> = Matrix::from_vec(n, n, matches); // km is a vector of size n where the value at each index // corresponds to the index of a blob let (_c, km) = kuhn_munkres_min(&m); for i in 0..n { result.insert(ids[i], blobs[km[i]].clone()); } Some(result) } // this will take commands_to_finalize from the old snapshot into the new // one if an error is found produced pub fn correct(&mut self, blobs: &[impl Blob]) -> Option<Snapshot> { let blob_matching = self.match_with_blobs(blobs)?; let mut was_error = false; let new_droplets: Map<_, _> = blob_matching .iter() .map(|(&id, blob)| { let d = self.droplets.get_mut(&id).unwrap(); let d_new = blob.to_droplet(id); if d.location != d_new.location || d.dimensions != d_new.dimensions { info!("Found error in droplet {:?}", id); debug!("Droplet error\n Expected: {:#?}\n Found: {:#?}", d, d_new); was_error = true; } // HACK FIXME this mutation is not great if (d.volume - d_new.volume).abs() > 1.0 { info!( "volume of {} changed: {} -> {}", id.id, d.volume, d_new.volume ) } d.volume = d_new.volume; (id, d_new) }).collect(); if was_error { let mut new_snapshot = Snapshot { droplets: new_droplets, commands_to_finalize: Vec::new(), }; ::std::mem::swap( &mut new_snapshot.commands_to_finalize, &mut self.commands_to_finalize, ); Some(new_snapshot) } else { None } } pub fn diff_droplet(&self, id: &DropletId, other: &Snapshot) -> DropletDiff { use self::DropletDiff::*; let droplet = self .droplets .get(id) .expect("id should be in self snapshot"); if let Some(other_droplet) = other.droplets.get(id) { // NOTE we only care about location diffs for now let loc = droplet.location; let other_loc = other_droplet.location; if loc != other_loc { // for now, just assert that we are only moving one spot at a time // FIXME HACK // assert_eq!((&loc - &other_loc).norm(), 1); Moved { from: loc, to: other_loc, } } else { DidNotMove } } else { Disappeared } } pub fn get_error_edges( &self, planned_outcome: &Snapshot, actual_outcome: &Snapshot, ) -> Vec<(Location, Location)> { use self::DropletDiff::*; self.droplets .keys() .filter_map(|id| { let planned_diff = self.diff_droplet(id, planned_outcome); let actual_diff = self.diff_droplet(id, actual_outcome); match (planned_diff, actual_diff) { (Moved { from, to }, DidNotMove) => { if (&from - &to).norm() == 1 { Some((from, to)) } else { warn!("Droplet {} jumped from {} to {}!", id.id, from, to); None } } _ => None, } }).collect() } } #[derive(Debug)] pub enum ExecResponse { Step(Snapshot), NotReady, Done, } impl GridView { pub fn new(grid: Grid) -> GridView { let mut planned = VecDeque::new(); planned.push_back(Snapshot::default()); #[cfg(feature = "pi")] let pi = match ::std::env::var("PUDDLE_PI") { Ok(s) => if s == "1" { let mut pi = RaspberryPi::new().unwrap(); info!("Initialized the pi!"); Some(pi) } else { warn!("Couldn't read PUDDLE_PI={}", s); None }, Err(_) => { info!("Did not start the pi!"); None } }; GridView { grid: grid, planned, completed: Vec::new(), done: false, bad_edges: Set::new(), #[cfg(feature = "pi")] pi, } } pub fn close(&mut self) { info!("Marking gridview as DONE!"); self.done = true; } pub fn execute(&mut self) -> ExecResponse { use self::ExecResponse::*; // compare with len - 1 because we wouldn't want to "write out" a state // that hasn't been fully planned let resp = if let Some(planned_snapshot) = self.planned.pop_front() { Step(planned_snapshot) } else if self.done { Done } else { NotReady }; trace!( "execute sending {:?}. Completed: {}, planned: {}.", resp, self.completed.len(), self.planned.len(), ); resp } pub fn commit_pending(&mut self, mut snapshot: Snapshot) { #[cfg(not(feature = "pi"))] snapshot.finalize(); #[cfg(feature = "pi")] snapshot.finalize(self.pi.as_mut()); self.completed.push(snapshot); } pub fn snapshot(&self) -> &Snapshot { self.planned.back().unwrap() } // TODO probably shouldn't provide this pub fn snapshot_mut(&mut self) -> &mut Snapshot
pub fn snapshot_ensure(&mut self) { if self.planned.is_empty() { let last = self.completed.last().unwrap(); self.planned.push_back(last.new_with_same_droplets()) } } pub fn exec_snapshot(&self) -> &Snapshot { self.completed.last().unwrap() } fn tick(&mut self) { let new_snapshot = { let just_planned = self.planned.back().unwrap(); if let Some(col) = just_planned.get_collision() { panic!("collision: {:#?}", col); }; just_planned.new_with_same_droplets() }; self.planned.push_back(new_snapshot); trace!("TICK! len={}", self.planned.len()); } fn update(&mut self, id: DropletId, func: impl FnOnce(&mut Droplet)) { let now = self.planned.back_mut().unwrap(); let droplet = now .droplets .get_mut(&id) .unwrap_or_else(|| panic!("Tried to remove a non-existent droplet: {:?}", id)); func(droplet); } pub fn plan_droplet_info(&self, pid_option: Option<ProcessId>) -> Vec<DropletInfo> { // gets from the planner for now self.planned.back().unwrap().droplet_info(pid_option) } pub fn take_paths(&mut self, paths: &Map<DropletId, Path>, final_tick: bool) { let max_len = paths.values().map(|path| path.len()).max().unwrap_or(0); // make sure that all droplets start where they are at this time step for (id, path) in paths.iter() { let snapshot = self.planned.back().unwrap(); let droplet = &snapshot.droplets[&id]; assert_eq!(droplet.location, path[0]); } for i in 1..max_len { for (&id, path) in paths.iter() { if i < path.len() { self.update(id, |droplet| { assert!(droplet.location.distance_to(&path[i]) <= 1); droplet.location = path[i]; }); } } if i < max_len - 1 || final_tick { self.tick(); } } } pub fn subview( &mut self, ids: impl IntoIterator<Item = DropletId>, mapping: Map<Location, Location>, ) -> GridSubView { GridSubView { backing_gridview: self, mapping: mapping, ids: ids.into
{ self.planned.back_mut().unwrap() }
identifier_body
gridview.rs
{ Disappeared, DidNotMove, Moved { from: Location, to: Location }, } impl Snapshot { pub fn new_with_same_droplets(&self) -> Snapshot { let mut new_snapshot = Snapshot::default(); new_snapshot.droplets = self.droplets.clone(); // clear out the destination because we're doing to replan for d in new_snapshot.droplets.values_mut() { d.destination = None; } new_snapshot } #[cfg(not(feature = "pi"))] fn finalize(&mut self) { // we need to drain this so we can mutate the command without mutating // self, as we need to pass self into cmd.finalize // this feels pretty ugly.... let mut x: Vec<_> = self.commands_to_finalize.drain(..).collect(); for cmd in &mut x { debug!("Finalizing command: {:#?}", cmd); cmd.finalize(self) } self.commands_to_finalize = x; } #[cfg(feature = "pi")] fn finalize(&mut self, pi: Option<&mut RaspberryPi>) { // we need to drain this so we can mutate the command without mutating // self, as we need to pass self into cmd.finalize // this feels pretty ugly.... let mut x: Vec<_> = self.commands_to_finalize.drain(..).collect(); if let Some(pi) = pi { for cmd in &mut x { debug!("Finalizing command: {:#?}", cmd); cmd.finalize(self, Some(pi)) } } else { for cmd in &mut x { debug!("Finalizing command: {:#?}", cmd); cmd.finalize(self, None) } } self.commands_to_finalize = x; } pub fn abort(mut self, gridview: &mut GridView) { for mut cmd in self.commands_to_finalize.drain(..) { debug!("Sending command back for replanning: {:#?}", cmd); if let Err((mut cmd, err)) = gridview.plan(cmd) { cmd.abort(err); } } } pub fn droplet_info(&self, pid_option: Option<ProcessId>) -> Vec<DropletInfo> { self.droplets .values() .filter(|&d| pid_option.map_or(true, |pid| d.id.process_id == pid)) .map(|d| d.info()) .collect() } /// Returns an invalid droplet, if any. fn get_collision(&self) -> Option<(i32, Droplet, Droplet)> { for (id1, droplet1) in &self.droplets { for (id2, droplet2) in &self.droplets { if id1 == id2 { continue; } if droplet1.collision_group == droplet2.collision_group { continue; } let distance = droplet1.collision_distance(droplet2); if distance <= 0 { return Some((distance, droplet1.clone(), droplet2.clone())); } } } None } pub fn to_blobs(&self) -> Vec<SimpleBlob> { self.droplets.values().map(|d| d.to_blob()).collect() } /// Takes a map of droplet ids to droplets (as in that /// of the planner/executor view) and a vector of blobs /// (as in that of the chip view) and returns a matching /// of droplet ids to closest matching blobs. /// /// Can currently only handle where both views contain /// the same number of 'droplets' fn match_with_blobs<B: Blob>(&self, blobs: &[B]) -> Option<Map<DropletId, B>> { // Ensure lengths are the same if self.droplets.len() != blobs.len() { error!("Expected and actual droplets are of different lengths"); return None; } let mut result = Map::new(); // to be returned let mut ids = vec![]; // store corresponding ids to indeces let mut matches = vec![]; // store similarity between blobs/droplets let n = blobs.len(); // store the id of each droplet in its corresponding // index in 'ids', then store the similarity of each // droplet to each blob in 'matches' for (&id, droplet) in &self.droplets { ids.push(id); for blob in blobs { let similarity = blob.get_similarity(&droplet); // must be non-negative for the algorithm to work assert!(similarity >= 0); matches.push(similarity); } } // convert the matches vector to a matrix // input should be [1,2,3,4], where the output // matrix is [[1,2],[3,4]] let m: Matrix<i32> = Matrix::from_vec(n, n, matches); // km is a vector of size n where the value at each index // corresponds to the index of a blob let (_c, km) = kuhn_munkres_min(&m); for i in 0..n { result.insert(ids[i], blobs[km[i]].clone()); } Some(result) } // this will take commands_to_finalize from the old snapshot into the new // one if an error is found produced pub fn correct(&mut self, blobs: &[impl Blob]) -> Option<Snapshot> { let blob_matching = self.match_with_blobs(blobs)?; let mut was_error = false; let new_droplets: Map<_, _> = blob_matching .iter() .map(|(&id, blob)| { let d = self.droplets.get_mut(&id).unwrap(); let d_new = blob.to_droplet(id); if d.location != d_new.location || d.dimensions != d_new.dimensions { info!("Found error in droplet {:?}", id); debug!("Droplet error\n Expected: {:#?}\n Found: {:#?}", d, d_new); was_error = true; } // HACK FIXME this mutation is not great if (d.volume - d_new.volume).abs() > 1.0 { info!( "volume of {} changed: {} -> {}", id.id, d.volume, d_new.volume ) } d.volume = d_new.volume; (id, d_new) }).collect(); if was_error { let mut new_snapshot = Snapshot { droplets: new_droplets, commands_to_finalize: Vec::new(), }; ::std::mem::swap( &mut new_snapshot.commands_to_finalize, &mut self.commands_to_finalize, ); Some(new_snapshot) } else { None } } pub fn diff_droplet(&self, id: &DropletId, other: &Snapshot) -> DropletDiff { use self::DropletDiff::*; let droplet = self .droplets .get(id) .expect("id should be in self snapshot"); if let Some(other_droplet) = other.droplets.get(id) { // NOTE we only care about location diffs for now let loc = droplet.location; let other_loc = other_droplet.location; if loc != other_loc { // for now, just assert that we are only moving one spot at a time // FIXME HACK // assert_eq!((&loc - &other_loc).norm(), 1); Moved { from: loc, to: other_loc, } } else { DidNotMove } } else { Disappeared } } pub fn get_error_edges( &self, planned_outcome: &Snapshot, actual_outcome: &Snapshot, ) -> Vec<(Location, Location)> { use self::DropletDiff::*; self.droplets .keys() .filter_map(|id| { let planned_diff = self.diff_droplet(id, planned_outcome); let actual_diff = self.diff_droplet(id, actual_outcome); match (planned_diff, actual_diff) { (Moved { from, to }, DidNotMove) => { if (&from - &to).norm() == 1 { Some((from, to)) } else { warn!("Droplet {} jumped from {} to {}!", id.id, from, to); None } } _ => None, } }).collect() } } #[derive(Debug)] pub enum ExecResponse { Step(Snapshot), NotReady, Done, } impl GridView { pub fn new(grid: Grid) -> GridView { let mut planned = VecDeque::new(); planned.push_back(Snapshot::default()); #[cfg(feature = "pi")] let pi = match ::std::env::var("PUDDLE_PI") { Ok(s) => if s == "1" { let mut pi = RaspberryPi::new().unwrap(); info!("Initialized the pi!"); Some(pi) } else { warn!("Couldn't read PUDDLE_PI={}", s); None }, Err(_) => { info!("Did not start the pi!"); None } };
DropletDiff
identifier_name
wix.rs
data structure is needed because WIX requires each path to have its own `id` and `guid`. #[derive(Serialize)] struct
{ /// the GUID to use on the WIX XML. guid: String, /// the id to use on the WIX XML. id: String, /// the binary path. path: String, } /// A Resource file to bundle with WIX. /// This data structure is needed because WIX requires each path to have its own `id` and `guid`. #[derive(Serialize, Clone)] struct ResourceFile { /// the GUID to use on the WIX XML. guid: String, /// the id to use on the WIX XML. id: String, /// the file path. path: String, } /// A resource directory to bundle with WIX. /// This data structure is needed because WIX requires each path to have its own `id` and `guid`. #[derive(Serialize)] struct ResourceDirectory { /// the directory name of the described resource. name: String, /// the files of the described resource directory. files: Vec<ResourceFile>, /// the directories that are children of the described resource directory. directories: Vec<ResourceDirectory>, } impl ResourceDirectory { /// Adds a file to this directory descriptor. fn add_file(&mut self, file: ResourceFile) { self.files.push(file); } /// Generates the wix XML string to bundle this directory resources recursively fn get_wix_data(self) -> crate::Result<(String, Vec<String>)> { let mut files = String::from(""); let mut file_ids = Vec::new(); for file in self.files { file_ids.push(file.id.clone()); files.push_str( format!( r#"<Component Id="{id}" Guid="{guid}" Win64="$(var.Win64)" KeyPath="yes"><File Id="PathFile_{id}" Source="{path}" /></Component>"#, id = file.id, guid = file.guid, path = file.path ).as_str() ); } let mut directories = String::from(""); for directory in self.directories { let (wix_string, ids) = directory.get_wix_data()?; for id in ids { file_ids.push(id) } directories.push_str(wix_string.as_str()); } let wix_string = format!( r#"<Directory Id="{name}" Name="{name}">{contents}</Directory>"#, name = self.name, contents = format!("{}{}", files, directories) ); Ok((wix_string, file_ids)) } } /// Copies the icons to the binary path, under the `resources` folder, /// and returns the path to that directory. fn copy_icons(settings: &Settings) -> crate::Result<PathBuf> { let base_dir = settings.project_out_directory(); let resource_dir = base_dir.join("resources"); let mut image_path = PathBuf::from(settings.project_out_directory()); // pop off till in tauri_src dir image_path.pop(); image_path.pop(); // get icon dir and icon file. let image_path = image_path.join("icons"); let opts = super::path_utils::Options::default(); copy( image_path, &resource_dir, &Options { copy_files: true, overwrite: true, ..opts }, )?; Ok(resource_dir) } /// Function used to download Wix and VC_REDIST. Checks SHA256 to verify the download. fn download_and_verify(url: &str, hash: &str) -> crate::Result<Vec<u8>> { common::print_info(format!("Downloading {}", url).as_str())?; let response = attohttpc::get(url).send()?; let data: Vec<u8> = response.bytes()?; common::print_info("validating hash")?; let mut hasher = sha2::Sha256::new(); hasher.update(&data); let url_hash = hasher.finalize().to_vec(); let expected_hash = hex::decode(hash)?; if expected_hash == url_hash { Ok(data) } else { Err(crate::Error::HashError) } } /// The installer directory of the app. fn app_installer_dir(settings: &Settings) -> crate::Result<PathBuf> { let arch = match settings.binary_arch() { "x86" => "x86", "x86_64" => "x64", target => { return Err(crate::Error::ArchError(format!( "Unsupported architecture: {}", target ))) } }; let package_base_name = format!( "{}_{}_{}", settings.main_binary_name().replace(".exe", ""), settings.version_string(), arch ); Ok( settings .project_out_directory() .to_path_buf() .join(format!("bundle/msi/{}.msi", package_base_name)), ) } /// Extracts the zips from Wix and VC_REDIST into a useable path. fn extract_zip(data: &Vec<u8>, path: &Path) -> crate::Result<()> { let cursor = Cursor::new(data); let mut zipa = ZipArchive::new(cursor)?; for i in 0..zipa.len() { let mut file = zipa.by_index(i)?; let dest_path = path.join(file.name()); let parent = dest_path.parent().expect("Failed to get parent"); if !parent.exists() { create_dir_all(parent)?; } let mut buff: Vec<u8> = Vec::new(); file.read_to_end(&mut buff)?; let mut fileout = File::create(dest_path).expect("Failed to open file"); fileout.write_all(&buff)?; } Ok(()) } /// Generates the UUID for the Wix template. fn generate_package_guid(settings: &Settings) -> Uuid { generate_guid(settings.bundle_identifier().as_bytes()) } /// Generates a GUID. fn generate_guid(key: &[u8]) -> Uuid { let namespace = Uuid::from_bytes(UUID_NAMESPACE); Uuid::new_v5(&namespace, key) } // Specifically goes and gets Wix and verifies the download via Sha256 pub fn get_and_extract_wix(path: &Path) -> crate::Result<()> { common::print_info("Verifying wix package")?; let data = download_and_verify(WIX_URL, WIX_SHA256)?; common::print_info("extracting WIX")?; extract_zip(&data, path) } // For if bundler needs DLL files. // fn run_heat_exe( // wix_toolset_path: &Path, // build_path: &Path, // harvest_dir: &Path, // platform: &str, // ) -> Result<(), String> { // let mut args = vec!["dir"]; // let harvest_str = harvest_dir.display().to_string(); // args.push(&harvest_str); // args.push("-platform"); // args.push(platform); // args.push("-cg"); // args.push("AppFiles"); // args.push("-dr"); // args.push("APPLICATIONFOLDER"); // args.push("-gg"); // args.push("-srd"); // args.push("-out"); // args.push("appdir.wxs"); // args.push("-var"); // args.push("var.SourceDir"); // let heat_exe = wix_toolset_path.join("heat.exe"); // let mut cmd = Command::new(&heat_exe) // .args(&args) // .stdout(Stdio::piped()) // .current_dir(build_path) // .spawn() // .expect("error running heat.exe"); // { // let stdout = cmd.stdout.as_mut().unwrap(); // let reader = BufReader::new(stdout); // for line in reader.lines() { // info!(logger, "{}", line.unwrap()); // } // } // let status = cmd.wait().unwrap(); // if status.success() { // Ok(()) // } else { // Err("error running heat.exe".to_string()) // } // } /// Runs the Candle.exe executable for Wix. Candle parses the wxs file and generates the code for building the installer. fn run_candle( settings: &Settings, wix_toolset_path: &Path, build_path: &Path, wxs_file_name: &str, ) -> crate::Result<()> { let arch = match settings.binary_arch() { "x86_64" => "x64", "x86" => "x86", target => { return Err(crate::Error::ArchError(format!( "unsupported target: {}", target ))) } }; let main_binary = settings .binaries() .iter() .find(|bin| bin.main()) .ok_or_else(|| anyhow::anyhow!("Failed to get main binary"))?; let args = vec![ "-arch".to_string(), arch.to_string(), wxs_file_name.to_string(), format!( "-dSourceDir={}", settings.binary_path(main_binary).display() ), ]; let candle_exe = wix_toolset_path.join("candle.exe"); common::print_info(format!("running candle for {}", wxs_file_name).as_str())?; let mut cmd = Command::new(&candle_exe); cmd
Binary
identifier_name
wix.rs
data structure is needed because WIX requires each path to have its own `id` and `guid`. #[derive(Serialize)] struct Binary { /// the GUID to use on the WIX XML. guid: String, /// the id to use on the WIX XML. id: String, /// the binary path. path: String, } /// A Resource file to bundle with WIX. /// This data structure is needed because WIX requires each path to have its own `id` and `guid`. #[derive(Serialize, Clone)] struct ResourceFile { /// the GUID to use on the WIX XML. guid: String, /// the id to use on the WIX XML. id: String, /// the file path. path: String, } /// A resource directory to bundle with WIX. /// This data structure is needed because WIX requires each path to have its own `id` and `guid`. #[derive(Serialize)] struct ResourceDirectory { /// the directory name of the described resource. name: String, /// the files of the described resource directory. files: Vec<ResourceFile>, /// the directories that are children of the described resource directory. directories: Vec<ResourceDirectory>, } impl ResourceDirectory { /// Adds a file to this directory descriptor. fn add_file(&mut self, file: ResourceFile) { self.files.push(file); } /// Generates the wix XML string to bundle this directory resources recursively fn get_wix_data(self) -> crate::Result<(String, Vec<String>)> { let mut files = String::from(""); let mut file_ids = Vec::new(); for file in self.files { file_ids.push(file.id.clone()); files.push_str( format!( r#"<Component Id="{id}" Guid="{guid}" Win64="$(var.Win64)" KeyPath="yes"><File Id="PathFile_{id}" Source="{path}" /></Component>"#, id = file.id, guid = file.guid, path = file.path ).as_str() ); } let mut directories = String::from(""); for directory in self.directories { let (wix_string, ids) = directory.get_wix_data()?; for id in ids { file_ids.push(id) } directories.push_str(wix_string.as_str()); } let wix_string = format!( r#"<Directory Id="{name}" Name="{name}">{contents}</Directory>"#, name = self.name, contents = format!("{}{}", files, directories) ); Ok((wix_string, file_ids)) } } /// Copies the icons to the binary path, under the `resources` folder, /// and returns the path to that directory. fn copy_icons(settings: &Settings) -> crate::Result<PathBuf> { let base_dir = settings.project_out_directory(); let resource_dir = base_dir.join("resources"); let mut image_path = PathBuf::from(settings.project_out_directory()); // pop off till in tauri_src dir image_path.pop(); image_path.pop(); // get icon dir and icon file. let image_path = image_path.join("icons"); let opts = super::path_utils::Options::default(); copy( image_path, &resource_dir, &Options { copy_files: true, overwrite: true, ..opts }, )?; Ok(resource_dir) } /// Function used to download Wix and VC_REDIST. Checks SHA256 to verify the download. fn download_and_verify(url: &str, hash: &str) -> crate::Result<Vec<u8>> { common::print_info(format!("Downloading {}", url).as_str())?; let response = attohttpc::get(url).send()?; let data: Vec<u8> = response.bytes()?; common::print_info("validating hash")?; let mut hasher = sha2::Sha256::new(); hasher.update(&data); let url_hash = hasher.finalize().to_vec(); let expected_hash = hex::decode(hash)?; if expected_hash == url_hash { Ok(data) } else { Err(crate::Error::HashError) } } /// The installer directory of the app. fn app_installer_dir(settings: &Settings) -> crate::Result<PathBuf> { let arch = match settings.binary_arch() { "x86" => "x86", "x86_64" => "x64", target => { return Err(crate::Error::ArchError(format!( "Unsupported architecture: {}", target ))) } }; let package_base_name = format!( "{}_{}_{}", settings.main_binary_name().replace(".exe", ""), settings.version_string(), arch ); Ok( settings .project_out_directory() .to_path_buf() .join(format!("bundle/msi/{}.msi", package_base_name)), ) } /// Extracts the zips from Wix and VC_REDIST into a useable path. fn extract_zip(data: &Vec<u8>, path: &Path) -> crate::Result<()> { let cursor = Cursor::new(data); let mut zipa = ZipArchive::new(cursor)?; for i in 0..zipa.len() { let mut file = zipa.by_index(i)?; let dest_path = path.join(file.name()); let parent = dest_path.parent().expect("Failed to get parent"); if !parent.exists() { create_dir_all(parent)?; } let mut buff: Vec<u8> = Vec::new(); file.read_to_end(&mut buff)?; let mut fileout = File::create(dest_path).expect("Failed to open file"); fileout.write_all(&buff)?; } Ok(()) } /// Generates the UUID for the Wix template. fn generate_package_guid(settings: &Settings) -> Uuid { generate_guid(settings.bundle_identifier().as_bytes()) } /// Generates a GUID. fn generate_guid(key: &[u8]) -> Uuid { let namespace = Uuid::from_bytes(UUID_NAMESPACE); Uuid::new_v5(&namespace, key) } // Specifically goes and gets Wix and verifies the download via Sha256 pub fn get_and_extract_wix(path: &Path) -> crate::Result<()> { common::print_info("Verifying wix package")?; let data = download_and_verify(WIX_URL, WIX_SHA256)?; common::print_info("extracting WIX")?; extract_zip(&data, path) } // For if bundler needs DLL files. // fn run_heat_exe( // wix_toolset_path: &Path, // build_path: &Path, // harvest_dir: &Path, // platform: &str, // ) -> Result<(), String> { // let mut args = vec!["dir"]; // let harvest_str = harvest_dir.display().to_string(); // args.push(&harvest_str); // args.push("-platform"); // args.push(platform); // args.push("-cg"); // args.push("AppFiles"); // args.push("-dr"); // args.push("APPLICATIONFOLDER"); // args.push("-gg"); // args.push("-srd"); // args.push("-out"); // args.push("appdir.wxs"); // args.push("-var"); // args.push("var.SourceDir"); // let heat_exe = wix_toolset_path.join("heat.exe"); // let mut cmd = Command::new(&heat_exe) // .args(&args) // .stdout(Stdio::piped()) // .current_dir(build_path) // .spawn() // .expect("error running heat.exe"); // { // let stdout = cmd.stdout.as_mut().unwrap(); // let reader = BufReader::new(stdout); // for line in reader.lines() { // info!(logger, "{}", line.unwrap()); // } // } // let status = cmd.wait().unwrap(); // if status.success() { // Ok(()) // } else { // Err("error running heat.exe".to_string()) // } // } /// Runs the Candle.exe executable for Wix. Candle parses the wxs file and generates the code for building the installer. fn run_candle( settings: &Settings, wix_toolset_path: &Path, build_path: &Path, wxs_file_name: &str, ) -> crate::Result<()>
arch.to_string(), wxs_file_name.to_string(), format!( "-dSourceDir={}", settings.binary_path(main_binary).display() ), ]; let candle_exe = wix_toolset_path.join("candle.exe"); common::print_info(format!("running candle for {}", wxs_file_name).as_str())?; let mut cmd = Command::new(&candle_exe); cmd
{ let arch = match settings.binary_arch() { "x86_64" => "x64", "x86" => "x86", target => { return Err(crate::Error::ArchError(format!( "unsupported target: {}", target ))) } }; let main_binary = settings .binaries() .iter() .find(|bin| bin.main()) .ok_or_else(|| anyhow::anyhow!("Failed to get main binary"))?; let args = vec![ "-arch".to_string(),
identifier_body
wix.rs
_else(|| anyhow::anyhow!("Failed to get main binary"))?; let args = vec![ "-arch".to_string(), arch.to_string(), wxs_file_name.to_string(), format!( "-dSourceDir={}", settings.binary_path(main_binary).display() ), ]; let candle_exe = wix_toolset_path.join("candle.exe"); common::print_info(format!("running candle for {}", wxs_file_name).as_str())?; let mut cmd = Command::new(&candle_exe); cmd .args(&args) .stdout(Stdio::piped()) .current_dir(build_path); common::print_info("running candle.exe")?; common::execute_with_output(&mut cmd).map_err(|_| crate::Error::CandleError) } /// Runs the Light.exe file. Light takes the generated code from Candle and produces an MSI Installer. fn run_light( wix_toolset_path: &Path, build_path: &Path, wixobjs: &[&str], output_path: &Path, ) -> crate::Result<PathBuf> { let light_exe = wix_toolset_path.join("light.exe"); let mut args: Vec<String> = vec![ "-ext".to_string(), "WixUIExtension".to_string(), "-o".to_string(), output_path.display().to_string(), ]; for p in wixobjs { args.push(p.to_string()); } let mut cmd = Command::new(&light_exe); cmd .args(&args) .stdout(Stdio::piped()) .current_dir(build_path); common::print_info(format!("running light to produce {}", output_path.display()).as_str())?; common::execute_with_output(&mut cmd) .map(|_| output_path.to_path_buf()) .map_err(|_| crate::Error::LightError) } // fn get_icon_data() -> crate::Result<()> { // Ok(()) // } // Entry point for bundling and creating the MSI installer. For now the only supported platform is Windows x64. pub fn build_wix_app_installer( settings: &Settings, wix_toolset_path: &Path, ) -> crate::Result<PathBuf> { let arch = match settings.binary_arch() { "x86_64" => "x64", "x86" => "x86", target => { return Err(crate::Error::ArchError(format!( "unsupported target: {}", target ))) } }; // target only supports x64. common::print_info(format!("Target: {}", arch).as_str())?; let output_path = settings .project_out_directory() .join("bundle/msi") .join(arch); let mut data = BTreeMap::new(); if let Ok(tauri_config) = crate::bundle::tauri_config::get() { data.insert( "embedded_server", to_json(tauri_config.tauri.embedded_server.active), ); } data.insert("product_name", to_json(settings.bundle_name())); data.insert("version", to_json(settings.version_string())); let manufacturer = settings.bundle_identifier().to_string(); data.insert("manufacturer", to_json(manufacturer.as_str())); let upgrade_code = Uuid::new_v5( &Uuid::NAMESPACE_DNS, format!("{}.app.x64", &settings.main_binary_name()).as_bytes(), ) .to_string(); data.insert("upgrade_code", to_json(&upgrade_code.as_str())); let path_guid = generate_package_guid(settings).to_string(); data.insert("path_component_guid", to_json(&path_guid.as_str())); let shortcut_guid = generate_package_guid(settings).to_string(); data.insert("shortcut_guid", to_json(&shortcut_guid.as_str())); let app_exe_name = settings.main_binary_name().to_string(); data.insert("app_exe_name", to_json(&app_exe_name)); let binaries = generate_binaries_data(&settings)?; let binaries_json = to_json(&binaries); data.insert("binaries", binaries_json); let resources = generate_resource_data(&settings)?; let mut resources_wix_string = String::from(""); let mut files_ids = Vec::new(); for (_, dir) in resources { let (wix_string, ids) = dir.get_wix_data()?; resources_wix_string.push_str(wix_string.as_str()); for id in ids { files_ids.push(id); } } data.insert("resources", to_json(resources_wix_string)); data.insert("resource_file_ids", to_json(files_ids)); let main_binary = settings .binaries() .iter() .find(|bin| bin.main()) .ok_or_else(|| anyhow::anyhow!("Failed to get main binary"))?; let app_exe_source = settings.binary_path(main_binary).display().to_string(); data.insert("app_exe_source", to_json(&app_exe_source)); // copy icons from icons folder to resource folder near msi let image_path = copy_icons(&settings)?; let path = image_path.join("icon.ico").display().to_string(); data.insert("icon_path", to_json(path.as_str())); let temp = HANDLEBARS.render("main.wxs", &data)?; if output_path.exists() { remove_dir_all(&output_path).or_else(|e| Err(e))?; } create_dir_all(&output_path).or_else(|e| Err(e))?; let main_wxs_path = output_path.join("main.wxs"); write(&main_wxs_path, temp).or_else(|e| Err(e))?; let input_basenames = vec!["main"]; for basename in &input_basenames { let wxs = format!("{}.wxs", basename); run_candle(settings, &wix_toolset_path, &output_path, &wxs)?; } let wixobjs = vec!["main.wixobj"]; let target = run_light( &wix_toolset_path, &output_path, &wixobjs, &app_installer_dir(settings)?, )?; Ok(target) } /// Generates the data required for the external binaries and extra binaries bundling. fn generate_binaries_data(settings: &Settings) -> crate::Result<Vec<Binary>> { let mut binaries = Vec::new(); let regex = Regex::new(r"[^\w\d\.]")?; let cwd = std::env::current_dir()?; for src in settings.external_binaries() { let src = src?; let filename = src .file_name() .expect("failed to extract external binary filename") .to_os_string() .into_string() .expect("failed to convert external binary filename to string"); let guid = generate_guid(filename.as_bytes()).to_string(); binaries.push(Binary { guid, path: cwd .join(src) .into_os_string() .into_string() .expect("failed to read external binary path"), id: regex.replace_all(&filename, "").to_string(), }); } for bin in settings.binaries() { let filename = bin.name(); let guid = generate_guid(filename.as_bytes()).to_string(); if !bin.main() { binaries.push(Binary { guid, path: settings .binary_path(bin) .into_os_string() .into_string() .expect("failed to read binary path"), id: regex.replace_all(&filename, "").to_string(), }) } } Ok(binaries) } /// Generates the data required for the resource bundling on wix fn generate_resource_data(settings: &Settings) -> crate::Result<ResourceMap> { let mut resources = ResourceMap::new(); let regex = Regex::new(r"[^\w\d\.]")?; let cwd = std::env::current_dir()?; for src in settings.resource_files() { let src = src?; let filename = src .file_name() .expect("failed to extract resource filename") .to_os_string() .into_string() .expect("failed to convert resource filename to string"); let resource_path = cwd .join(src.clone()) .into_os_string() .into_string() .expect("failed to read resource path"); let resource_entry = ResourceFile { guid: generate_guid(filename.as_bytes()).to_string(), path: resource_path, id: regex.replace_all(&filename, "").to_string(), }; // split the resource path directories let mut directories = src .components() .filter(|component| { let comp = component.as_os_str(); comp != "." && comp != ".." }) .collect::<Vec<_>>(); directories.truncate(directories.len() - 1); // transform the directory structure to a chained vec structure for directory in directories { let directory_name = directory .as_os_str() .to_os_string() .into_string() .expect("failed to read resource folder name"); // if the directory is already on the map if resources.contains_key(&directory_name) { let directory_entry = &mut resources .get_mut(&directory_name) .expect("Unable to handle resources"); if directory_entry.name == directory_name { // the directory entry is the root of the chain directory_entry.add_file(resource_entry.clone()); } else { let index = directory_entry
random_line_split
collision.go
Collisions(cg *quad.CollisionGroup, now stime.Time) ([]entity.Entity, []entity.Entity) { // Reset the resolved slice phase.solved = phase.solved[:0] // Generate a collision index for the collision group phase.collisionIndex = cg.CollisionIndex() // A map to store entities that still remain in the world remaining := make(map[entity.Id]entity.Entity, len(cg.Entities)) remainingSlice := func() []entity.Entity { // Build a slice from the `remaining` map s := make([]entity.Entity, 0, len(remaining)) for _, e := range remaining { s = append(s, e) } return s } for _, c := range cg.Collisions { if phase.hasSolved(c) { continue } var entities []entity.Entity // Resolve type of entity in collision.A switch e := c.A.(type) { case actorEntity: // Resolve the type of entity in collision.B entities = phase.resolveActorEntity(phase.actorIndex[e.ActorId()], c.B, c, now) default: switch e := c.B.(type) { case actorEntity: // Resolve the type of entity in collision.B entities = phase.resolveActorEntity(phase.actorIndex[e.ActorId()], c.A, c, now) } } // As collisions are solved they return entities // that have been created or modified and we store // them in a map by their Id. Multiple collisions // may modify and entity, therefor we only will // one version of the entity back to engine when // we return. for _, e := range entities { remaining[e.Id()] = e } } return remainingSlice(), nil } func (phase *narrowPhase) resolveActorEntity(a *actor, with entity.Entity, collision quad.Collision, now stime.Time) []entity.Entity { switch e := with.(type) { case actorEntity: b := phase.actorIndex[e.ActorId()] return phase.solveActorActor(&solverActorActor{}, a, b, collision) case assailEntity: return phase.solveActorAssail(a, e, collision, now) case wallEntity: a.revertMoveAction() return []entity.Entity{a.Entity(), e} } return nil }
} var percentDamage float64 switch a.pathAction { case nil: if a.Cell() == assail.Cell() { percentDamage = 1.0 } default: coordCollision := coord.NewCellCollision(*a.pathAction, assail.Cell()) percentDamage = coordCollision.OverlapAt(now) } damage := int(math.Floor(float64(assail.damage) * percentDamage)) a.hp -= damage if a.hp <= 0 { a.hp = 100 a.actorEntity.cell = origin a.actorEntity.facing = coord.South a.actorEntity.pathAction = nil } return []entity.Entity{a.Entity()} } func newActorActorCollision(a, b *actor) (*actor, *actor, coord.Collision) { var collision coord.Collision switch { case a.pathAction == nil && b.pathAction != nil: a, b = b, a fallthrough case a.pathAction != nil && b.pathAction == nil: collision = coord.NewCellCollision(*a.pathAction, b.Cell()) // A or B may have had a previous collision resolved that // caused this collision to not be possible anymore. // It is more relevant to return nil here then a // coord.Collision with type CT_NONE if collision.Type() == coord.CT_NONE { return a, b, nil } case a.pathAction != nil && b.pathAction != nil: pathCollision := coord.NewPathCollision(*a.pathAction, *b.pathAction) // coord.NewPathCollision can flip the, // A and B paths to simplify the number // of collision types. This normalizes // actor A with pathCollision.A if *a.pathAction != pathCollision.A { a, b = b, a } collision = pathCollision case a.pathAction == nil && b.pathAction == nil: // This case handles actors being on the same square, // but not moving at all. // There isn't a coord.CollisionType for this case. // Maybe there should be? return a, b, nil default: panic(fmt.Sprintf("impossible collision between {%v} {%v}", a, b)) } return a, b, collision } type node struct { actor *actor entity entity.Entity } // Move forward in the directed graph. This movement is based on // which entity is occupying the destination of the other's path action. func followGraph(a, b *actor, collision quad.Collision) node { // normalize a, b to collision.[A, B] if a.actorEntity.Id() != collision.A.Id() { a, b = b, a } var actor *actor var entity entity.Entity switch { case a.pathAction.Orig == b.pathAction.Dest: entity = collision.A actor = a case b.pathAction.Orig == a.pathAction.Dest: entity = collision.B actor = b default: panic(fmt.Sprintf("unexpected graph state %v between %v & %v", collision, a, b)) } return node{actor, entity} } // Used to figure out which actor is "A" if // the collision was CT_A_INTO_B instead of CT_NONE func currentNode(a, b *actor, collision quad.Collision) *actor { switch { case a.pathAction.Orig == b.pathAction.Dest: return b case b.pathAction.Orig == a.pathAction.Dest: return a default: panic(fmt.Sprintf("unexpected graph state %v between %v & %v", collision, a, b)) } } // Compare entity Id's with the entities in // a collision and return the one that isn't // the actor. func otherEntityIn(a *actor, collision quad.Collision) entity.Entity { var e entity.Entity // figure out is prioritized actor is A or B in the collision switch { case a.actorEntity.Id() != collision.A.Id(): e = collision.A case a.actorEntity.Id() != collision.B.Id(): e = collision.B default: panic(fmt.Sprintf("unexpected graph state %v actor %v", collision, a)) } return e } // Store what actor's have been visited during // a recursive solve. Used to avoid infinite // recursion through a cycle in the graph. type solverActorActor struct { visited []*actor } func (s solverActorActor) hasVisited(actor *actor) bool { for _, a := range s.visited { if actor == a { return true } } return false } func (phase *narrowPhase) solveActorActor(solver *solverActorActor, a, b *actor, collision quad.Collision) []entity.Entity { // When this functions returns the // collision will have been solved defer func() { phase.solved = append(phase.solved, collision) }() var entities []entity.Entity attemptSolve: a, b, coordCollision := newActorActorCollision(a, b) if coordCollision == nil { goto resolved } switch coordCollision.Type() { case coord.CT_NONE: // This may not be entirely accurate. // We should walk through the collision index // of our partner too see if they should resolve // some of there collisions first. They may // appear to be moving to us right now, but // have a collision that when solved will // render them motionless, thus we would become // motionless as well. e, err := phase.solveDependencies(solver, a, b, collision) switch err { case nil: if len(e) > 0 { entities = append(entities, e...) } // Try solving again goto attemptSolve case errCycleDetected: // Detected a cycle, we can't move currentNode(a, b, collision).revertMoveAction() goto resolved case errNoDependencies: // All dependencies have been solved // We can move goto resolved } case coord.CT_CELL_DEST: a.revertMoveAction() goto resolved case coord.CT_SWAP: a.revertMoveAction() b.revertMoveAction() goto resolved case coord.CT_A_INTO_B_FROM_SIDE: // This may not be entirely accurate. // We should walk through the collision index // of our partner too see if they should resolve // some of there collisions first. They may // appear to be moving to us right
func (phase *narrowPhase) solveActorAssail(a *actor, assail assailEntity, collision quad.Collision, now stime.Time) []entity.Entity { // Don't damage yourself if assail.spawnedBy == a.actorEntity.Id() { return []entity.Entity{a.Entity()}
random_line_split
collision.go
return false } func (phase narrowPhaseLocker) ResolveCollisions(cg *quad.CollisionGroup, now stime.Time) ([]entity.Entity, []entity.Entity) { defer phase.ActorIndexLocker.RUnlock() return newNarrowPhase(phase.ActorIndexLocker.RLock()).ResolveCollisions(cg, now) } // Implementation of the quad.NarrowPhaseHandler interface. func (phase narrowPhase) ResolveCollisions(cg *quad.CollisionGroup, now stime.Time) ([]entity.Entity, []entity.Entity) { // Reset the resolved slice phase.solved = phase.solved[:0] // Generate a collision index for the collision group phase.collisionIndex = cg.CollisionIndex() // A map to store entities that still remain in the world remaining := make(map[entity.Id]entity.Entity, len(cg.Entities)) remainingSlice := func() []entity.Entity { // Build a slice from the `remaining` map s := make([]entity.Entity, 0, len(remaining)) for _, e := range remaining { s = append(s, e) } return s } for _, c := range cg.Collisions { if phase.hasSolved(c) { continue } var entities []entity.Entity // Resolve type of entity in collision.A switch e := c.A.(type) { case actorEntity: // Resolve the type of entity in collision.B entities = phase.resolveActorEntity(phase.actorIndex[e.ActorId()], c.B, c, now) default: switch e := c.B.(type) { case actorEntity: // Resolve the type of entity in collision.B entities = phase.resolveActorEntity(phase.actorIndex[e.ActorId()], c.A, c, now) } } // As collisions are solved they return entities // that have been created or modified and we store // them in a map by their Id. Multiple collisions // may modify and entity, therefor we only will // one version of the entity back to engine when // we return. for _, e := range entities { remaining[e.Id()] = e } } return remainingSlice(), nil } func (phase *narrowPhase) resolveActorEntity(a *actor, with entity.Entity, collision quad.Collision, now stime.Time) []entity.Entity { switch e := with.(type) { case actorEntity: b := phase.actorIndex[e.ActorId()] return phase.solveActorActor(&solverActorActor{}, a, b, collision) case assailEntity: return phase.solveActorAssail(a, e, collision, now) case wallEntity: a.revertMoveAction() return []entity.Entity{a.Entity(), e} } return nil } func (phase *narrowPhase) solveActorAssail(a *actor, assail assailEntity, collision quad.Collision, now stime.Time) []entity.Entity { // Don't damage yourself if assail.spawnedBy == a.actorEntity.Id() { return []entity.Entity{a.Entity()} } var percentDamage float64 switch a.pathAction { case nil: if a.Cell() == assail.Cell() { percentDamage = 1.0 } default: coordCollision := coord.NewCellCollision(*a.pathAction, assail.Cell()) percentDamage = coordCollision.OverlapAt(now) } damage := int(math.Floor(float64(assail.damage) * percentDamage)) a.hp -= damage if a.hp <= 0 { a.hp = 100 a.actorEntity.cell = origin a.actorEntity.facing = coord.South a.actorEntity.pathAction = nil } return []entity.Entity{a.Entity()} } func newActorActorCollision(a, b *actor) (*actor, *actor, coord.Collision) { var collision coord.Collision switch { case a.pathAction == nil && b.pathAction != nil: a, b = b, a fallthrough case a.pathAction != nil && b.pathAction == nil: collision = coord.NewCellCollision(*a.pathAction, b.Cell()) // A or B may have had a previous collision resolved that // caused this collision to not be possible anymore. // It is more relevant to return nil here then a // coord.Collision with type CT_NONE if collision.Type() == coord.CT_NONE { return a, b, nil } case a.pathAction != nil && b.pathAction != nil: pathCollision := coord.NewPathCollision(*a.pathAction, *b.pathAction) // coord.NewPathCollision can flip the, // A and B paths to simplify the number // of collision types. This normalizes // actor A with pathCollision.A if *a.pathAction != pathCollision.A { a, b = b, a } collision = pathCollision case a.pathAction == nil && b.pathAction == nil: // This case handles actors being on the same square, // but not moving at all. // There isn't a coord.CollisionType for this case. // Maybe there should be? return a, b, nil default: panic(fmt.Sprintf("impossible collision between {%v} {%v}", a, b)) } return a, b, collision } type node struct { actor *actor entity entity.Entity } // Move forward in the directed graph. This movement is based on // which entity is occupying the destination of the other's path action. func followGraph(a, b *actor, collision quad.Collision) node { // normalize a, b to collision.[A, B] if a.actorEntity.Id() != collision.A.Id() { a, b = b, a } var actor *actor var entity entity.Entity switch { case a.pathAction.Orig == b.pathAction.Dest: entity = collision.A actor = a case b.pathAction.Orig == a.pathAction.Dest: entity = collision.B actor = b default: panic(fmt.Sprintf("unexpected graph state %v between %v & %v", collision, a, b)) } return node{actor, entity} } // Used to figure out which actor is "A" if // the collision was CT_A_INTO_B instead of CT_NONE func currentNode(a, b *actor, collision quad.Collision) *actor { switch { case a.pathAction.Orig == b.pathAction.Dest: return b case b.pathAction.Orig == a.pathAction.Dest: return a default: panic(fmt.Sprintf("unexpected graph state %v between %v & %v", collision, a, b)) } } // Compare entity Id's with the entities in // a collision and return the one that isn't // the actor. func otherEntityIn(a *actor, collision quad.Collision) entity.Entity { var e entity.Entity // figure out is prioritized actor is A or B in the collision switch { case a.actorEntity.Id() != collision.A.Id(): e = collision.A case a.actorEntity.Id() != collision.B.Id(): e = collision.B default: panic(fmt.Sprintf("unexpected graph state %v actor %v", collision, a)) } return e } // Store what actor's have been visited during // a recursive solve. Used to avoid infinite // recursion through a cycle in the graph. type solverActorActor struct { visited []*actor } func (s solverActorActor) hasVisited(actor *actor) bool { for _, a := range s.visited { if actor == a { return true } } return false } func (phase *narrowPhase) solveActorActor(solver *solverActorActor, a, b *actor, collision quad.Collision) []entity.Entity { // When this functions returns the // collision will have been solved defer func() { phase.solved = append(phase.solved, collision) }() var entities []entity.Entity attemptSolve: a, b, coordCollision := newActorActorCollision(a, b) if coordCollision == nil { goto resolved } switch coordCollision.Type() { case coord.CT_NONE: // This may not be entirely accurate. // We should walk through the collision index // of our partner too see if they should resolve // some of there collisions first. They may // appear to be moving to us right now, but // have a collision that when solved will // render them motionless, thus we would become // motionless as well. e, err := phase.solveDependencies(solver, a, b, collision) switch err { case nil: if len(e) > 0 { entities = append(entities, e...) } // Try solving again goto attemptSolve case errCycleDetected: // Detected a cycle, we can't move currentNode(a, b, collision).revertMoveAction() goto resolved case errNoDependencies: // All dependencies have been solved // We can move goto resolved } case coord.CT_CELL_DEST
{ if c.IsSameAs(solved) { return true } }
conditional_block
collision.go
isions(cg *quad.CollisionGroup, now stime.Time) ([]entity.Entity, []entity.Entity) { // Reset the resolved slice phase.solved = phase.solved[:0] // Generate a collision index for the collision group phase.collisionIndex = cg.CollisionIndex() // A map to store entities that still remain in the world remaining := make(map[entity.Id]entity.Entity, len(cg.Entities)) remainingSlice := func() []entity.Entity { // Build a slice from the `remaining` map s := make([]entity.Entity, 0, len(remaining)) for _, e := range remaining { s = append(s, e) } return s } for _, c := range cg.Collisions { if phase.hasSolved(c) { continue } var entities []entity.Entity // Resolve type of entity in collision.A switch e := c.A.(type) { case actorEntity: // Resolve the type of entity in collision.B entities = phase.resolveActorEntity(phase.actorIndex[e.ActorId()], c.B, c, now) default: switch e := c.B.(type) { case actorEntity: // Resolve the type of entity in collision.B entities = phase.resolveActorEntity(phase.actorIndex[e.ActorId()], c.A, c, now) } } // As collisions are solved they return entities // that have been created or modified and we store // them in a map by their Id. Multiple collisions // may modify and entity, therefor we only will // one version of the entity back to engine when // we return. for _, e := range entities { remaining[e.Id()] = e } } return remainingSlice(), nil } func (phase *narrowPhase) resolveActorEntity(a *actor, with entity.Entity, collision quad.Collision, now stime.Time) []entity.Entity { switch e := with.(type) { case actorEntity: b := phase.actorIndex[e.ActorId()] return phase.solveActorActor(&solverActorActor{}, a, b, collision) case assailEntity: return phase.solveActorAssail(a, e, collision, now) case wallEntity: a.revertMoveAction() return []entity.Entity{a.Entity(), e} } return nil } func (phase *narrowPhase) solveActorAssail(a *actor, assail assailEntity, collision quad.Collision, now stime.Time) []entity.Entity { // Don't damage yourself if assail.spawnedBy == a.actorEntity.Id() { return []entity.Entity{a.Entity()} } var percentDamage float64 switch a.pathAction { case nil: if a.Cell() == assail.Cell() { percentDamage = 1.0 } default: coordCollision := coord.NewCellCollision(*a.pathAction, assail.Cell()) percentDamage = coordCollision.OverlapAt(now) } damage := int(math.Floor(float64(assail.damage) * percentDamage)) a.hp -= damage if a.hp <= 0 { a.hp = 100 a.actorEntity.cell = origin a.actorEntity.facing = coord.South a.actorEntity.pathAction = nil } return []entity.Entity{a.Entity()} } func newActorActorCollision(a, b *actor) (*actor, *actor, coord.Collision) { var collision coord.Collision switch { case a.pathAction == nil && b.pathAction != nil: a, b = b, a fallthrough case a.pathAction != nil && b.pathAction == nil: collision = coord.NewCellCollision(*a.pathAction, b.Cell()) // A or B may have had a previous collision resolved that // caused this collision to not be possible anymore. // It is more relevant to return nil here then a // coord.Collision with type CT_NONE if collision.Type() == coord.CT_NONE { return a, b, nil } case a.pathAction != nil && b.pathAction != nil: pathCollision := coord.NewPathCollision(*a.pathAction, *b.pathAction) // coord.NewPathCollision can flip the, // A and B paths to simplify the number // of collision types. This normalizes // actor A with pathCollision.A if *a.pathAction != pathCollision.A { a, b = b, a } collision = pathCollision case a.pathAction == nil && b.pathAction == nil: // This case handles actors being on the same square, // but not moving at all. // There isn't a coord.CollisionType for this case. // Maybe there should be? return a, b, nil default: panic(fmt.Sprintf("impossible collision between {%v} {%v}", a, b)) } return a, b, collision } type node struct { actor *actor entity entity.Entity } // Move forward in the directed graph. This movement is based on // which entity is occupying the destination of the other's path action. func followGraph(a, b *actor, collision quad.Collision) node { // normalize a, b to collision.[A, B] if a.actorEntity.Id() != collision.A.Id() { a, b = b, a } var actor *actor var entity entity.Entity switch { case a.pathAction.Orig == b.pathAction.Dest: entity = collision.A actor = a case b.pathAction.Orig == a.pathAction.Dest: entity = collision.B actor = b default: panic(fmt.Sprintf("unexpected graph state %v between %v & %v", collision, a, b)) } return node{actor, entity} } // Used to figure out which actor is "A" if // the collision was CT_A_INTO_B instead of CT_NONE func currentNode(a, b *actor, collision quad.Collision) *actor
// Compare entity Id's with the entities in // a collision and return the one that isn't // the actor. func otherEntityIn(a *actor, collision quad.Collision) entity.Entity { var e entity.Entity // figure out is prioritized actor is A or B in the collision switch { case a.actorEntity.Id() != collision.A.Id(): e = collision.A case a.actorEntity.Id() != collision.B.Id(): e = collision.B default: panic(fmt.Sprintf("unexpected graph state %v actor %v", collision, a)) } return e } // Store what actor's have been visited during // a recursive solve. Used to avoid infinite // recursion through a cycle in the graph. type solverActorActor struct { visited []*actor } func (s solverActorActor) hasVisited(actor *actor) bool { for _, a := range s.visited { if actor == a { return true } } return false } func (phase *narrowPhase) solveActorActor(solver *solverActorActor, a, b *actor, collision quad.Collision) []entity.Entity { // When this functions returns the // collision will have been solved defer func() { phase.solved = append(phase.solved, collision) }() var entities []entity.Entity attemptSolve: a, b, coordCollision := newActorActorCollision(a, b) if coordCollision == nil { goto resolved } switch coordCollision.Type() { case coord.CT_NONE: // This may not be entirely accurate. // We should walk through the collision index // of our partner too see if they should resolve // some of there collisions first. They may // appear to be moving to us right now, but // have a collision that when solved will // render them motionless, thus we would become // motionless as well. e, err := phase.solveDependencies(solver, a, b, collision) switch err { case nil: if len(e) > 0 { entities = append(entities, e...) } // Try solving again goto attemptSolve case errCycleDetected: // Detected a cycle, we can't move currentNode(a, b, collision).revertMoveAction() goto resolved case errNoDependencies: // All dependencies have been solved // We can move goto resolved } case coord.CT_CELL_DEST: a.revertMoveAction() goto resolved case coord.CT_SWAP: a.revertMoveAction() b.revertMoveAction() goto resolved case coord.CT_A_INTO_B_FROM_SIDE: // This may not be entirely accurate. // We should walk through the collision index // of our partner too see if they should resolve // some of there collisions first. They may // appear to be moving to us right
{ switch { case a.pathAction.Orig == b.pathAction.Dest: return b case b.pathAction.Orig == a.pathAction.Dest: return a default: panic(fmt.Sprintf("unexpected graph state %v between %v & %v", collision, a, b)) } }
identifier_body
collision.go
Collisions(cg *quad.CollisionGroup, now stime.Time) ([]entity.Entity, []entity.Entity) { // Reset the resolved slice phase.solved = phase.solved[:0] // Generate a collision index for the collision group phase.collisionIndex = cg.CollisionIndex() // A map to store entities that still remain in the world remaining := make(map[entity.Id]entity.Entity, len(cg.Entities)) remainingSlice := func() []entity.Entity { // Build a slice from the `remaining` map s := make([]entity.Entity, 0, len(remaining)) for _, e := range remaining { s = append(s, e) } return s } for _, c := range cg.Collisions { if phase.hasSolved(c) { continue } var entities []entity.Entity // Resolve type of entity in collision.A switch e := c.A.(type) { case actorEntity: // Resolve the type of entity in collision.B entities = phase.resolveActorEntity(phase.actorIndex[e.ActorId()], c.B, c, now) default: switch e := c.B.(type) { case actorEntity: // Resolve the type of entity in collision.B entities = phase.resolveActorEntity(phase.actorIndex[e.ActorId()], c.A, c, now) } } // As collisions are solved they return entities // that have been created or modified and we store // them in a map by their Id. Multiple collisions // may modify and entity, therefor we only will // one version of the entity back to engine when // we return. for _, e := range entities { remaining[e.Id()] = e } } return remainingSlice(), nil } func (phase *narrowPhase) resolveActorEntity(a *actor, with entity.Entity, collision quad.Collision, now stime.Time) []entity.Entity { switch e := with.(type) { case actorEntity: b := phase.actorIndex[e.ActorId()] return phase.solveActorActor(&solverActorActor{}, a, b, collision) case assailEntity: return phase.solveActorAssail(a, e, collision, now) case wallEntity: a.revertMoveAction() return []entity.Entity{a.Entity(), e} } return nil } func (phase *narrowPhase) solveActorAssail(a *actor, assail assailEntity, collision quad.Collision, now stime.Time) []entity.Entity { // Don't damage yourself if assail.spawnedBy == a.actorEntity.Id() { return []entity.Entity{a.Entity()} } var percentDamage float64 switch a.pathAction { case nil: if a.Cell() == assail.Cell() { percentDamage = 1.0 } default: coordCollision := coord.NewCellCollision(*a.pathAction, assail.Cell()) percentDamage = coordCollision.OverlapAt(now) } damage := int(math.Floor(float64(assail.damage) * percentDamage)) a.hp -= damage if a.hp <= 0 { a.hp = 100 a.actorEntity.cell = origin a.actorEntity.facing = coord.South a.actorEntity.pathAction = nil } return []entity.Entity{a.Entity()} } func
(a, b *actor) (*actor, *actor, coord.Collision) { var collision coord.Collision switch { case a.pathAction == nil && b.pathAction != nil: a, b = b, a fallthrough case a.pathAction != nil && b.pathAction == nil: collision = coord.NewCellCollision(*a.pathAction, b.Cell()) // A or B may have had a previous collision resolved that // caused this collision to not be possible anymore. // It is more relevant to return nil here then a // coord.Collision with type CT_NONE if collision.Type() == coord.CT_NONE { return a, b, nil } case a.pathAction != nil && b.pathAction != nil: pathCollision := coord.NewPathCollision(*a.pathAction, *b.pathAction) // coord.NewPathCollision can flip the, // A and B paths to simplify the number // of collision types. This normalizes // actor A with pathCollision.A if *a.pathAction != pathCollision.A { a, b = b, a } collision = pathCollision case a.pathAction == nil && b.pathAction == nil: // This case handles actors being on the same square, // but not moving at all. // There isn't a coord.CollisionType for this case. // Maybe there should be? return a, b, nil default: panic(fmt.Sprintf("impossible collision between {%v} {%v}", a, b)) } return a, b, collision } type node struct { actor *actor entity entity.Entity } // Move forward in the directed graph. This movement is based on // which entity is occupying the destination of the other's path action. func followGraph(a, b *actor, collision quad.Collision) node { // normalize a, b to collision.[A, B] if a.actorEntity.Id() != collision.A.Id() { a, b = b, a } var actor *actor var entity entity.Entity switch { case a.pathAction.Orig == b.pathAction.Dest: entity = collision.A actor = a case b.pathAction.Orig == a.pathAction.Dest: entity = collision.B actor = b default: panic(fmt.Sprintf("unexpected graph state %v between %v & %v", collision, a, b)) } return node{actor, entity} } // Used to figure out which actor is "A" if // the collision was CT_A_INTO_B instead of CT_NONE func currentNode(a, b *actor, collision quad.Collision) *actor { switch { case a.pathAction.Orig == b.pathAction.Dest: return b case b.pathAction.Orig == a.pathAction.Dest: return a default: panic(fmt.Sprintf("unexpected graph state %v between %v & %v", collision, a, b)) } } // Compare entity Id's with the entities in // a collision and return the one that isn't // the actor. func otherEntityIn(a *actor, collision quad.Collision) entity.Entity { var e entity.Entity // figure out is prioritized actor is A or B in the collision switch { case a.actorEntity.Id() != collision.A.Id(): e = collision.A case a.actorEntity.Id() != collision.B.Id(): e = collision.B default: panic(fmt.Sprintf("unexpected graph state %v actor %v", collision, a)) } return e } // Store what actor's have been visited during // a recursive solve. Used to avoid infinite // recursion through a cycle in the graph. type solverActorActor struct { visited []*actor } func (s solverActorActor) hasVisited(actor *actor) bool { for _, a := range s.visited { if actor == a { return true } } return false } func (phase *narrowPhase) solveActorActor(solver *solverActorActor, a, b *actor, collision quad.Collision) []entity.Entity { // When this functions returns the // collision will have been solved defer func() { phase.solved = append(phase.solved, collision) }() var entities []entity.Entity attemptSolve: a, b, coordCollision := newActorActorCollision(a, b) if coordCollision == nil { goto resolved } switch coordCollision.Type() { case coord.CT_NONE: // This may not be entirely accurate. // We should walk through the collision index // of our partner too see if they should resolve // some of there collisions first. They may // appear to be moving to us right now, but // have a collision that when solved will // render them motionless, thus we would become // motionless as well. e, err := phase.solveDependencies(solver, a, b, collision) switch err { case nil: if len(e) > 0 { entities = append(entities, e...) } // Try solving again goto attemptSolve case errCycleDetected: // Detected a cycle, we can't move currentNode(a, b, collision).revertMoveAction() goto resolved case errNoDependencies: // All dependencies have been solved // We can move goto resolved } case coord.CT_CELL_DEST: a.revertMoveAction() goto resolved case coord.CT_SWAP: a.revertMoveAction() b.revertMoveAction() goto resolved case coord.CT_A_INTO_B_FROM_SIDE: // This may not be entirely accurate. // We should walk through the collision index // of our partner too see if they should resolve // some of there collisions first. They may // appear to be moving to us right
newActorActorCollision
identifier_name
main.go
var trowner, trname string var hbrev, brbd uint var osvs []string var logger = log.New(os.Stderr, "", log.LstdFlags) func ferr(msg string, args ...interface{}) { fmt.Printf(msg+"\n", args...) os.Exit(1) } var ghc *github.Client func init() { pflag.StringVar(&rname, "release", "", "release name (ex: v1.0.0)") pflag.StringVar(&npath, "notes-path", "relnotes.md", "path to release notes") pflag.StringVar(&commitsha, "commit", "", "commit SHA to release") pflag.StringVar(&taprepo, "tap-repo", "dollarshaveclub/homebrew-public", "name of tap GitHub repository ([owner]/[repo])") pflag.StringVar(&tapref, "tap-repo-ref", "master", "tap repository ref (branch/tag/SHA)") pflag.StringVar(&fpath, "formula", "Formula/psst.rb", "path to formula within tap repo") pflag.StringVar(&ftpath, "formula-template", "Formula/psst.rb.tmpl", "path to formula template within tap repo") pflag.StringVar(&targetoslist, "macos-versions", "el_capitan,high_sierra,sierra", "Supported MacOS versions (comma-delimited)") pflag.UintVar(&hbrev, "homebrew-rev", 0, "Homebrew revision (bump to force reinstall/rebuild)") pflag.UintVar(&brbd, "bottle-rebuild", 1, "Bottle rebuild (bump to force bottle reinstall)") pflag.BoolVar(&draft, "draft", false, "Draft release (unpublished)") pflag.BoolVar(&prerelease, "prerelease", false, "Prerelease") pflag.BoolVar(&dobuild, "build", true, "Build binaries first") pflag.Parse() trs := strings.Split(taprepo, "/") if len(trs) != 2 { ferr("malformed tap repo (expected [owner]/[repo]): %v", taprepo) } if rname == "" { ferr("release name is required") } trowner = trs[0] trname = trs[1] osvs = strings.Split(targetoslist, ",") if len(osvs) == 0 { ferr("At least one MacOS version is required") } ghtoken = os.Getenv("GITHUB_TOKEN") if ghtoken == "" { ferr("GITHUB_TOKEN missing from environment") } if err := checkFiles(npath); err != nil { ferr("file path error: %v", err) } checkLocalRepoVersion() ghc = newGHClient() } func newGHClient() *github.Client { tc := oauth2.NewClient(context.Background(), oauth2.StaticTokenSource( &oauth2.Token{AccessToken: ghtoken}, )) return github.NewClient(tc) } func checkLocalRepoVersion() { cmd := exec.Command("git", "rev-parse", "HEAD") out, err := cmd.Output() if err != nil { ferr("error getting git command output: %v", err) } if strings.TrimRight(string(out), "\n") != commitsha { ferr("current git revision does not match requested release version: %v (expected %v)", string(out), commitsha) } } func checkFiles(paths ...string) error { for _, p := range paths { if _, err := os.Stat(p); err != nil { return errors.Wrap(err, "file error") } } return nil } func createGitTag() error { msg := fmt.Sprintf("release %v", rname) ot := "commit" tag := github.Tag{ Tag: &rname, Message: &msg, Object: &github.GitObject{ Type: &ot, SHA: &commitsha, }, } log.Printf("creating tag...\n") _, _, err := ghc.Git.CreateTag(context.Background(), repoOwner, repoName, &tag) if err != nil { return errors.Wrap(err, "error creating tag") } refstr := fmt.Sprintf("refs/tags/%v", rname) objt := "commit" ref := github.Reference{ Ref: &refstr, Object: &github.GitObject{ Type: &objt, SHA: &commitsha, }, } log.Printf("creating tag ref...\n") _, _, err = ghc.Git.CreateRef(context.Background(), repoOwner, repoName, &ref) if err != nil { return errors.Wrap(err, "error creating tag ref") } return nil } type bottleDefinition struct { Hash string TargetOS string } type formulaTemplateData struct { Tag string CommitSHA string HomebrewRevision uint BaseDownloadURL string Bottled bool BottleRebuild uint BottleDefs []bottleDefinition } func (ftd *formulaTemplateData) populate(bdefs []bottleDefinition) { ftd.Tag = rname ftd.CommitSHA = commitsha if hbrev > 0 { ftd.HomebrewRevision = hbrev } ftd.BaseDownloadURL = fmt.Sprintf("https://github.com/%v/%v/releases/download/%v", repoOwner, repoName, rname) ftd.BottleRebuild = brbd ftd.Bottled = true ftd.BottleDefs = bdefs } const header = "# GENERATED FROM TEMPLATE. DO NOT EDIT!\n" // generateFormula fetches the template from github, executes the template with ftd and returns the raw data or error, if any func generateFormula(ftd formulaTemplateData) ([]byte, error) { logger.Printf("Generating Homebrew formula") // get template fc, _, _, err := ghc.Repositories.GetContents(context.Background(), trowner, trname, ftpath, &github.RepositoryContentGetOptions{Ref: tapref}) if err != nil { return nil, errors.Wrap(err, "error getting formula template") } rt, err := fc.GetContent() if err != nil { return nil, errors.Wrap(err, "error getting formula template content") } // generate new formula tmpl, err := template.New("formula").Parse(rt) if err != nil { return nil, errors.Wrap(err, "error parsing formula template") } buf := bytes.NewBuffer([]byte{}) if err = tmpl.Execute(buf, &ftd); err != nil { return nil, errors.Wrap(err, "error executing template") } return append([]byte(header), buf.Bytes()...), nil } func pushFormula(fd []byte) error { logger.Printf("Pushing Homebrew formula") // Get the current file for the SHA fc, _, _, err := ghc.Repositories.GetContents(context.Background(), trowner, trname, fpath, &github.RepositoryContentGetOptions{Ref: tapref}) if err != nil { return errors.Wrap(err, "error getting formula contents") } sp := func(s string) *string { return &s } _, _, err = ghc.Repositories.UpdateFile(context.Background(), trowner, trname, fpath, &github.RepositoryContentFileOptions{ Message: sp(fmt.Sprintf("updated for release %v", rname)), Content: fd, SHA: fc.SHA, Branch: &tapref, }) if err != nil { return errors.Wrap(err, "error updating formula") } return nil } const ( linuxBinName = "psst-linux-amd64" ) var buildopts = []string{"-ldflags", "-X github.com/dollarshaveclub/psst/cmd.CommitSHA=%v -X github.com/dollarshaveclub/psst/cmd.Version=%v -X github.com/dollarshaveclub/psst/cmd.CompiledDirectory=github -X github.com/dollarshaveclub/psst/cmd.CompiledStorage=vault -X github.com/dollarshaveclub/psst/cmd.Org=dollarshaveclub"} func buildBins() error { if err := os.MkdirAll("bins", os.ModeDir|0755); err != nil { return errors.Wrap(err, "error creating bins directory") } cwd, err := os.Getwd() if err != nil { return errors.Wrap(err, "error getting working directory") } wd := filepath.Join(cwd, "..") buildopts[1] = fmt.Sprintf(buildopts[1], commitsha, rname) build := func(osn string) ([]byte, error) { cmd := exec.Command("go", append([]string{"build"}, buildopts...)...) cmd.Env = append(os.Environ(), []string{fmt.Sprintf("GOOS=%v", osn), "GOARCH=amd64"}...) cmd.Dir = wd return cmd.CombinedOutput() } logger.Printf("Building binaries...\n") logger.Printf("...macOS amd64") if out, err := build("darwin"); err != nil { return errors.Wrapf(err, "error running build command: %s", out) } if err := os.Rename(filepath.Join(wd, "psst"), filepath.Join(cwd
var rname, npath, commitsha, ghtoken, taprepo, tapref, fpath, ftpath, targetoslist string var draft, prerelease, dobuild bool
random_line_split
main.go
oauth2.StaticTokenSource( &oauth2.Token{AccessToken: ghtoken}, )) return github.NewClient(tc) } func checkLocalRepoVersion() { cmd := exec.Command("git", "rev-parse", "HEAD") out, err := cmd.Output() if err != nil { ferr("error getting git command output: %v", err) } if strings.TrimRight(string(out), "\n") != commitsha { ferr("current git revision does not match requested release version: %v (expected %v)", string(out), commitsha) } } func checkFiles(paths ...string) error { for _, p := range paths { if _, err := os.Stat(p); err != nil { return errors.Wrap(err, "file error") } } return nil } func createGitTag() error { msg := fmt.Sprintf("release %v", rname) ot := "commit" tag := github.Tag{ Tag: &rname, Message: &msg, Object: &github.GitObject{ Type: &ot, SHA: &commitsha, }, } log.Printf("creating tag...\n") _, _, err := ghc.Git.CreateTag(context.Background(), repoOwner, repoName, &tag) if err != nil { return errors.Wrap(err, "error creating tag") } refstr := fmt.Sprintf("refs/tags/%v", rname) objt := "commit" ref := github.Reference{ Ref: &refstr, Object: &github.GitObject{ Type: &objt, SHA: &commitsha, }, } log.Printf("creating tag ref...\n") _, _, err = ghc.Git.CreateRef(context.Background(), repoOwner, repoName, &ref) if err != nil { return errors.Wrap(err, "error creating tag ref") } return nil } type bottleDefinition struct { Hash string TargetOS string } type formulaTemplateData struct { Tag string CommitSHA string HomebrewRevision uint BaseDownloadURL string Bottled bool BottleRebuild uint BottleDefs []bottleDefinition } func (ftd *formulaTemplateData) populate(bdefs []bottleDefinition) { ftd.Tag = rname ftd.CommitSHA = commitsha if hbrev > 0 { ftd.HomebrewRevision = hbrev } ftd.BaseDownloadURL = fmt.Sprintf("https://github.com/%v/%v/releases/download/%v", repoOwner, repoName, rname) ftd.BottleRebuild = brbd ftd.Bottled = true ftd.BottleDefs = bdefs } const header = "# GENERATED FROM TEMPLATE. DO NOT EDIT!\n" // generateFormula fetches the template from github, executes the template with ftd and returns the raw data or error, if any func generateFormula(ftd formulaTemplateData) ([]byte, error) { logger.Printf("Generating Homebrew formula") // get template fc, _, _, err := ghc.Repositories.GetContents(context.Background(), trowner, trname, ftpath, &github.RepositoryContentGetOptions{Ref: tapref}) if err != nil { return nil, errors.Wrap(err, "error getting formula template") } rt, err := fc.GetContent() if err != nil { return nil, errors.Wrap(err, "error getting formula template content") } // generate new formula tmpl, err := template.New("formula").Parse(rt) if err != nil { return nil, errors.Wrap(err, "error parsing formula template") } buf := bytes.NewBuffer([]byte{}) if err = tmpl.Execute(buf, &ftd); err != nil { return nil, errors.Wrap(err, "error executing template") } return append([]byte(header), buf.Bytes()...), nil } func pushFormula(fd []byte) error { logger.Printf("Pushing Homebrew formula") // Get the current file for the SHA fc, _, _, err := ghc.Repositories.GetContents(context.Background(), trowner, trname, fpath, &github.RepositoryContentGetOptions{Ref: tapref}) if err != nil { return errors.Wrap(err, "error getting formula contents") } sp := func(s string) *string { return &s } _, _, err = ghc.Repositories.UpdateFile(context.Background(), trowner, trname, fpath, &github.RepositoryContentFileOptions{ Message: sp(fmt.Sprintf("updated for release %v", rname)), Content: fd, SHA: fc.SHA, Branch: &tapref, }) if err != nil { return errors.Wrap(err, "error updating formula") } return nil } const ( linuxBinName = "psst-linux-amd64" ) var buildopts = []string{"-ldflags", "-X github.com/dollarshaveclub/psst/cmd.CommitSHA=%v -X github.com/dollarshaveclub/psst/cmd.Version=%v -X github.com/dollarshaveclub/psst/cmd.CompiledDirectory=github -X github.com/dollarshaveclub/psst/cmd.CompiledStorage=vault -X github.com/dollarshaveclub/psst/cmd.Org=dollarshaveclub"} func buildBins() error { if err := os.MkdirAll("bins", os.ModeDir|0755); err != nil { return errors.Wrap(err, "error creating bins directory") } cwd, err := os.Getwd() if err != nil { return errors.Wrap(err, "error getting working directory") } wd := filepath.Join(cwd, "..") buildopts[1] = fmt.Sprintf(buildopts[1], commitsha, rname) build := func(osn string) ([]byte, error) { cmd := exec.Command("go", append([]string{"build"}, buildopts...)...) cmd.Env = append(os.Environ(), []string{fmt.Sprintf("GOOS=%v", osn), "GOARCH=amd64"}...) cmd.Dir = wd return cmd.CombinedOutput() } logger.Printf("Building binaries...\n") logger.Printf("...macOS amd64") if out, err := build("darwin"); err != nil { return errors.Wrapf(err, "error running build command: %s", out) } if err := os.Rename(filepath.Join(wd, "psst"), filepath.Join(cwd, "bins", "psst-darwin")); err != nil { return errors.Wrap(err, "error renaming binary") } logger.Printf("...Linux amd64") if out, err := build("linux"); err != nil { return errors.Wrapf(err, "error running build command: %s", out) } lfn := filepath.Join(cwd, "bins", linuxBinName) if err := os.Rename(filepath.Join(wd, "psst"), lfn); err != nil { return errors.Wrap(err, "error renaming binary") } // compress linux binary logger.Printf("...compressing Linux binary\n") d, err := ioutil.ReadFile(lfn) if err != nil { return errors.Wrap(err, "error reading linux binary") } f, err := os.Create(lfn + ".gz") if err != nil { return errors.Wrap(err, "error creating compressed linux binary") } defer f.Close() gw := gzip.NewWriter(f) defer gw.Close() if _, err := gw.Write(d); err != nil { return errors.Wrap(err, "error writing compressed linux binary") } return nil } // "copy" (link) a file if it doesn't exist func cpifneeded(src, dest string) error { if _, err := os.Stat(dest); err != nil { if os.IsNotExist(err) { return os.Link(src, dest) } return errors.Wrap(err, "error getting destination") } return nil } var bottleNameTmpl = template.Must(template.New("bn").Parse("psst-{{ .Release }}{{ if .HomebrewRevision }}_{{ .HomebrewRevision }}{{ end }}.{{ .OS }}.bottle.{{ .BottleRebuild }}.tar.gz")) // createBottle synthetically creates a bottle tarball returning the bottle definitions, local bottle filenames and error if any func createBottle() ([]bottleDefinition, []string, error)
{ logger.Printf("Creating Homebrew bottle...\n") cwd, err := os.Getwd() if err != nil { return nil, nil, errors.Wrap(err, "error getting working directory") } rver := regexp.MustCompile("([0-9.]+)").FindString(rname) basepath := filepath.Join(".", "psst", rver) binpath := filepath.Join(basepath, "bin") if err := os.MkdirAll(binpath, os.ModeDir|0755); err != nil { return nil, nil, errors.Wrap(err, "error creating bottle directory path") } // .brew if err := os.MkdirAll(filepath.Join(basepath, ".brew"), os.ModeDir|0755); err != nil { return nil, nil, errors.Wrap(err, "error creating .brew directory") } // copy README if err := cpifneeded(filepath.Join(cwd, "..", "README.md"), filepath.Join(basepath, "README.md")); err != nil { return nil, nil, errors.Wrap(err, "error copying README") }
identifier_body