file_name
large_stringlengths
4
140
prefix
large_stringlengths
0
12.1k
suffix
large_stringlengths
0
12k
middle
large_stringlengths
0
7.51k
fim_type
large_stringclasses
4 values
flap.py
FPS = 30 # tweak this to change how quickly the bird falls GRAVITY = 1 # how big the gaps between the pipes are GAP = 100 # how frequently the pipes spawn (sec) FREQ = 2 # how fast the bird flies at the pipes PIPE_SPEED = 3 # how powerful is a flap? FLAP_SPEED = 15 # set up asset folders game_dir = path.dirname(__file__) img_dir = path.join(game_dir, 'img') snd_dir = path.join(game_dir, 'snd') class SpriteSheet: """Utility class to load and parse spritesheets""" def __init__(self, filename): self.sprite_sheet = pygame.image.load(filename).convert() def get_image(self, x, y, width, height): # grab an image out of a larger spritesheet image = pygame.Surface([width, height]).convert() image.blit(self.sprite_sheet, (0, 0), (x, y, width, height)) image.set_colorkey([0, 0, 0]) return image class Bird(pygame.sprite.Sprite): # player controlled bird, can only flap width = 36 height = 24 def __init__(self): # when you make a Pygame Sprite object, you have to call the # Sprite init function pygame.sprite.Sprite.__init__(self) self.alive = True self.speed_x = 0 self.speed_y = 0 self.flap_snd = pygame.mixer.Sound(path.join(snd_dir, "bird_flap.wav")) self.flap_snd.set_volume(0.2) self.frames = [] sprite_sheet = SpriteSheet(path.join(img_dir, "bird_sprites.png")) image = sprite_sheet.get_image(3, 7, 34, 24) image.set_colorkey(BLACK) self.frames.append(image) self.image_dead = pygame.transform.rotate(image, -90) image = sprite_sheet.get_image(59, 7, 34, 24) image.set_colorkey(BLACK) self.frames.append(image) image = sprite_sheet.get_image(115, 7, 34, 24) image.set_colorkey(BLACK) self.frames.append(image) self.index = 0 self.image = self.frames[self.index] self.mask = pygame.mask.from_surface(self.image) self.rect = self.image.get_rect() # start in the middle of the screen self.rect.centerx = WIDTH / 2 self.rect.y = HEIGHT / 2 def update(self): # gravity pulls downward self.speed_y += gravity # move self.rect.x += self.speed_x self.rect.y += self.speed_y if self.alive: # animate self.index += 1 if self.index >= len(self.frames): self.index = 0 self.image = self.frames[self.index] else: self.image = self.image_dead # stop at the top/bottom if self.rect.top < 0: self.rect.top = 0 self.speed_y = 0 if self.rect.bottom > HEIGHT-50: self.rect.bottom = HEIGHT-50 self.speed_y = 0 def move(self): # player hit SPACEBAR if not self.alive: return self.speed_y -= FLAP_SPEED self.flap_snd.play() class Pipe(pygame.sprite.Sprite): # pipe segment class speed_x = -PIPE_SPEED width = 36 def __init__(self, loc, y): # loc = upper or lower # y = where to place it pygame.sprite.Sprite.__init__(self) sprite_sheet = SpriteSheet(path.join(img_dir, "pipes.png")) if loc == 'u': self.image = sprite_sheet.get_image(2, 8, 52, 320) else: self.image = sprite_sheet.get_image(58, 8, 52, 320) self.image.set_colorkey(BLACK) self.mask = pygame.mask.from_surface(self.image) self.rect = self.image.get_rect() # start offscreen to the right self.rect.x = WIDTH + 50 if loc == 'u':
else: self.rect.top = y # has the bird passed this pipe? self.passed = False def update(self): # move self.rect.x += self.speed_x def offscreen(self): # test to see if the pipe has moved offscreen if self.rect.right < 0: return True else: return False def new_pipe(): # create a new pair of pipes (upper and lower) y = random.randrange(30, HEIGHT-50-GAP) pipe_u = Pipe('u', y) pipe_l = Pipe('l', y + GAP) return pipe_u, pipe_l def draw_background(): # draw the background (tiled) background_rect.bottom = HEIGHT + 20 background_rect.left = 0 screen.blit(background, background_rect) if background_rect.width < WIDTH: background_rect.left = background_rect.width screen.blit(background, background_rect) def draw_ground(): # draw the ground tiles, moving at the same speed as pipes for image in ground_list: image.x -= PIPE_SPEED if image.right <= 0: # if the image has completely moved off the screen, move it to the right image.left = 2 * ground.get_width() + image.right screen.blit(ground, image) def draw_text(text, size, x, y): # utility function to draw text at a given location font_name = pygame.font.match_font('arial') font = pygame.font.Font(font_name, size) text_surface = font.render(text, True, WHITE) text_rect = text_surface.get_rect() text_rect.midtop = (x, y) screen.blit(text_surface, text_rect) def show_go_image(): go_rect.midtop = (WIDTH/2, HEIGHT/2) screen.blit(go_image, go_rect) def show_ready_image(): ready_rect.midtop = (WIDTH/2, HEIGHT*2/3) screen.blit(ready_image, ready_rect) def load_score_images(): sprite_sheet = SpriteSheet(path.join(img_dir, 'numbers.png')) score_images = [] image = sprite_sheet.get_image(114, 45, 24, 36) score_images.append(image) image = sprite_sheet.get_image(2, 4, 24, 36) score_images.append(image) image = sprite_sheet.get_image(30, 4, 24, 36) score_images.append(image) image = sprite_sheet.get_image(58, 4, 24, 36) score_images.append(image) image = sprite_sheet.get_image(86, 4, 24, 36) score_images.append(image) image = sprite_sheet.get_image(114, 4, 24, 36) score_images.append(image) image = sprite_sheet.get_image(2, 45, 24, 36) score_images.append(image) image = sprite_sheet.get_image(30, 45, 24, 36) score_images.append(image) image = sprite_sheet.get_image(58, 45, 24, 36) score_images.append(image) image = sprite_sheet.get_image(86, 45, 24, 36) score_images.append(image) return score_images def show_score(score): # show the score using the score images # draw_text(str(int(score)), 22, WIDTH/2, 10) digits = [int(c) for c in str(score)] for i, digit in enumerate(digits): img = score_images[digit] img_rect = img.get_rect() img_rect.y = 5 img_rect.x = i * img_rect.width + 5 screen.blit(img, img_rect) # initialize pygame pygame.init() # initialize sound pygame.mixer.init() screen = pygame.display.set_mode((WIDTH, HEIGHT)) pygame.display.set_caption("Flap") try: pygame.mixer.music.load(path.join(snd_dir, "Choro_bavario_loop.ogg")) pygame.mixer.music.set_volume(0.5) pygame.mixer.music.play(loops=-1) except: print("Can't load music.") # background background = pygame.image.load(path.join(img_dir, "background.png")).convert() background_rect = background.get_rect() background_rect.bottom = HEIGHT background_rect.left = 0 # load some other images we need go_image = pygame.image.load(path.join(img_dir, "gameover.png")).convert() go_image.set_colorkey(BLACK) go_rect = go_image.get_rect() ready_image = pygame.image.load(path.join(img_dir, "getready.png")).convert() ready_image.set_colorkey(BLACK) ready_rect = ready_image.get_rect() score_images = load_score_images() # load the ground tile images
self.rect.bottom = y
conditional_block
flap.py
= 30 # tweak this to change how quickly the bird falls GRAVITY = 1 # how big the gaps between the pipes are GAP = 100 # how frequently the pipes spawn (sec) FREQ = 2 # how fast the bird flies at the pipes PIPE_SPEED = 3 # how powerful is a flap? FLAP_SPEED = 15 # set up asset folders game_dir = path.dirname(__file__) img_dir = path.join(game_dir, 'img') snd_dir = path.join(game_dir, 'snd') class SpriteSheet: """Utility class to load and parse spritesheets""" def __init__(self, filename): self.sprite_sheet = pygame.image.load(filename).convert() def get_image(self, x, y, width, height): # grab an image out of a larger spritesheet image = pygame.Surface([width, height]).convert() image.blit(self.sprite_sheet, (0, 0), (x, y, width, height)) image.set_colorkey([0, 0, 0]) return image class Bird(pygame.sprite.Sprite): # player controlled bird, can only flap width = 36 height = 24 def __init__(self): # when you make a Pygame Sprite object, you have to call the # Sprite init function pygame.sprite.Sprite.__init__(self) self.alive = True self.speed_x = 0 self.speed_y = 0 self.flap_snd = pygame.mixer.Sound(path.join(snd_dir, "bird_flap.wav")) self.flap_snd.set_volume(0.2) self.frames = [] sprite_sheet = SpriteSheet(path.join(img_dir, "bird_sprites.png")) image = sprite_sheet.get_image(3, 7, 34, 24) image.set_colorkey(BLACK) self.frames.append(image) self.image_dead = pygame.transform.rotate(image, -90) image = sprite_sheet.get_image(59, 7, 34, 24) image.set_colorkey(BLACK) self.frames.append(image) image = sprite_sheet.get_image(115, 7, 34, 24) image.set_colorkey(BLACK) self.frames.append(image) self.index = 0 self.image = self.frames[self.index] self.mask = pygame.mask.from_surface(self.image) self.rect = self.image.get_rect() # start in the middle of the screen self.rect.centerx = WIDTH / 2 self.rect.y = HEIGHT / 2 def
(self): # gravity pulls downward self.speed_y += gravity # move self.rect.x += self.speed_x self.rect.y += self.speed_y if self.alive: # animate self.index += 1 if self.index >= len(self.frames): self.index = 0 self.image = self.frames[self.index] else: self.image = self.image_dead # stop at the top/bottom if self.rect.top < 0: self.rect.top = 0 self.speed_y = 0 if self.rect.bottom > HEIGHT-50: self.rect.bottom = HEIGHT-50 self.speed_y = 0 def move(self): # player hit SPACEBAR if not self.alive: return self.speed_y -= FLAP_SPEED self.flap_snd.play() class Pipe(pygame.sprite.Sprite): # pipe segment class speed_x = -PIPE_SPEED width = 36 def __init__(self, loc, y): # loc = upper or lower # y = where to place it pygame.sprite.Sprite.__init__(self) sprite_sheet = SpriteSheet(path.join(img_dir, "pipes.png")) if loc == 'u': self.image = sprite_sheet.get_image(2, 8, 52, 320) else: self.image = sprite_sheet.get_image(58, 8, 52, 320) self.image.set_colorkey(BLACK) self.mask = pygame.mask.from_surface(self.image) self.rect = self.image.get_rect() # start offscreen to the right self.rect.x = WIDTH + 50 if loc == 'u': self.rect.bottom = y else: self.rect.top = y # has the bird passed this pipe? self.passed = False def update(self): # move self.rect.x += self.speed_x def offscreen(self): # test to see if the pipe has moved offscreen if self.rect.right < 0: return True else: return False def new_pipe(): # create a new pair of pipes (upper and lower) y = random.randrange(30, HEIGHT-50-GAP) pipe_u = Pipe('u', y) pipe_l = Pipe('l', y + GAP) return pipe_u, pipe_l def draw_background(): # draw the background (tiled) background_rect.bottom = HEIGHT + 20 background_rect.left = 0 screen.blit(background, background_rect) if background_rect.width < WIDTH: background_rect.left = background_rect.width screen.blit(background, background_rect) def draw_ground(): # draw the ground tiles, moving at the same speed as pipes for image in ground_list: image.x -= PIPE_SPEED if image.right <= 0: # if the image has completely moved off the screen, move it to the right image.left = 2 * ground.get_width() + image.right screen.blit(ground, image) def draw_text(text, size, x, y): # utility function to draw text at a given location font_name = pygame.font.match_font('arial') font = pygame.font.Font(font_name, size) text_surface = font.render(text, True, WHITE) text_rect = text_surface.get_rect() text_rect.midtop = (x, y) screen.blit(text_surface, text_rect) def show_go_image(): go_rect.midtop = (WIDTH/2, HEIGHT/2) screen.blit(go_image, go_rect) def show_ready_image(): ready_rect.midtop = (WIDTH/2, HEIGHT*2/3) screen.blit(ready_image, ready_rect) def load_score_images(): sprite_sheet = SpriteSheet(path.join(img_dir, 'numbers.png')) score_images = [] image = sprite_sheet.get_image(114, 45, 24, 36) score_images.append(image) image = sprite_sheet.get_image(2, 4, 24, 36) score_images.append(image) image = sprite_sheet.get_image(30, 4, 24, 36) score_images.append(image) image = sprite_sheet.get_image(58, 4, 24, 36) score_images.append(image) image = sprite_sheet.get_image(86, 4, 24, 36) score_images.append(image) image = sprite_sheet.get_image(114, 4, 24, 36) score_images.append(image) image = sprite_sheet.get_image(2, 45, 24, 36) score_images.append(image) image = sprite_sheet.get_image(30, 45, 24, 36) score_images.append(image) image = sprite_sheet.get_image(58, 45, 24, 36) score_images.append(image) image = sprite_sheet.get_image(86, 45, 24, 36) score_images.append(image) return score_images def show_score(score): # show the score using the score images # draw_text(str(int(score)), 22, WIDTH/2, 10) digits = [int(c) for c in str(score)] for i, digit in enumerate(digits): img = score_images[digit] img_rect = img.get_rect() img_rect.y = 5 img_rect.x = i * img_rect.width + 5 screen.blit(img, img_rect) # initialize pygame pygame.init() # initialize sound pygame.mixer.init() screen = pygame.display.set_mode((WIDTH, HEIGHT)) pygame.display.set_caption("Flap") try: pygame.mixer.music.load(path.join(snd_dir, "Choro_bavario_loop.ogg")) pygame.mixer.music.set_volume(0.5) pygame.mixer.music.play(loops=-1) except: print("Can't load music.") # background background = pygame.image.load(path.join(img_dir, "background.png")).convert() background_rect = background.get_rect() background_rect.bottom = HEIGHT background_rect.left = 0 # load some other images we need go_image = pygame.image.load(path.join(img_dir, "gameover.png")).convert() go_image.set_colorkey(BLACK) go_rect = go_image.get_rect() ready_image = pygame.image.load(path.join(img_dir, "getready.png")).convert() ready_image.set_colorkey(BLACK) ready_rect = ready_image.get_rect() score_images = load_score_images() # load the ground tile images
update
identifier_name
flap.py
FPS = 30 # tweak this to change how quickly the bird falls GRAVITY = 1 # how big the gaps between the pipes are GAP = 100 # how frequently the pipes spawn (sec) FREQ = 2 # how fast the bird flies at the pipes PIPE_SPEED = 3 # how powerful is a flap?
FLAP_SPEED = 15 # set up asset folders game_dir = path.dirname(__file__) img_dir = path.join(game_dir, 'img') snd_dir = path.join(game_dir, 'snd') class SpriteSheet: """Utility class to load and parse spritesheets""" def __init__(self, filename): self.sprite_sheet = pygame.image.load(filename).convert() def get_image(self, x, y, width, height): # grab an image out of a larger spritesheet image = pygame.Surface([width, height]).convert() image.blit(self.sprite_sheet, (0, 0), (x, y, width, height)) image.set_colorkey([0, 0, 0]) return image class Bird(pygame.sprite.Sprite): # player controlled bird, can only flap width = 36 height = 24 def __init__(self): # when you make a Pygame Sprite object, you have to call the # Sprite init function pygame.sprite.Sprite.__init__(self) self.alive = True self.speed_x = 0 self.speed_y = 0 self.flap_snd = pygame.mixer.Sound(path.join(snd_dir, "bird_flap.wav")) self.flap_snd.set_volume(0.2) self.frames = [] sprite_sheet = SpriteSheet(path.join(img_dir, "bird_sprites.png")) image = sprite_sheet.get_image(3, 7, 34, 24) image.set_colorkey(BLACK) self.frames.append(image) self.image_dead = pygame.transform.rotate(image, -90) image = sprite_sheet.get_image(59, 7, 34, 24) image.set_colorkey(BLACK) self.frames.append(image) image = sprite_sheet.get_image(115, 7, 34, 24) image.set_colorkey(BLACK) self.frames.append(image) self.index = 0 self.image = self.frames[self.index] self.mask = pygame.mask.from_surface(self.image) self.rect = self.image.get_rect() # start in the middle of the screen self.rect.centerx = WIDTH / 2 self.rect.y = HEIGHT / 2 def update(self): # gravity pulls downward self.speed_y += gravity # move self.rect.x += self.speed_x self.rect.y += self.speed_y if self.alive: # animate self.index += 1 if self.index >= len(self.frames): self.index = 0 self.image = self.frames[self.index] else: self.image = self.image_dead # stop at the top/bottom if self.rect.top < 0: self.rect.top = 0 self.speed_y = 0 if self.rect.bottom > HEIGHT-50: self.rect.bottom = HEIGHT-50 self.speed_y = 0 def move(self): # player hit SPACEBAR if not self.alive: return self.speed_y -= FLAP_SPEED self.flap_snd.play() class Pipe(pygame.sprite.Sprite): # pipe segment class speed_x = -PIPE_SPEED width = 36 def __init__(self, loc, y): # loc = upper or lower # y = where to place it pygame.sprite.Sprite.__init__(self) sprite_sheet = SpriteSheet(path.join(img_dir, "pipes.png")) if loc == 'u': self.image = sprite_sheet.get_image(2, 8, 52, 320) else: self.image = sprite_sheet.get_image(58, 8, 52, 320) self.image.set_colorkey(BLACK) self.mask = pygame.mask.from_surface(self.image) self.rect = self.image.get_rect() # start offscreen to the right self.rect.x = WIDTH + 50 if loc == 'u': self.rect.bottom = y else: self.rect.top = y # has the bird passed this pipe? self.passed = False def update(self): # move self.rect.x += self.speed_x def offscreen(self): # test to see if the pipe has moved offscreen if self.rect.right < 0: return True else: return False def new_pipe(): # create a new pair of pipes (upper and lower) y = random.randrange(30, HEIGHT-50-GAP) pipe_u = Pipe('u', y) pipe_l = Pipe('l', y + GAP) return pipe_u, pipe_l def draw_background(): # draw the background (tiled) background_rect.bottom = HEIGHT + 20 background_rect.left = 0 screen.blit(background, background_rect) if background_rect.width < WIDTH: background_rect.left = background_rect.width screen.blit(background, background_rect) def draw_ground(): # draw the ground tiles, moving at the same speed as pipes for image in ground_list: image.x -= PIPE_SPEED if image.right <= 0: # if the image has completely moved off the screen, move it to the right image.left = 2 * ground.get_width() + image.right screen.blit(ground, image) def draw_text(text, size, x, y): # utility function to draw text at a given location font_name = pygame.font.match_font('arial') font = pygame.font.Font(font_name, size) text_surface = font.render(text, True, WHITE) text_rect = text_surface.get_rect() text_rect.midtop = (x, y) screen.blit(text_surface, text_rect) def show_go_image(): go_rect.midtop = (WIDTH/2, HEIGHT/2) screen.blit(go_image, go_rect) def show_ready_image(): ready_rect.midtop = (WIDTH/2, HEIGHT*2/3) screen.blit(ready_image, ready_rect) def load_score_images(): sprite_sheet = SpriteSheet(path.join(img_dir, 'numbers.png')) score_images = [] image = sprite_sheet.get_image(114, 45, 24, 36) score_images.append(image) image = sprite_sheet.get_image(2, 4, 24, 36) score_images.append(image) image = sprite_sheet.get_image(30, 4, 24, 36) score_images.append(image) image = sprite_sheet.get_image(58, 4, 24, 36) score_images.append(image) image = sprite_sheet.get_image(86, 4, 24, 36) score_images.append(image) image = sprite_sheet.get_image(114, 4, 24, 36) score_images.append(image) image = sprite_sheet.get_image(2, 45, 24, 36) score_images.append(image) image = sprite_sheet.get_image(30, 45, 24, 36) score_images.append(image) image = sprite_sheet.get_image(58, 45, 24, 36) score_images.append(image) image = sprite_sheet.get_image(86, 45, 24, 36) score_images.append(image) return score_images def show_score(score): # show the score using the score images # draw_text(str(int(score)), 22, WIDTH/2, 10) digits = [int(c) for c in str(score)] for i, digit in enumerate(digits): img = score_images[digit] img_rect = img.get_rect() img_rect.y = 5 img_rect.x = i * img_rect.width + 5 screen.blit(img, img_rect) # initialize pygame pygame.init() # initialize sound pygame.mixer.init() screen = pygame.display.set_mode((WIDTH, HEIGHT)) pygame.display.set_caption("Flap") try: pygame.mixer.music.load(path.join(snd_dir, "Choro_bavario_loop.ogg")) pygame.mixer.music.set_volume(0.5) pygame.mixer.music.play(loops=-1) except: print("Can't load music.") # background background = pygame.image.load(path.join(img_dir, "background.png")).convert() background_rect = background.get_rect() background_rect.bottom = HEIGHT background_rect.left = 0 # load some other images we need go_image = pygame.image.load(path.join(img_dir, "gameover.png")).convert() go_image.set_colorkey(BLACK) go_rect = go_image.get_rect() ready_image = pygame.image.load(path.join(img_dir, "getready.png")).convert() ready_image.set_colorkey(BLACK) ready_rect = ready_image.get_rect() score_images = load_score_images() # load the ground tile images
random_line_split
kubernetes_decorator.py
a step to execute on Kubernetes regardless of the ```--with``` specification. To use, annotate your step as follows: ``` @kubernetes @step def my_step(self): ... ``` Parameters ---------- cpu : int Number of CPUs required for this step. Defaults to 1. If @resources is also present, the maximum value from all decorators is used memory : int Memory size (in MB) required for this step. Defaults to 4096. If @resources is also present, the maximum value from all decorators is used disk : int Disk size (in MB) required for this step. Defaults to 10GB. If @resources is also present, the maximum value from all decorators is used image : string Docker image to use when launching on Kubernetes. If not specified, a default docker image mapping to the current version of Python is used """ name = "kubernetes" defaults = { "cpu": "1", "memory": "4096", "disk": "10240", "image": None, "service_account": None, "secrets": None, # e.g., mysecret "node_selector": None, # e.g., kubernetes.io/os=linux "namespace": None, "gpu": None, # value of 0 implies that the scheduled node should not have GPUs "gpu_vendor": None, } package_url = None package_sha = None run_time_limit = None def __init__(self, attributes=None, statically_defined=False): super(KubernetesDecorator, self).__init__(attributes, statically_defined) if not self.attributes["namespace"]: self.attributes["namespace"] = KUBERNETES_NAMESPACE if not self.attributes["service_account"]: self.attributes["service_account"] = KUBERNETES_SERVICE_ACCOUNT if not self.attributes["gpu_vendor"]: self.attributes["gpu_vendor"] = KUBERNETES_GPU_VENDOR # TODO: Handle node_selector in a better manner. Currently it is special # cased in kubernetes_client.py # If no docker image is explicitly specified, impute a default image. if not self.attributes["image"]: # If metaflow-config specifies a docker image, just use that. if KUBERNETES_CONTAINER_IMAGE: self.attributes["image"] = KUBERNETES_CONTAINER_IMAGE # If metaflow-config doesn't specify a docker image, assign a # default docker image. else: # Default to vanilla Python image corresponding to major.minor # version of the Python interpreter launching the flow. self.attributes["image"] = "python:%s.%s" % ( platform.python_version_tuple()[0], platform.python_version_tuple()[1], ) # Assign docker registry URL for the image. if not get_docker_registry(self.attributes["image"]): if KUBERNETES_CONTAINER_REGISTRY: self.attributes["image"] = "%s/%s" % ( KUBERNETES_CONTAINER_REGISTRY.rstrip("/"), self.attributes["image"], ) # Refer https://github.com/Netflix/metaflow/blob/master/docs/lifecycle.png def step_init(self, flow, graph, step, decos, environment, flow_datastore, logger): # Executing Kubernetes jobs requires a non-local datastore. if flow_datastore.TYPE != "s3": raise KubernetesException( "The *@kubernetes* decorator requires --datastore=s3 at the moment." ) # Set internal state. self.logger = logger self.environment = environment self.step = step self.flow_datastore = flow_datastore if any([deco.name == "batch" for deco in decos]): raise MetaflowException( "Step *{step}* is marked for execution both on AWS Batch and " "Kubernetes. Please use one or the other.".format(step=step) ) for deco in decos: if getattr(deco, "IS_PARALLEL", False): raise KubernetesException( "@kubernetes does not support parallel execution currently." ) # Set run time limit for the Kubernetes job. self.run_time_limit = get_run_time_limit_for_task(decos) if self.run_time_limit < 60: raise KubernetesException( "The timeout for step *{step}* should be at least 60 seconds for " "execution on Kubernetes.".format(step=step) ) for deco in decos: if isinstance(deco, ResourcesDecorator): for k, v in deco.attributes.items(): # TODO: Special case GPUs when they are introduced in @resources. if k in self.attributes: if self.defaults[k] is None: # skip if expected value isn't an int/float
# We use the larger of @resources and @batch attributes # TODO: Fix https://github.com/Netflix/metaflow/issues/467 my_val = self.attributes.get(k) if not (my_val is None and v is None): self.attributes[k] = str( max(float(my_val or 0), float(v or 0)) ) # Check GPU vendor. if self.attributes["gpu_vendor"].lower() not in ("amd", "nvidia"): raise KubernetesException( "GPU vendor *{}* for step *{step}* is not currently supported.".format( self.attributes["gpu_vendor"], step=step ) ) # CPU, Disk, and Memory values should be greater than 0. for attr in ["cpu", "disk", "memory"]: if not ( isinstance(self.attributes[attr], (int, unicode, basestring, float)) and float(self.attributes[attr]) > 0 ): raise KubernetesException( "Invalid {} value *{}* for step *{step}*; it should be greater than 0".format( attr, self.attributes[attr], step=step ) ) if self.attributes["gpu"] is not None and not ( isinstance(self.attributes["gpu"], (int, unicode, basestring)) and float(self.attributes["gpu"]).is_integer() ): raise KubernetesException( "Invalid GPU value *{}* for step *{step}*; it should be an integer".format( self.attributes["gpu"], step=step ) ) def package_init(self, flow, step_name, environment): try: # Kubernetes is a soft dependency. from kubernetes import client, config except (NameError, ImportError): raise KubernetesException( "Could not import module 'kubernetes'.\n\nInstall Kubernetes " "Python package (https://pypi.org/project/kubernetes/) first.\n" "You can install the module by executing - " "%s -m pip install kubernetes\n" "or equivalent through your favorite Python package manager." % sys.executable ) def runtime_init(self, flow, graph, package, run_id): # Set some more internal state. self.flow = flow self.graph = graph self.package = package self.run_id = run_id def runtime_task_created( self, task_datastore, task_id, split_index, input_paths, is_cloned, ubf_context ): # To execute the Kubernetes job, the job container needs to have # access to the code package. We store the package in the datastore # which the pod is able to download as part of it's entrypoint. if not is_cloned: self._save_package_once(self.flow_datastore, self.package) def runtime_step_cli( self, cli_args, retry_count, max_user_code_retries, ubf_context ): if retry_count <= max_user_code_retries: # After all attempts to run the user code have failed, we don't need # to execute on Kubernetes anymore. We can execute possible fallback # code locally. cli_args.commands = ["kubernetes", "step"] cli_args.command_args.append(self.package_sha) cli_args.command_args.append(self.package_url) # --namespace is used to specify Metaflow namespace (a different # concept from k8s namespace). for k, v in self.attributes.items(): if k == "namespace": cli_args.command_options["k8s_namespace"] = v else: cli_args.command_options[k] = v cli_args.command_options["run-time-limit"] = self.run_time_limit cli_args.entrypoint[0] = sys.executable def task_pre_step( self, step_name, task_datastore, metadata, run_id, task_id, flow, graph, retry_count, max_retries, ubf_context, inputs, ): self.metadata = metadata self.task_datastore = task_datastore # task_pre_step may run locally if fallback is activated for @catch # decorator. In that scenario, we skip collecting Kubernetes execution # metadata. A rudimentary way to detect non-local execution is to # check for the existence of METAFLOW_KUBERNETES_WORKLOAD environment # variable. if "METAFLOW_KUBERNETES_WORKLOAD" in os.environ: meta = {} meta["kubernetes-pod-name"] =
continue
conditional_block
kubernetes_decorator.py
a step to execute on Kubernetes regardless of the ```--with``` specification. To use, annotate your step as follows: ``` @kubernetes @step def my_step(self): ... ``` Parameters ---------- cpu : int Number of CPUs required for this step. Defaults to 1. If @resources is also present, the maximum value from all decorators is used memory : int Memory size (in MB) required for this step. Defaults to 4096. If @resources is also present, the maximum value from all decorators is used disk : int Disk size (in MB) required for this step. Defaults to 10GB. If @resources is also present, the maximum value from all decorators is used image : string Docker image to use when launching on Kubernetes. If not specified, a default docker image mapping to the current version of Python is used """ name = "kubernetes" defaults = { "cpu": "1", "memory": "4096", "disk": "10240", "image": None, "service_account": None, "secrets": None, # e.g., mysecret "node_selector": None, # e.g., kubernetes.io/os=linux "namespace": None, "gpu": None, # value of 0 implies that the scheduled node should not have GPUs "gpu_vendor": None, } package_url = None package_sha = None run_time_limit = None def __init__(self, attributes=None, statically_defined=False): super(KubernetesDecorator, self).__init__(attributes, statically_defined) if not self.attributes["namespace"]: self.attributes["namespace"] = KUBERNETES_NAMESPACE if not self.attributes["service_account"]: self.attributes["service_account"] = KUBERNETES_SERVICE_ACCOUNT if not self.attributes["gpu_vendor"]: self.attributes["gpu_vendor"] = KUBERNETES_GPU_VENDOR # TODO: Handle node_selector in a better manner. Currently it is special # cased in kubernetes_client.py # If no docker image is explicitly specified, impute a default image. if not self.attributes["image"]: # If metaflow-config specifies a docker image, just use that. if KUBERNETES_CONTAINER_IMAGE: self.attributes["image"] = KUBERNETES_CONTAINER_IMAGE # If metaflow-config doesn't specify a docker image, assign a # default docker image. else: # Default to vanilla Python image corresponding to major.minor # version of the Python interpreter launching the flow. self.attributes["image"] = "python:%s.%s" % ( platform.python_version_tuple()[0], platform.python_version_tuple()[1], ) # Assign docker registry URL for the image. if not get_docker_registry(self.attributes["image"]): if KUBERNETES_CONTAINER_REGISTRY: self.attributes["image"] = "%s/%s" % ( KUBERNETES_CONTAINER_REGISTRY.rstrip("/"), self.attributes["image"], ) # Refer https://github.com/Netflix/metaflow/blob/master/docs/lifecycle.png def step_init(self, flow, graph, step, decos, environment, flow_datastore, logger): # Executing Kubernetes jobs requires a non-local datastore. if flow_datastore.TYPE != "s3": raise KubernetesException( "The *@kubernetes* decorator requires --datastore=s3 at the moment." ) # Set internal state. self.logger = logger self.environment = environment self.step = step self.flow_datastore = flow_datastore if any([deco.name == "batch" for deco in decos]): raise MetaflowException( "Step *{step}* is marked for execution both on AWS Batch and " "Kubernetes. Please use one or the other.".format(step=step) ) for deco in decos: if getattr(deco, "IS_PARALLEL", False): raise KubernetesException( "@kubernetes does not support parallel execution currently." ) # Set run time limit for the Kubernetes job. self.run_time_limit = get_run_time_limit_for_task(decos) if self.run_time_limit < 60: raise KubernetesException( "The timeout for step *{step}* should be at least 60 seconds for " "execution on Kubernetes.".format(step=step) ) for deco in decos: if isinstance(deco, ResourcesDecorator): for k, v in deco.attributes.items(): # TODO: Special case GPUs when they are introduced in @resources. if k in self.attributes: if self.defaults[k] is None: # skip if expected value isn't an int/float continue # We use the larger of @resources and @batch attributes # TODO: Fix https://github.com/Netflix/metaflow/issues/467 my_val = self.attributes.get(k) if not (my_val is None and v is None): self.attributes[k] = str( max(float(my_val or 0), float(v or 0)) ) # Check GPU vendor. if self.attributes["gpu_vendor"].lower() not in ("amd", "nvidia"): raise KubernetesException( "GPU vendor *{}* for step *{step}* is not currently supported.".format( self.attributes["gpu_vendor"], step=step ) ) # CPU, Disk, and Memory values should be greater than 0. for attr in ["cpu", "disk", "memory"]: if not ( isinstance(self.attributes[attr], (int, unicode, basestring, float)) and float(self.attributes[attr]) > 0 ): raise KubernetesException( "Invalid {} value *{}* for step *{step}*; it should be greater than 0".format( attr, self.attributes[attr], step=step ) ) if self.attributes["gpu"] is not None and not ( isinstance(self.attributes["gpu"], (int, unicode, basestring)) and float(self.attributes["gpu"]).is_integer() ): raise KubernetesException( "Invalid GPU value *{}* for step *{step}*; it should be an integer".format( self.attributes["gpu"], step=step ) ) def package_init(self, flow, step_name, environment): try: # Kubernetes is a soft dependency. from kubernetes import client, config except (NameError, ImportError): raise KubernetesException( "Could not import module 'kubernetes'.\n\nInstall Kubernetes " "Python package (https://pypi.org/project/kubernetes/) first.\n" "You can install the module by executing - " "%s -m pip install kubernetes\n" "or equivalent through your favorite Python package manager." % sys.executable ) def runtime_init(self, flow, graph, package, run_id): # Set some more internal state. self.flow = flow self.graph = graph self.package = package self.run_id = run_id def runtime_task_created( self, task_datastore, task_id, split_index, input_paths, is_cloned, ubf_context ): # To execute the Kubernetes job, the job container needs to have # access to the code package. We store the package in the datastore # which the pod is able to download as part of it's entrypoint. if not is_cloned: self._save_package_once(self.flow_datastore, self.package) def runtime_step_cli( self, cli_args, retry_count, max_user_code_retries, ubf_context ): if retry_count <= max_user_code_retries: # After all attempts to run the user code have failed, we don't need # to execute on Kubernetes anymore. We can execute possible fallback # code locally. cli_args.commands = ["kubernetes", "step"] cli_args.command_args.append(self.package_sha) cli_args.command_args.append(self.package_url) # --namespace is used to specify Metaflow namespace (a different # concept from k8s namespace). for k, v in self.attributes.items(): if k == "namespace": cli_args.command_options["k8s_namespace"] = v else: cli_args.command_options[k] = v cli_args.command_options["run-time-limit"] = self.run_time_limit cli_args.entrypoint[0] = sys.executable def task_pre_step( self, step_name, task_datastore, metadata, run_id, task_id, flow, graph, retry_count, max_retries, ubf_context, inputs, ): self.metadata = metadata self.task_datastore = task_datastore # task_pre_step may run locally if fallback is activated for @catch # decorator. In that scenario, we skip collecting Kubernetes execution # metadata. A rudimentary way to detect non-local execution is to # check for the existence of METAFLOW_KUBERNETES_WORKLOAD environment
meta["kubernetes-pod-name"] = os
# variable. if "METAFLOW_KUBERNETES_WORKLOAD" in os.environ: meta = {}
random_line_split
kubernetes_decorator.py
step to execute on Kubernetes regardless of the ```--with``` specification. To use, annotate your step as follows: ``` @kubernetes @step def my_step(self): ... ``` Parameters ---------- cpu : int Number of CPUs required for this step. Defaults to 1. If @resources is also present, the maximum value from all decorators is used memory : int Memory size (in MB) required for this step. Defaults to 4096. If @resources is also present, the maximum value from all decorators is used disk : int Disk size (in MB) required for this step. Defaults to 10GB. If @resources is also present, the maximum value from all decorators is used image : string Docker image to use when launching on Kubernetes. If not specified, a default docker image mapping to the current version of Python is used """ name = "kubernetes" defaults = { "cpu": "1", "memory": "4096", "disk": "10240", "image": None, "service_account": None, "secrets": None, # e.g., mysecret "node_selector": None, # e.g., kubernetes.io/os=linux "namespace": None, "gpu": None, # value of 0 implies that the scheduled node should not have GPUs "gpu_vendor": None, } package_url = None package_sha = None run_time_limit = None def __init__(self, attributes=None, statically_defined=False): super(KubernetesDecorator, self).__init__(attributes, statically_defined) if not self.attributes["namespace"]: self.attributes["namespace"] = KUBERNETES_NAMESPACE if not self.attributes["service_account"]: self.attributes["service_account"] = KUBERNETES_SERVICE_ACCOUNT if not self.attributes["gpu_vendor"]: self.attributes["gpu_vendor"] = KUBERNETES_GPU_VENDOR # TODO: Handle node_selector in a better manner. Currently it is special # cased in kubernetes_client.py # If no docker image is explicitly specified, impute a default image. if not self.attributes["image"]: # If metaflow-config specifies a docker image, just use that. if KUBERNETES_CONTAINER_IMAGE: self.attributes["image"] = KUBERNETES_CONTAINER_IMAGE # If metaflow-config doesn't specify a docker image, assign a # default docker image. else: # Default to vanilla Python image corresponding to major.minor # version of the Python interpreter launching the flow. self.attributes["image"] = "python:%s.%s" % ( platform.python_version_tuple()[0], platform.python_version_tuple()[1], ) # Assign docker registry URL for the image. if not get_docker_registry(self.attributes["image"]): if KUBERNETES_CONTAINER_REGISTRY: self.attributes["image"] = "%s/%s" % ( KUBERNETES_CONTAINER_REGISTRY.rstrip("/"), self.attributes["image"], ) # Refer https://github.com/Netflix/metaflow/blob/master/docs/lifecycle.png def
(self, flow, graph, step, decos, environment, flow_datastore, logger): # Executing Kubernetes jobs requires a non-local datastore. if flow_datastore.TYPE != "s3": raise KubernetesException( "The *@kubernetes* decorator requires --datastore=s3 at the moment." ) # Set internal state. self.logger = logger self.environment = environment self.step = step self.flow_datastore = flow_datastore if any([deco.name == "batch" for deco in decos]): raise MetaflowException( "Step *{step}* is marked for execution both on AWS Batch and " "Kubernetes. Please use one or the other.".format(step=step) ) for deco in decos: if getattr(deco, "IS_PARALLEL", False): raise KubernetesException( "@kubernetes does not support parallel execution currently." ) # Set run time limit for the Kubernetes job. self.run_time_limit = get_run_time_limit_for_task(decos) if self.run_time_limit < 60: raise KubernetesException( "The timeout for step *{step}* should be at least 60 seconds for " "execution on Kubernetes.".format(step=step) ) for deco in decos: if isinstance(deco, ResourcesDecorator): for k, v in deco.attributes.items(): # TODO: Special case GPUs when they are introduced in @resources. if k in self.attributes: if self.defaults[k] is None: # skip if expected value isn't an int/float continue # We use the larger of @resources and @batch attributes # TODO: Fix https://github.com/Netflix/metaflow/issues/467 my_val = self.attributes.get(k) if not (my_val is None and v is None): self.attributes[k] = str( max(float(my_val or 0), float(v or 0)) ) # Check GPU vendor. if self.attributes["gpu_vendor"].lower() not in ("amd", "nvidia"): raise KubernetesException( "GPU vendor *{}* for step *{step}* is not currently supported.".format( self.attributes["gpu_vendor"], step=step ) ) # CPU, Disk, and Memory values should be greater than 0. for attr in ["cpu", "disk", "memory"]: if not ( isinstance(self.attributes[attr], (int, unicode, basestring, float)) and float(self.attributes[attr]) > 0 ): raise KubernetesException( "Invalid {} value *{}* for step *{step}*; it should be greater than 0".format( attr, self.attributes[attr], step=step ) ) if self.attributes["gpu"] is not None and not ( isinstance(self.attributes["gpu"], (int, unicode, basestring)) and float(self.attributes["gpu"]).is_integer() ): raise KubernetesException( "Invalid GPU value *{}* for step *{step}*; it should be an integer".format( self.attributes["gpu"], step=step ) ) def package_init(self, flow, step_name, environment): try: # Kubernetes is a soft dependency. from kubernetes import client, config except (NameError, ImportError): raise KubernetesException( "Could not import module 'kubernetes'.\n\nInstall Kubernetes " "Python package (https://pypi.org/project/kubernetes/) first.\n" "You can install the module by executing - " "%s -m pip install kubernetes\n" "or equivalent through your favorite Python package manager." % sys.executable ) def runtime_init(self, flow, graph, package, run_id): # Set some more internal state. self.flow = flow self.graph = graph self.package = package self.run_id = run_id def runtime_task_created( self, task_datastore, task_id, split_index, input_paths, is_cloned, ubf_context ): # To execute the Kubernetes job, the job container needs to have # access to the code package. We store the package in the datastore # which the pod is able to download as part of it's entrypoint. if not is_cloned: self._save_package_once(self.flow_datastore, self.package) def runtime_step_cli( self, cli_args, retry_count, max_user_code_retries, ubf_context ): if retry_count <= max_user_code_retries: # After all attempts to run the user code have failed, we don't need # to execute on Kubernetes anymore. We can execute possible fallback # code locally. cli_args.commands = ["kubernetes", "step"] cli_args.command_args.append(self.package_sha) cli_args.command_args.append(self.package_url) # --namespace is used to specify Metaflow namespace (a different # concept from k8s namespace). for k, v in self.attributes.items(): if k == "namespace": cli_args.command_options["k8s_namespace"] = v else: cli_args.command_options[k] = v cli_args.command_options["run-time-limit"] = self.run_time_limit cli_args.entrypoint[0] = sys.executable def task_pre_step( self, step_name, task_datastore, metadata, run_id, task_id, flow, graph, retry_count, max_retries, ubf_context, inputs, ): self.metadata = metadata self.task_datastore = task_datastore # task_pre_step may run locally if fallback is activated for @catch # decorator. In that scenario, we skip collecting Kubernetes execution # metadata. A rudimentary way to detect non-local execution is to # check for the existence of METAFLOW_KUBERNETES_WORKLOAD environment # variable. if "METAFLOW_KUBERNETES_WORKLOAD" in os.environ: meta = {} meta["kubernetes-pod-name"] =
step_init
identifier_name
kubernetes_decorator.py
step to execute on Kubernetes regardless of the ```--with``` specification. To use, annotate your step as follows: ``` @kubernetes @step def my_step(self): ... ``` Parameters ---------- cpu : int Number of CPUs required for this step. Defaults to 1. If @resources is also present, the maximum value from all decorators is used memory : int Memory size (in MB) required for this step. Defaults to 4096. If @resources is also present, the maximum value from all decorators is used disk : int Disk size (in MB) required for this step. Defaults to 10GB. If @resources is also present, the maximum value from all decorators is used image : string Docker image to use when launching on Kubernetes. If not specified, a default docker image mapping to the current version of Python is used """ name = "kubernetes" defaults = { "cpu": "1", "memory": "4096", "disk": "10240", "image": None, "service_account": None, "secrets": None, # e.g., mysecret "node_selector": None, # e.g., kubernetes.io/os=linux "namespace": None, "gpu": None, # value of 0 implies that the scheduled node should not have GPUs "gpu_vendor": None, } package_url = None package_sha = None run_time_limit = None def __init__(self, attributes=None, statically_defined=False): super(KubernetesDecorator, self).__init__(attributes, statically_defined) if not self.attributes["namespace"]: self.attributes["namespace"] = KUBERNETES_NAMESPACE if not self.attributes["service_account"]: self.attributes["service_account"] = KUBERNETES_SERVICE_ACCOUNT if not self.attributes["gpu_vendor"]: self.attributes["gpu_vendor"] = KUBERNETES_GPU_VENDOR # TODO: Handle node_selector in a better manner. Currently it is special # cased in kubernetes_client.py # If no docker image is explicitly specified, impute a default image. if not self.attributes["image"]: # If metaflow-config specifies a docker image, just use that. if KUBERNETES_CONTAINER_IMAGE: self.attributes["image"] = KUBERNETES_CONTAINER_IMAGE # If metaflow-config doesn't specify a docker image, assign a # default docker image. else: # Default to vanilla Python image corresponding to major.minor # version of the Python interpreter launching the flow. self.attributes["image"] = "python:%s.%s" % ( platform.python_version_tuple()[0], platform.python_version_tuple()[1], ) # Assign docker registry URL for the image. if not get_docker_registry(self.attributes["image"]): if KUBERNETES_CONTAINER_REGISTRY: self.attributes["image"] = "%s/%s" % ( KUBERNETES_CONTAINER_REGISTRY.rstrip("/"), self.attributes["image"], ) # Refer https://github.com/Netflix/metaflow/blob/master/docs/lifecycle.png def step_init(self, flow, graph, step, decos, environment, flow_datastore, logger): # Executing Kubernetes jobs requires a non-local datastore. if flow_datastore.TYPE != "s3": raise KubernetesException( "The *@kubernetes* decorator requires --datastore=s3 at the moment." ) # Set internal state. self.logger = logger self.environment = environment self.step = step self.flow_datastore = flow_datastore if any([deco.name == "batch" for deco in decos]): raise MetaflowException( "Step *{step}* is marked for execution both on AWS Batch and " "Kubernetes. Please use one or the other.".format(step=step) ) for deco in decos: if getattr(deco, "IS_PARALLEL", False): raise KubernetesException( "@kubernetes does not support parallel execution currently." ) # Set run time limit for the Kubernetes job. self.run_time_limit = get_run_time_limit_for_task(decos) if self.run_time_limit < 60: raise KubernetesException( "The timeout for step *{step}* should be at least 60 seconds for " "execution on Kubernetes.".format(step=step) ) for deco in decos: if isinstance(deco, ResourcesDecorator): for k, v in deco.attributes.items(): # TODO: Special case GPUs when they are introduced in @resources. if k in self.attributes: if self.defaults[k] is None: # skip if expected value isn't an int/float continue # We use the larger of @resources and @batch attributes # TODO: Fix https://github.com/Netflix/metaflow/issues/467 my_val = self.attributes.get(k) if not (my_val is None and v is None): self.attributes[k] = str( max(float(my_val or 0), float(v or 0)) ) # Check GPU vendor. if self.attributes["gpu_vendor"].lower() not in ("amd", "nvidia"): raise KubernetesException( "GPU vendor *{}* for step *{step}* is not currently supported.".format( self.attributes["gpu_vendor"], step=step ) ) # CPU, Disk, and Memory values should be greater than 0. for attr in ["cpu", "disk", "memory"]: if not ( isinstance(self.attributes[attr], (int, unicode, basestring, float)) and float(self.attributes[attr]) > 0 ): raise KubernetesException( "Invalid {} value *{}* for step *{step}*; it should be greater than 0".format( attr, self.attributes[attr], step=step ) ) if self.attributes["gpu"] is not None and not ( isinstance(self.attributes["gpu"], (int, unicode, basestring)) and float(self.attributes["gpu"]).is_integer() ): raise KubernetesException( "Invalid GPU value *{}* for step *{step}*; it should be an integer".format( self.attributes["gpu"], step=step ) ) def package_init(self, flow, step_name, environment): try: # Kubernetes is a soft dependency. from kubernetes import client, config except (NameError, ImportError): raise KubernetesException( "Could not import module 'kubernetes'.\n\nInstall Kubernetes " "Python package (https://pypi.org/project/kubernetes/) first.\n" "You can install the module by executing - " "%s -m pip install kubernetes\n" "or equivalent through your favorite Python package manager." % sys.executable ) def runtime_init(self, flow, graph, package, run_id): # Set some more internal state. self.flow = flow self.graph = graph self.package = package self.run_id = run_id def runtime_task_created( self, task_datastore, task_id, split_index, input_paths, is_cloned, ubf_context ): # To execute the Kubernetes job, the job container needs to have # access to the code package. We store the package in the datastore # which the pod is able to download as part of it's entrypoint.
def runtime_step_cli( self, cli_args, retry_count, max_user_code_retries, ubf_context ): if retry_count <= max_user_code_retries: # After all attempts to run the user code have failed, we don't need # to execute on Kubernetes anymore. We can execute possible fallback # code locally. cli_args.commands = ["kubernetes", "step"] cli_args.command_args.append(self.package_sha) cli_args.command_args.append(self.package_url) # --namespace is used to specify Metaflow namespace (a different # concept from k8s namespace). for k, v in self.attributes.items(): if k == "namespace": cli_args.command_options["k8s_namespace"] = v else: cli_args.command_options[k] = v cli_args.command_options["run-time-limit"] = self.run_time_limit cli_args.entrypoint[0] = sys.executable def task_pre_step( self, step_name, task_datastore, metadata, run_id, task_id, flow, graph, retry_count, max_retries, ubf_context, inputs, ): self.metadata = metadata self.task_datastore = task_datastore # task_pre_step may run locally if fallback is activated for @catch # decorator. In that scenario, we skip collecting Kubernetes execution # metadata. A rudimentary way to detect non-local execution is to # check for the existence of METAFLOW_KUBERNETES_WORKLOAD environment # variable. if "METAFLOW_KUBERNETES_WORKLOAD" in os.environ: meta = {} meta["kubernetes-pod-name"] =
if not is_cloned: self._save_package_once(self.flow_datastore, self.package)
identifier_body
account_control.js
License for the specific language governing permissions and * limitations under the License. */ import("stringutils"); import("stringutils.*"); import("funhtml.*"); import("email.sendEmail"); import("cache_utils.syncedWithCache"); import("etherpad.helpers"); import("etherpad.utils.*"); import("etherpad.sessions.getSession"); import("etherpad.pro.pro_accounts"); import("etherpad.pro.pro_accounts.getSessionProAccount"); import("etherpad.pro.domains"); import("etherpad.pro.pro_utils"); import("etherpad.pro.pro_account_auto_signin"); import("etherpad.pro.pro_config"); import("etherpad.pad.pad_security"); import("etherpad.pad.padutils"); import("etherpad.pad.padusers"); import("etherpad.collab.collab_server"); function onRequest() { if (!getSession().tempFormData) { getSession().tempFormData = {}; } return false; // path not handled here } //-------------------------------------------------------------------------------- // helpers //-------------------------------------------------------------------------------- function _redirOnError(m, clearQuery) { if (m) { getSession().accountFormError = m; var dest = request.url; if (clearQuery) { dest = request.path; } response.redirect(dest); } } function setSigninNotice(m) { getSession().accountSigninNotice = m; } function setSessionError(m) { getSession().accountFormError = m; } function _topDiv(id, name) { var m = getSession()[name]; if (m) { delete getSession()[name]; return DIV({id: id}, m); } else { return ''; } } function _messageDiv() { return _topDiv('account-message', 'accountMessage'); } function _errorDiv() { return _topDiv('account-error', 'accountFormError'); } function
() { return _topDiv('signin-notice', 'accountSigninNotice'); } function _renderTemplate(name, data) { data.messageDiv = _messageDiv; data.errorDiv = _errorDiv; data.signinNotice = _signinNoticeDiv; data.tempFormData = getSession().tempFormData; renderFramed('pro/account/'+name+'.ejs', data); } //---------------------------------------------------------------- // /ep/account/ //---------------------------------------------------------------- function render_main_get() { _renderTemplate('my-account', { account: getSessionProAccount(), changePass: getSession().changePass }); } function render_update_info_get() { response.redirect('/ep/account/'); } function render_update_info_post() { var fullName = request.params.fullName; var email = trim(request.params.email); getSession().tempFormData.email = email; getSession().tempFormData.fullName = fullName; _redirOnError(pro_accounts.validateEmail(email)); _redirOnError(pro_accounts.validateFullName(fullName)); pro_accounts.setEmail(getSessionProAccount(), email); pro_accounts.setFullName(getSessionProAccount(), fullName); getSession().accountMessage = "Info updated."; response.redirect('/ep/account/'); } function render_update_password_get() { response.redirect('/ep/account/'); } function render_update_password_post() { var password = request.params.password; var passwordConfirm = request.params.passwordConfirm; if (password != passwordConfirm) { _redirOnError('Passwords did not match.'); } _redirOnError(pro_accounts.validatePassword(password)); pro_accounts.setPassword(getSessionProAccount(), password); if (getSession().changePass) { delete getSession().changePass; response.redirect('/'); } getSession().accountMessage = "Password updated."; response.redirect('/ep/account/'); } //-------------------------------------------------------------------------------- // signin/signout //-------------------------------------------------------------------------------- function render_sign_in_get() { if (request.params.uid && request.params.tp) { var m = pro_accounts.authenticateTempSignIn(Number(request.params.uid), request.params.tp); if (m) { getSession().accountFormError = m; response.redirect('/ep/account/'); } } if (request.params.instantSigninKey) { _attemptInstantSignin(request.params.instantSigninKey); } if (getSession().recentlySignedOut && getSession().accountFormError) { delete getSession().accountFormError; delete getSession().recentlySignedOut; } // Note: must check isAccountSignedIn before calling checkAutoSignin()! if (pro_accounts.isAccountSignedIn()) { _redirectToPostSigninDestination(); } pro_account_auto_signin.checkAutoSignin(); var domainRecord = domains.getRequestDomainRecord(); var showGuestBox = false; if (request.params.guest && request.params.padId) { showGuestBox = true; } _renderTemplate('signin', { domain: pro_utils.getFullProDomain(), siteName: toHTML(pro_config.getConfig().siteName), email: getSession().tempFormData.email || "", password: getSession().tempFormData.password || "", rememberMe: getSession().tempFormData.rememberMe || false, showGuestBox: showGuestBox, localPadId: request.params.padId }); } function _attemptInstantSignin(key) { // See src/etherpad/control/global_pro_account_control.js var email = null; var password = null; syncedWithCache('global_signin_passwords', function(c) { if (c[key]) { email = c[key].email; password = c[key].password; } delete c[key]; }); getSession().tempFormData.email = email; _redirOnError(pro_accounts.authenticateSignIn(email, password), true); } function render_sign_in_post() { var email = trim(request.params.email); var password = request.params.password; getSession().tempFormData.email = email; getSession().tempFormData.rememberMe = request.params.rememberMe; _redirOnError(pro_accounts.authenticateSignIn(email, password)); pro_account_auto_signin.setAutoSigninCookie(request.params.rememberMe); _redirectToPostSigninDestination(); } function render_guest_sign_in_get() { var localPadId = request.params.padId; var domainId = domains.getRequestDomainId(); var globalPadId = padutils.makeGlobalId(domainId, localPadId); var userId = padusers.getUserId(); pro_account_auto_signin.checkAutoSignin(); pad_security.clearKnockStatus(userId, globalPadId); _renderTemplate('signin-guest', { localPadId: localPadId, errorMessage: getSession().guestAccessError, siteName: toHTML(pro_config.getConfig().siteName), guestName: padusers.getUserName() || "" }); } function render_guest_sign_in_post() { function _err(m) { if (m) { getSession().guestAccessError = m; response.redirect(request.url); } } var displayName = request.params.guestDisplayName; var localPadId = request.params.localPadId; if (!(displayName && displayName.length > 0)) { _err("Please enter a display name"); } getSession().guestDisplayName = displayName; response.redirect('/ep/account/guest-knock?padId='+encodeURIComponent(localPadId)+ "&guestDisplayName="+encodeURIComponent(displayName)); } function render_guest_knock_get() { var localPadId = request.params.padId; helpers.addClientVars({ localPadId: localPadId, guestDisplayName: request.params.guestDisplayName, padUrl: "http://"+httpHost(request.host)+"/"+localPadId }); _renderTemplate('guest-knock', {}); } function render_guest_knock_post() { var localPadId = request.params.padId; var displayName = request.params.guestDisplayName; var domainId = domains.getRequestDomainId(); var globalPadId = padutils.makeGlobalId(domainId, localPadId); var userId = padusers.getUserId(); response.setContentType("text/plain; charset=utf-8"); // has the knock already been answsered? var currentAnswer = pad_security.getKnockAnswer(userId, globalPadId); if (currentAnswer) { response.write(currentAnswer); } else { collab_server.guestKnock(globalPadId, userId, displayName); response.write("wait"); } } function _redirectToPostSigninDestination() { var cont = request.params.cont; if (!cont) { cont = '/'; } response.redirect(cont); } function render_sign_out() { pro_account_auto_signin.setAutoSigninCookie(false); pro_accounts.signOut(); delete getSession().padPasswordAuth; getSession().recentlySignedOut = true; response.redirect("/"); } //-------------------------------------------------------------------------------- // create-admin-account (eepnet only) //-------------------------------------------------------------------------------- function render_create_admin_account_get() { if (pro_accounts.doesAdminExist()) { renderFramedError("An admin account already exists on this domain."); response.stop(); } _renderTemplate('create-admin-account', {}); } function render_create_admin_account_post() { var email = trim(request.params.email); var password = request.params.password; var passwordConfirm = request.params.passwordConfirm; var fullName = request.params.fullName; getSession().tempFormData.email = email; getSession().tempFormData.fullName = fullName; if (password != passwordConfirm) { _redirOnError('Passwords did not match.'); } _redirOnError(pro_accounts.validateEmail(email)); _redirOnError(pro_accounts.validateFullName(fullName)); _redirOnError(pro_accounts.validatePassword(password)); pro_accounts.createNewAccount(null, fullName, email, password, true); var u = pro_accounts.getAccountByEmail(email, null); // TODO: should we send a welcome email here? //pro_accounts.sendWelcomeEmail(u);
_signinNoticeDiv
identifier_name
account_control.js
License for the specific language governing permissions and * limitations under the License. */ import("stringutils"); import("stringutils.*"); import("funhtml.*"); import("email.sendEmail"); import("cache_utils.syncedWithCache"); import("etherpad.helpers"); import("etherpad.utils.*"); import("etherpad.sessions.getSession"); import("etherpad.pro.pro_accounts"); import("etherpad.pro.pro_accounts.getSessionProAccount"); import("etherpad.pro.domains"); import("etherpad.pro.pro_utils"); import("etherpad.pro.pro_account_auto_signin"); import("etherpad.pro.pro_config"); import("etherpad.pad.pad_security"); import("etherpad.pad.padutils"); import("etherpad.pad.padusers"); import("etherpad.collab.collab_server"); function onRequest() { if (!getSession().tempFormData) { getSession().tempFormData = {}; } return false; // path not handled here } //-------------------------------------------------------------------------------- // helpers //-------------------------------------------------------------------------------- function _redirOnError(m, clearQuery) { if (m) { getSession().accountFormError = m; var dest = request.url; if (clearQuery) { dest = request.path; } response.redirect(dest); } } function setSigninNotice(m) { getSession().accountSigninNotice = m; } function setSessionError(m) { getSession().accountFormError = m; } function _topDiv(id, name) { var m = getSession()[name]; if (m) { delete getSession()[name]; return DIV({id: id}, m); } else { return ''; } } function _messageDiv() { return _topDiv('account-message', 'accountMessage'); } function _errorDiv() { return _topDiv('account-error', 'accountFormError'); } function _signinNoticeDiv() { return _topDiv('signin-notice', 'accountSigninNotice'); } function _renderTemplate(name, data) { data.messageDiv = _messageDiv; data.errorDiv = _errorDiv; data.signinNotice = _signinNoticeDiv; data.tempFormData = getSession().tempFormData; renderFramed('pro/account/'+name+'.ejs', data); } //---------------------------------------------------------------- // /ep/account/ //---------------------------------------------------------------- function render_main_get() { _renderTemplate('my-account', { account: getSessionProAccount(), changePass: getSession().changePass }); } function render_update_info_get() { response.redirect('/ep/account/'); } function render_update_info_post() { var fullName = request.params.fullName; var email = trim(request.params.email); getSession().tempFormData.email = email; getSession().tempFormData.fullName = fullName; _redirOnError(pro_accounts.validateEmail(email)); _redirOnError(pro_accounts.validateFullName(fullName)); pro_accounts.setEmail(getSessionProAccount(), email); pro_accounts.setFullName(getSessionProAccount(), fullName); getSession().accountMessage = "Info updated."; response.redirect('/ep/account/'); } function render_update_password_get() { response.redirect('/ep/account/'); } function render_update_password_post() { var password = request.params.password; var passwordConfirm = request.params.passwordConfirm; if (password != passwordConfirm) { _redirOnError('Passwords did not match.'); } _redirOnError(pro_accounts.validatePassword(password)); pro_accounts.setPassword(getSessionProAccount(), password); if (getSession().changePass) { delete getSession().changePass; response.redirect('/'); } getSession().accountMessage = "Password updated."; response.redirect('/ep/account/'); } //-------------------------------------------------------------------------------- // signin/signout //-------------------------------------------------------------------------------- function render_sign_in_get() { if (request.params.uid && request.params.tp) { var m = pro_accounts.authenticateTempSignIn(Number(request.params.uid), request.params.tp); if (m) { getSession().accountFormError = m; response.redirect('/ep/account/'); } } if (request.params.instantSigninKey) { _attemptInstantSignin(request.params.instantSigninKey); } if (getSession().recentlySignedOut && getSession().accountFormError) { delete getSession().accountFormError; delete getSession().recentlySignedOut; } // Note: must check isAccountSignedIn before calling checkAutoSignin()! if (pro_accounts.isAccountSignedIn()) { _redirectToPostSigninDestination(); } pro_account_auto_signin.checkAutoSignin(); var domainRecord = domains.getRequestDomainRecord(); var showGuestBox = false; if (request.params.guest && request.params.padId) { showGuestBox = true; } _renderTemplate('signin', { domain: pro_utils.getFullProDomain(), siteName: toHTML(pro_config.getConfig().siteName), email: getSession().tempFormData.email || "", password: getSession().tempFormData.password || "", rememberMe: getSession().tempFormData.rememberMe || false, showGuestBox: showGuestBox, localPadId: request.params.padId }); } function _attemptInstantSignin(key) { // See src/etherpad/control/global_pro_account_control.js var email = null; var password = null; syncedWithCache('global_signin_passwords', function(c) { if (c[key]) { email = c[key].email; password = c[key].password; } delete c[key]; }); getSession().tempFormData.email = email; _redirOnError(pro_accounts.authenticateSignIn(email, password), true); } function render_sign_in_post() { var email = trim(request.params.email); var password = request.params.password; getSession().tempFormData.email = email; getSession().tempFormData.rememberMe = request.params.rememberMe; _redirOnError(pro_accounts.authenticateSignIn(email, password)); pro_account_auto_signin.setAutoSigninCookie(request.params.rememberMe); _redirectToPostSigninDestination(); } function render_guest_sign_in_get()
function render_guest_sign_in_post() { function _err(m) { if (m) { getSession().guestAccessError = m; response.redirect(request.url); } } var displayName = request.params.guestDisplayName; var localPadId = request.params.localPadId; if (!(displayName && displayName.length > 0)) { _err("Please enter a display name"); } getSession().guestDisplayName = displayName; response.redirect('/ep/account/guest-knock?padId='+encodeURIComponent(localPadId)+ "&guestDisplayName="+encodeURIComponent(displayName)); } function render_guest_knock_get() { var localPadId = request.params.padId; helpers.addClientVars({ localPadId: localPadId, guestDisplayName: request.params.guestDisplayName, padUrl: "http://"+httpHost(request.host)+"/"+localPadId }); _renderTemplate('guest-knock', {}); } function render_guest_knock_post() { var localPadId = request.params.padId; var displayName = request.params.guestDisplayName; var domainId = domains.getRequestDomainId(); var globalPadId = padutils.makeGlobalId(domainId, localPadId); var userId = padusers.getUserId(); response.setContentType("text/plain; charset=utf-8"); // has the knock already been answsered? var currentAnswer = pad_security.getKnockAnswer(userId, globalPadId); if (currentAnswer) { response.write(currentAnswer); } else { collab_server.guestKnock(globalPadId, userId, displayName); response.write("wait"); } } function _redirectToPostSigninDestination() { var cont = request.params.cont; if (!cont) { cont = '/'; } response.redirect(cont); } function render_sign_out() { pro_account_auto_signin.setAutoSigninCookie(false); pro_accounts.signOut(); delete getSession().padPasswordAuth; getSession().recentlySignedOut = true; response.redirect("/"); } //-------------------------------------------------------------------------------- // create-admin-account (eepnet only) //-------------------------------------------------------------------------------- function render_create_admin_account_get() { if (pro_accounts.doesAdminExist()) { renderFramedError("An admin account already exists on this domain."); response.stop(); } _renderTemplate('create-admin-account', {}); } function render_create_admin_account_post() { var email = trim(request.params.email); var password = request.params.password; var passwordConfirm = request.params.passwordConfirm; var fullName = request.params.fullName; getSession().tempFormData.email = email; getSession().tempFormData.fullName = fullName; if (password != passwordConfirm) { _redirOnError('Passwords did not match.'); } _redirOnError(pro_accounts.validateEmail(email)); _redirOnError(pro_accounts.validateFullName(fullName)); _redirOnError(pro_accounts.validatePassword(password)); pro_accounts.createNewAccount(null, fullName, email, password, true); var u = pro_accounts.getAccountByEmail(email, null); // TODO: should we send a welcome email here? //pro_accounts.sendWelcomeEmail(u);
{ var localPadId = request.params.padId; var domainId = domains.getRequestDomainId(); var globalPadId = padutils.makeGlobalId(domainId, localPadId); var userId = padusers.getUserId(); pro_account_auto_signin.checkAutoSignin(); pad_security.clearKnockStatus(userId, globalPadId); _renderTemplate('signin-guest', { localPadId: localPadId, errorMessage: getSession().guestAccessError, siteName: toHTML(pro_config.getConfig().siteName), guestName: padusers.getUserName() || "" }); }
identifier_body
account_control.js
License for the specific language governing permissions and * limitations under the License. */ import("stringutils"); import("stringutils.*"); import("funhtml.*"); import("email.sendEmail"); import("cache_utils.syncedWithCache"); import("etherpad.helpers"); import("etherpad.utils.*"); import("etherpad.sessions.getSession"); import("etherpad.pro.pro_accounts"); import("etherpad.pro.pro_accounts.getSessionProAccount"); import("etherpad.pro.domains"); import("etherpad.pro.pro_utils"); import("etherpad.pro.pro_account_auto_signin"); import("etherpad.pro.pro_config"); import("etherpad.pad.pad_security"); import("etherpad.pad.padutils"); import("etherpad.pad.padusers"); import("etherpad.collab.collab_server"); function onRequest() { if (!getSession().tempFormData) { getSession().tempFormData = {}; } return false; // path not handled here } //-------------------------------------------------------------------------------- // helpers //-------------------------------------------------------------------------------- function _redirOnError(m, clearQuery) { if (m) { getSession().accountFormError = m; var dest = request.url; if (clearQuery) { dest = request.path; } response.redirect(dest); } } function setSigninNotice(m) { getSession().accountSigninNotice = m; } function setSessionError(m) { getSession().accountFormError = m; } function _topDiv(id, name) { var m = getSession()[name]; if (m) { delete getSession()[name]; return DIV({id: id}, m); } else { return ''; } } function _messageDiv() { return _topDiv('account-message', 'accountMessage'); } function _errorDiv() { return _topDiv('account-error', 'accountFormError'); } function _signinNoticeDiv() { return _topDiv('signin-notice', 'accountSigninNotice'); } function _renderTemplate(name, data) { data.messageDiv = _messageDiv; data.errorDiv = _errorDiv; data.signinNotice = _signinNoticeDiv; data.tempFormData = getSession().tempFormData; renderFramed('pro/account/'+name+'.ejs', data); } //---------------------------------------------------------------- // /ep/account/ //---------------------------------------------------------------- function render_main_get() { _renderTemplate('my-account', { account: getSessionProAccount(), changePass: getSession().changePass }); } function render_update_info_get() { response.redirect('/ep/account/'); } function render_update_info_post() { var fullName = request.params.fullName; var email = trim(request.params.email); getSession().tempFormData.email = email; getSession().tempFormData.fullName = fullName; _redirOnError(pro_accounts.validateEmail(email)); _redirOnError(pro_accounts.validateFullName(fullName)); pro_accounts.setEmail(getSessionProAccount(), email); pro_accounts.setFullName(getSessionProAccount(), fullName); getSession().accountMessage = "Info updated."; response.redirect('/ep/account/'); } function render_update_password_get() { response.redirect('/ep/account/'); } function render_update_password_post() { var password = request.params.password; var passwordConfirm = request.params.passwordConfirm; if (password != passwordConfirm) { _redirOnError('Passwords did not match.'); } _redirOnError(pro_accounts.validatePassword(password)); pro_accounts.setPassword(getSessionProAccount(), password); if (getSession().changePass) { delete getSession().changePass; response.redirect('/'); } getSession().accountMessage = "Password updated."; response.redirect('/ep/account/'); } //-------------------------------------------------------------------------------- // signin/signout //-------------------------------------------------------------------------------- function render_sign_in_get() { if (request.params.uid && request.params.tp) { var m = pro_accounts.authenticateTempSignIn(Number(request.params.uid), request.params.tp); if (m) { getSession().accountFormError = m; response.redirect('/ep/account/'); } } if (request.params.instantSigninKey) { _attemptInstantSignin(request.params.instantSigninKey); } if (getSession().recentlySignedOut && getSession().accountFormError) { delete getSession().accountFormError; delete getSession().recentlySignedOut; } // Note: must check isAccountSignedIn before calling checkAutoSignin()! if (pro_accounts.isAccountSignedIn()) { _redirectToPostSigninDestination(); } pro_account_auto_signin.checkAutoSignin(); var domainRecord = domains.getRequestDomainRecord(); var showGuestBox = false; if (request.params.guest && request.params.padId) { showGuestBox = true; } _renderTemplate('signin', { domain: pro_utils.getFullProDomain(), siteName: toHTML(pro_config.getConfig().siteName), email: getSession().tempFormData.email || "", password: getSession().tempFormData.password || "", rememberMe: getSession().tempFormData.rememberMe || false, showGuestBox: showGuestBox, localPadId: request.params.padId }); } function _attemptInstantSignin(key) { // See src/etherpad/control/global_pro_account_control.js var email = null; var password = null; syncedWithCache('global_signin_passwords', function(c) { if (c[key]) { email = c[key].email; password = c[key].password; } delete c[key]; }); getSession().tempFormData.email = email; _redirOnError(pro_accounts.authenticateSignIn(email, password), true); } function render_sign_in_post() { var email = trim(request.params.email); var password = request.params.password; getSession().tempFormData.email = email; getSession().tempFormData.rememberMe = request.params.rememberMe; _redirOnError(pro_accounts.authenticateSignIn(email, password)); pro_account_auto_signin.setAutoSigninCookie(request.params.rememberMe); _redirectToPostSigninDestination(); } function render_guest_sign_in_get() { var localPadId = request.params.padId; var domainId = domains.getRequestDomainId(); var globalPadId = padutils.makeGlobalId(domainId, localPadId); var userId = padusers.getUserId(); pro_account_auto_signin.checkAutoSignin(); pad_security.clearKnockStatus(userId, globalPadId); _renderTemplate('signin-guest', { localPadId: localPadId, errorMessage: getSession().guestAccessError, siteName: toHTML(pro_config.getConfig().siteName), guestName: padusers.getUserName() || "" }); } function render_guest_sign_in_post() { function _err(m) { if (m) { getSession().guestAccessError = m; response.redirect(request.url); } } var displayName = request.params.guestDisplayName; var localPadId = request.params.localPadId; if (!(displayName && displayName.length > 0)) { _err("Please enter a display name"); } getSession().guestDisplayName = displayName; response.redirect('/ep/account/guest-knock?padId='+encodeURIComponent(localPadId)+ "&guestDisplayName="+encodeURIComponent(displayName)); } function render_guest_knock_get() { var localPadId = request.params.padId; helpers.addClientVars({ localPadId: localPadId, guestDisplayName: request.params.guestDisplayName, padUrl: "http://"+httpHost(request.host)+"/"+localPadId }); _renderTemplate('guest-knock', {}); } function render_guest_knock_post() { var localPadId = request.params.padId; var displayName = request.params.guestDisplayName; var domainId = domains.getRequestDomainId(); var globalPadId = padutils.makeGlobalId(domainId, localPadId); var userId = padusers.getUserId(); response.setContentType("text/plain; charset=utf-8"); // has the knock already been answsered? var currentAnswer = pad_security.getKnockAnswer(userId, globalPadId); if (currentAnswer) { response.write(currentAnswer); } else { collab_server.guestKnock(globalPadId, userId, displayName); response.write("wait"); } } function _redirectToPostSigninDestination() { var cont = request.params.cont; if (!cont) { cont = '/'; } response.redirect(cont); } function render_sign_out() { pro_account_auto_signin.setAutoSigninCookie(false); pro_accounts.signOut(); delete getSession().padPasswordAuth; getSession().recentlySignedOut = true; response.redirect("/"); } //-------------------------------------------------------------------------------- // create-admin-account (eepnet only) //-------------------------------------------------------------------------------- function render_create_admin_account_get() { if (pro_accounts.doesAdminExist()) { renderFramedError("An admin account already exists on this domain."); response.stop(); } _renderTemplate('create-admin-account', {}); } function render_create_admin_account_post() { var email = trim(request.params.email); var password = request.params.password; var passwordConfirm = request.params.passwordConfirm; var fullName = request.params.fullName;
if (password != passwordConfirm) { _redirOnError('Passwords did not match.'); } _redirOnError(pro_accounts.validateEmail(email)); _redirOnError(pro_accounts.validateFullName(fullName)); _redirOnError(pro_accounts.validatePassword(password)); pro_accounts.createNewAccount(null, fullName, email, password, true); var u = pro_accounts.getAccountByEmail(email, null); // TODO: should we send a welcome email here? //pro_accounts.sendWelcomeEmail(u);
getSession().tempFormData.email = email; getSession().tempFormData.fullName = fullName;
random_line_split
account_control.js
License for the specific language governing permissions and * limitations under the License. */ import("stringutils"); import("stringutils.*"); import("funhtml.*"); import("email.sendEmail"); import("cache_utils.syncedWithCache"); import("etherpad.helpers"); import("etherpad.utils.*"); import("etherpad.sessions.getSession"); import("etherpad.pro.pro_accounts"); import("etherpad.pro.pro_accounts.getSessionProAccount"); import("etherpad.pro.domains"); import("etherpad.pro.pro_utils"); import("etherpad.pro.pro_account_auto_signin"); import("etherpad.pro.pro_config"); import("etherpad.pad.pad_security"); import("etherpad.pad.padutils"); import("etherpad.pad.padusers"); import("etherpad.collab.collab_server"); function onRequest() { if (!getSession().tempFormData) { getSession().tempFormData = {}; } return false; // path not handled here } //-------------------------------------------------------------------------------- // helpers //-------------------------------------------------------------------------------- function _redirOnError(m, clearQuery) { if (m) { getSession().accountFormError = m; var dest = request.url; if (clearQuery) { dest = request.path; } response.redirect(dest); } } function setSigninNotice(m) { getSession().accountSigninNotice = m; } function setSessionError(m) { getSession().accountFormError = m; } function _topDiv(id, name) { var m = getSession()[name]; if (m) { delete getSession()[name]; return DIV({id: id}, m); } else { return ''; } } function _messageDiv() { return _topDiv('account-message', 'accountMessage'); } function _errorDiv() { return _topDiv('account-error', 'accountFormError'); } function _signinNoticeDiv() { return _topDiv('signin-notice', 'accountSigninNotice'); } function _renderTemplate(name, data) { data.messageDiv = _messageDiv; data.errorDiv = _errorDiv; data.signinNotice = _signinNoticeDiv; data.tempFormData = getSession().tempFormData; renderFramed('pro/account/'+name+'.ejs', data); } //---------------------------------------------------------------- // /ep/account/ //---------------------------------------------------------------- function render_main_get() { _renderTemplate('my-account', { account: getSessionProAccount(), changePass: getSession().changePass }); } function render_update_info_get() { response.redirect('/ep/account/'); } function render_update_info_post() { var fullName = request.params.fullName; var email = trim(request.params.email); getSession().tempFormData.email = email; getSession().tempFormData.fullName = fullName; _redirOnError(pro_accounts.validateEmail(email)); _redirOnError(pro_accounts.validateFullName(fullName)); pro_accounts.setEmail(getSessionProAccount(), email); pro_accounts.setFullName(getSessionProAccount(), fullName); getSession().accountMessage = "Info updated."; response.redirect('/ep/account/'); } function render_update_password_get() { response.redirect('/ep/account/'); } function render_update_password_post() { var password = request.params.password; var passwordConfirm = request.params.passwordConfirm; if (password != passwordConfirm) { _redirOnError('Passwords did not match.'); } _redirOnError(pro_accounts.validatePassword(password)); pro_accounts.setPassword(getSessionProAccount(), password); if (getSession().changePass) { delete getSession().changePass; response.redirect('/'); } getSession().accountMessage = "Password updated."; response.redirect('/ep/account/'); } //-------------------------------------------------------------------------------- // signin/signout //-------------------------------------------------------------------------------- function render_sign_in_get() { if (request.params.uid && request.params.tp) { var m = pro_accounts.authenticateTempSignIn(Number(request.params.uid), request.params.tp); if (m) { getSession().accountFormError = m; response.redirect('/ep/account/'); } } if (request.params.instantSigninKey) { _attemptInstantSignin(request.params.instantSigninKey); } if (getSession().recentlySignedOut && getSession().accountFormError) { delete getSession().accountFormError; delete getSession().recentlySignedOut; } // Note: must check isAccountSignedIn before calling checkAutoSignin()! if (pro_accounts.isAccountSignedIn()) { _redirectToPostSigninDestination(); } pro_account_auto_signin.checkAutoSignin(); var domainRecord = domains.getRequestDomainRecord(); var showGuestBox = false; if (request.params.guest && request.params.padId)
_renderTemplate('signin', { domain: pro_utils.getFullProDomain(), siteName: toHTML(pro_config.getConfig().siteName), email: getSession().tempFormData.email || "", password: getSession().tempFormData.password || "", rememberMe: getSession().tempFormData.rememberMe || false, showGuestBox: showGuestBox, localPadId: request.params.padId }); } function _attemptInstantSignin(key) { // See src/etherpad/control/global_pro_account_control.js var email = null; var password = null; syncedWithCache('global_signin_passwords', function(c) { if (c[key]) { email = c[key].email; password = c[key].password; } delete c[key]; }); getSession().tempFormData.email = email; _redirOnError(pro_accounts.authenticateSignIn(email, password), true); } function render_sign_in_post() { var email = trim(request.params.email); var password = request.params.password; getSession().tempFormData.email = email; getSession().tempFormData.rememberMe = request.params.rememberMe; _redirOnError(pro_accounts.authenticateSignIn(email, password)); pro_account_auto_signin.setAutoSigninCookie(request.params.rememberMe); _redirectToPostSigninDestination(); } function render_guest_sign_in_get() { var localPadId = request.params.padId; var domainId = domains.getRequestDomainId(); var globalPadId = padutils.makeGlobalId(domainId, localPadId); var userId = padusers.getUserId(); pro_account_auto_signin.checkAutoSignin(); pad_security.clearKnockStatus(userId, globalPadId); _renderTemplate('signin-guest', { localPadId: localPadId, errorMessage: getSession().guestAccessError, siteName: toHTML(pro_config.getConfig().siteName), guestName: padusers.getUserName() || "" }); } function render_guest_sign_in_post() { function _err(m) { if (m) { getSession().guestAccessError = m; response.redirect(request.url); } } var displayName = request.params.guestDisplayName; var localPadId = request.params.localPadId; if (!(displayName && displayName.length > 0)) { _err("Please enter a display name"); } getSession().guestDisplayName = displayName; response.redirect('/ep/account/guest-knock?padId='+encodeURIComponent(localPadId)+ "&guestDisplayName="+encodeURIComponent(displayName)); } function render_guest_knock_get() { var localPadId = request.params.padId; helpers.addClientVars({ localPadId: localPadId, guestDisplayName: request.params.guestDisplayName, padUrl: "http://"+httpHost(request.host)+"/"+localPadId }); _renderTemplate('guest-knock', {}); } function render_guest_knock_post() { var localPadId = request.params.padId; var displayName = request.params.guestDisplayName; var domainId = domains.getRequestDomainId(); var globalPadId = padutils.makeGlobalId(domainId, localPadId); var userId = padusers.getUserId(); response.setContentType("text/plain; charset=utf-8"); // has the knock already been answsered? var currentAnswer = pad_security.getKnockAnswer(userId, globalPadId); if (currentAnswer) { response.write(currentAnswer); } else { collab_server.guestKnock(globalPadId, userId, displayName); response.write("wait"); } } function _redirectToPostSigninDestination() { var cont = request.params.cont; if (!cont) { cont = '/'; } response.redirect(cont); } function render_sign_out() { pro_account_auto_signin.setAutoSigninCookie(false); pro_accounts.signOut(); delete getSession().padPasswordAuth; getSession().recentlySignedOut = true; response.redirect("/"); } //-------------------------------------------------------------------------------- // create-admin-account (eepnet only) //-------------------------------------------------------------------------------- function render_create_admin_account_get() { if (pro_accounts.doesAdminExist()) { renderFramedError("An admin account already exists on this domain."); response.stop(); } _renderTemplate('create-admin-account', {}); } function render_create_admin_account_post() { var email = trim(request.params.email); var password = request.params.password; var passwordConfirm = request.params.passwordConfirm; var fullName = request.params.fullName; getSession().tempFormData.email = email; getSession().tempFormData.fullName = fullName; if (password != passwordConfirm) { _redirOnError('Passwords did not match.'); } _redirOnError(pro_accounts.validateEmail(email)); _redirOnError(pro_accounts.validateFullName(fullName)); _redirOnError(pro_accounts.validatePassword(password)); pro_accounts.createNewAccount(null, fullName, email, password, true); var u = pro_accounts.getAccountByEmail(email, null); // TODO: should we send a welcome email here? //pro_accounts.sendWelcomeEmail(u);
{ showGuestBox = true; }
conditional_block
gui_debugging.py
deltaP = [mean_power - P for P in trap_powers] dmags = [(dP / abs(dP)) * sqrt(abs(dP)) * L for dP in deltaP] mags = np.add(mags, dmags) print("Magnitudes: ", mags) break # self._update_magnitudes(mags) _ = analyze_image(im, ntraps, verbose=verbose) def _run_cam(which_cam, verbose=False): names = ['ThorCam', 'ChamberCam'] # False, True ## https://instrumental-lib.readthedocs.io/en/stable/uc480-cameras.html ## cam = instrument(names[which_cam]) ## Cam Live Stream ## cam.start_live_video(framerate=10 * u.hertz) exp_t = cam._get_exposure() ## Create Figure ## fig = plt.figure() ax1 = fig.add_subplot(1, 1, 1) ## Animation Frame ## def animate(i): if cam.wait_for_frame(): im = cam.latest_frame() ax1.clear() ax1.imshow(im) ## Button: Automatic Exposure Adjustment ## def find_exposure(event): fix_exposure(cam, set_exposure, verbose) ## Button: Intensity Feedback ## def stabilize(event): # Wrapper for Intensity Feedback function. im = cam.latest_frame() print(analyze_image(which_cam, im, 12, 1, True)) # stabilize_intensity(which_cam, cam, verbose) def snapshot(event): im = cam.latest_frame() guess_image(which_cam, im, 12) def switch_cam(event): nonlocal cam, which_cam cam.close() which_cam = not which_cam cam = instrument(names[which_cam]) cam.start_live_video(framerate=10 * u.hertz) # ## Button: Pause ## # def playback(event): # if playback.running: # spcm_dwSetParam_i32(self.hCard, SPC_M2CMD, M2CMD_CARD_STOP) # playback.running = 0 # else: # spcm_dwSetParam_i32(self.hCard, SPC_M2CMD, M2CMD_CARD_START | M2CMD_CARD_ENABLETRIGGER) # playback.running = 1 # playback.running = 1 ## Slider: Exposure ## def adjust_exposure(exp_t): cam._set_exposure(exp_t * u.milliseconds) ## Button Construction ## axspos = plt.axes([0.56, 0.0, 0.13, 0.05]) axstab = plt.axes([0.7, 0.0, 0.1, 0.05]) # axstop = plt.axes([0.81, 0.0, 0.12, 0.05]) axplot = plt.axes([0.81, 0.0, 0.09, 0.05]) ### ! axswch = plt.axes([0.91, 0.0, 0.09, 0.05]) axspar = plt.axes([0.14, 0.9, 0.73, 0.05]) correct_exposure = Button(axspos, 'AutoExpose') stabilize_button = Button(axstab, 'Stabilize') # pause_play = Button(axstop, 'Pause/Play') plot_snapshot = Button(axplot, 'Plot') switch_cameras = Button(axswch, 'Switch') set_exposure = Slider(axspar, 'Exposure', valmin=0.1, valmax=MAX_EXP, valinit=exp_t.magnitude) correct_exposure.on_clicked(find_exposure) stabilize_button.on_clicked(stabilize) # pause_play.on_clicked(playback) plot_snapshot.on_clicked(snapshot) switch_cameras.on_clicked(switch_cam) set_exposure.on_changed(adjust_exposure) ## Begin Animation ## _ = animation.FuncAnimation(fig, animate, interval=100) plt.show() plt.close(fig) ########## Helper Functions ############### # noinspection PyPep8Naming def gaussian1d(x, x0, w0, A, offset): """ Returns intensity profile of 1d gaussian beam x0: x-offset w0: waist of Gaussian beam A: Amplitude offset: Global offset """ if w0 == 0: return 0 return A * np.exp(-2 * (x - x0) ** 2 / (w0 ** 2)) + offset # noinspection PyPep8Naming def gaussianarray1d(x, x0_vec, wx_vec, A_vec, offset, ntraps): """ Returns intensity profile of trap array x0_vec: 1-by-ntraps array of x-offsets of traps wx_vec: 1-by-ntraps array of waists of traps A_vec: 1-by-ntraps array of amplitudes of traps offset: global offset ntraps: Number of traps """ array = np.zeros(np.shape(x)) for k in range(ntraps): array = array + gaussian1d(x, x0_vec[k], wx_vec[k], A_vec[k], 0) return array + offset def wrapper_fit_func(x, ntraps, *args): """ Juggles parameters in order to be able to fit a list of parameters """ a, b, c = list(args[0][:ntraps]), list(args[0][ntraps:2 * ntraps]), list(args[0][2 * ntraps:3 * ntraps]) offset = args[0][-1] return gaussianarray1d(x, a, b, c, offset, ntraps) def guess_image(which_cam, image, ntraps): """ Scans the given image for the 'ntraps' number of trap intensity peaks. Then extracts the 1-dimensional gaussian profiles across the traps and returns a list of the amplitudes. """ threshes = [0.5, 0.65] ## Image Conditioning ## margin = 10 threshold = np.max(image)*threshes[which_cam] im = image.transpose() x_len = len(im) peak_locs = np.zeros(x_len) peak_vals = np.zeros(x_len) ## Trap Peak Detection ## for i in range(x_len): if i < margin or x_len - i < margin: peak_locs[i] = 0 peak_vals[i] = 0 else: peak_locs[i] = np.argmax(im[i]) peak_vals[i] = max(im[i]) ## Trap Range Detection ## first = True pos_first, pos_last = 0, 0 left_pos = 0 for i, p in enumerate(peak_vals): if p > threshold: left_pos = i elif p < threshold and left_pos != 0: if first: pos_first = (left_pos + i) // 2 first = False pos_last = (left_pos + i) // 2 left_pos = 0 ## Separation Value ## separation = (pos_last - pos_first) / ntraps # In Pixels ## Initial Guesses ## means0 = np.linspace(pos_first, pos_last, ntraps).tolist() waists0 = (separation * np.ones(ntraps) / 2).tolist() ampls0 = (max(peak_vals) * 0.7 * np.ones(ntraps)).tolist() _params0 = [means0, waists0, ampls0, [0.06]] params0 = [item for sublist in _params0 for item in sublist] xdata = np.arange(x_len) plt.figure() plt.plot(xdata, peak_vals) plt.plot(xdata, wrapper_fit_func(xdata, ntraps, params0), '--r') # Initial Guess plt.xlim((pos_first - margin, pos_last + margin)) plt.legend(["Data", "Guess", "Fit"]) plt.show(block=False) def analyze_image(which_cam, image, ntraps, iteration=0, verbose=False): """ Scans the given image for the 'ntraps' number of trap intensity peaks. Then extracts the 1-dimensional gaussian profiles across the traps and returns a list of the amplitudes. """ threshes = [0.5, 0.6] margin = 10 threshold = np.max(image) * threshes[which_cam] im = image.transpose() x_len = len(im) peak_locs = np.zeros(x_len) peak_vals = np.zeros(x_len) ## Trap Peak Detection ## for i in range(x_len): if i < margin or x_len - i < margin: peak_locs[i] = 0 peak_vals[i] = 0 else: peak_locs[i] = np.argmax(im[i]) peak_vals[i] = max(im[i]) ## Trap Range Detection ## first =
print("WOW") break
conditional_block
gui_debugging.py
sqrt(abs(dP)) * L for dP in deltaP] mags = np.add(mags, dmags) print("Magnitudes: ", mags) break # self._update_magnitudes(mags) _ = analyze_image(im, ntraps, verbose=verbose) def _run_cam(which_cam, verbose=False): names = ['ThorCam', 'ChamberCam'] # False, True ## https://instrumental-lib.readthedocs.io/en/stable/uc480-cameras.html ## cam = instrument(names[which_cam]) ## Cam Live Stream ## cam.start_live_video(framerate=10 * u.hertz) exp_t = cam._get_exposure() ## Create Figure ## fig = plt.figure() ax1 = fig.add_subplot(1, 1, 1) ## Animation Frame ## def
(i): if cam.wait_for_frame(): im = cam.latest_frame() ax1.clear() ax1.imshow(im) ## Button: Automatic Exposure Adjustment ## def find_exposure(event): fix_exposure(cam, set_exposure, verbose) ## Button: Intensity Feedback ## def stabilize(event): # Wrapper for Intensity Feedback function. im = cam.latest_frame() print(analyze_image(which_cam, im, 12, 1, True)) # stabilize_intensity(which_cam, cam, verbose) def snapshot(event): im = cam.latest_frame() guess_image(which_cam, im, 12) def switch_cam(event): nonlocal cam, which_cam cam.close() which_cam = not which_cam cam = instrument(names[which_cam]) cam.start_live_video(framerate=10 * u.hertz) # ## Button: Pause ## # def playback(event): # if playback.running: # spcm_dwSetParam_i32(self.hCard, SPC_M2CMD, M2CMD_CARD_STOP) # playback.running = 0 # else: # spcm_dwSetParam_i32(self.hCard, SPC_M2CMD, M2CMD_CARD_START | M2CMD_CARD_ENABLETRIGGER) # playback.running = 1 # playback.running = 1 ## Slider: Exposure ## def adjust_exposure(exp_t): cam._set_exposure(exp_t * u.milliseconds) ## Button Construction ## axspos = plt.axes([0.56, 0.0, 0.13, 0.05]) axstab = plt.axes([0.7, 0.0, 0.1, 0.05]) # axstop = plt.axes([0.81, 0.0, 0.12, 0.05]) axplot = plt.axes([0.81, 0.0, 0.09, 0.05]) ### ! axswch = plt.axes([0.91, 0.0, 0.09, 0.05]) axspar = plt.axes([0.14, 0.9, 0.73, 0.05]) correct_exposure = Button(axspos, 'AutoExpose') stabilize_button = Button(axstab, 'Stabilize') # pause_play = Button(axstop, 'Pause/Play') plot_snapshot = Button(axplot, 'Plot') switch_cameras = Button(axswch, 'Switch') set_exposure = Slider(axspar, 'Exposure', valmin=0.1, valmax=MAX_EXP, valinit=exp_t.magnitude) correct_exposure.on_clicked(find_exposure) stabilize_button.on_clicked(stabilize) # pause_play.on_clicked(playback) plot_snapshot.on_clicked(snapshot) switch_cameras.on_clicked(switch_cam) set_exposure.on_changed(adjust_exposure) ## Begin Animation ## _ = animation.FuncAnimation(fig, animate, interval=100) plt.show() plt.close(fig) ########## Helper Functions ############### # noinspection PyPep8Naming def gaussian1d(x, x0, w0, A, offset): """ Returns intensity profile of 1d gaussian beam x0: x-offset w0: waist of Gaussian beam A: Amplitude offset: Global offset """ if w0 == 0: return 0 return A * np.exp(-2 * (x - x0) ** 2 / (w0 ** 2)) + offset # noinspection PyPep8Naming def gaussianarray1d(x, x0_vec, wx_vec, A_vec, offset, ntraps): """ Returns intensity profile of trap array x0_vec: 1-by-ntraps array of x-offsets of traps wx_vec: 1-by-ntraps array of waists of traps A_vec: 1-by-ntraps array of amplitudes of traps offset: global offset ntraps: Number of traps """ array = np.zeros(np.shape(x)) for k in range(ntraps): array = array + gaussian1d(x, x0_vec[k], wx_vec[k], A_vec[k], 0) return array + offset def wrapper_fit_func(x, ntraps, *args): """ Juggles parameters in order to be able to fit a list of parameters """ a, b, c = list(args[0][:ntraps]), list(args[0][ntraps:2 * ntraps]), list(args[0][2 * ntraps:3 * ntraps]) offset = args[0][-1] return gaussianarray1d(x, a, b, c, offset, ntraps) def guess_image(which_cam, image, ntraps): """ Scans the given image for the 'ntraps' number of trap intensity peaks. Then extracts the 1-dimensional gaussian profiles across the traps and returns a list of the amplitudes. """ threshes = [0.5, 0.65] ## Image Conditioning ## margin = 10 threshold = np.max(image)*threshes[which_cam] im = image.transpose() x_len = len(im) peak_locs = np.zeros(x_len) peak_vals = np.zeros(x_len) ## Trap Peak Detection ## for i in range(x_len): if i < margin or x_len - i < margin: peak_locs[i] = 0 peak_vals[i] = 0 else: peak_locs[i] = np.argmax(im[i]) peak_vals[i] = max(im[i]) ## Trap Range Detection ## first = True pos_first, pos_last = 0, 0 left_pos = 0 for i, p in enumerate(peak_vals): if p > threshold: left_pos = i elif p < threshold and left_pos != 0: if first: pos_first = (left_pos + i) // 2 first = False pos_last = (left_pos + i) // 2 left_pos = 0 ## Separation Value ## separation = (pos_last - pos_first) / ntraps # In Pixels ## Initial Guesses ## means0 = np.linspace(pos_first, pos_last, ntraps).tolist() waists0 = (separation * np.ones(ntraps) / 2).tolist() ampls0 = (max(peak_vals) * 0.7 * np.ones(ntraps)).tolist() _params0 = [means0, waists0, ampls0, [0.06]] params0 = [item for sublist in _params0 for item in sublist] xdata = np.arange(x_len) plt.figure() plt.plot(xdata, peak_vals) plt.plot(xdata, wrapper_fit_func(xdata, ntraps, params0), '--r') # Initial Guess plt.xlim((pos_first - margin, pos_last + margin)) plt.legend(["Data", "Guess", "Fit"]) plt.show(block=False) def analyze_image(which_cam, image, ntraps, iteration=0, verbose=False): """ Scans the given image for the 'ntraps' number of trap intensity peaks. Then extracts the 1-dimensional gaussian profiles across the traps and returns a list of the amplitudes. """ threshes = [0.5, 0.6] margin = 10 threshold = np.max(image) * threshes[which_cam] im = image.transpose() x_len = len(im) peak_locs = np.zeros(x_len) peak_vals = np.zeros(x_len) ## Trap Peak Detection ## for i in range(x_len): if i < margin or x_len - i < margin: peak_locs[i] = 0 peak_vals[i] = 0 else: peak_locs[i] = np.argmax(im[i]) peak_vals[i] = max(im[i]) ## Trap Range Detection ## first = True pos_first, pos_last = 0, 0 left_pos = 0 for i, p in enumerate(peak_vals): if p
animate
identifier_name
gui_debugging.py
sqrt(abs(dP)) * L for dP in deltaP] mags = np.add(mags, dmags) print("Magnitudes: ", mags) break # self._update_magnitudes(mags) _ = analyze_image(im, ntraps, verbose=verbose) def _run_cam(which_cam, verbose=False): names = ['ThorCam', 'ChamberCam'] # False, True ## https://instrumental-lib.readthedocs.io/en/stable/uc480-cameras.html ## cam = instrument(names[which_cam]) ## Cam Live Stream ## cam.start_live_video(framerate=10 * u.hertz) exp_t = cam._get_exposure() ## Create Figure ## fig = plt.figure() ax1 = fig.add_subplot(1, 1, 1) ## Animation Frame ## def animate(i): if cam.wait_for_frame(): im = cam.latest_frame() ax1.clear() ax1.imshow(im) ## Button: Automatic Exposure Adjustment ## def find_exposure(event): fix_exposure(cam, set_exposure, verbose) ## Button: Intensity Feedback ## def stabilize(event): # Wrapper for Intensity Feedback function. im = cam.latest_frame() print(analyze_image(which_cam, im, 12, 1, True)) # stabilize_intensity(which_cam, cam, verbose) def snapshot(event): im = cam.latest_frame() guess_image(which_cam, im, 12) def switch_cam(event): nonlocal cam, which_cam cam.close() which_cam = not which_cam cam = instrument(names[which_cam]) cam.start_live_video(framerate=10 * u.hertz) # ## Button: Pause ## # def playback(event): # if playback.running: # spcm_dwSetParam_i32(self.hCard, SPC_M2CMD, M2CMD_CARD_STOP) # playback.running = 0 # else: # spcm_dwSetParam_i32(self.hCard, SPC_M2CMD, M2CMD_CARD_START | M2CMD_CARD_ENABLETRIGGER) # playback.running = 1 # playback.running = 1 ## Slider: Exposure ## def adjust_exposure(exp_t): cam._set_exposure(exp_t * u.milliseconds) ## Button Construction ## axspos = plt.axes([0.56, 0.0, 0.13, 0.05]) axstab = plt.axes([0.7, 0.0, 0.1, 0.05]) # axstop = plt.axes([0.81, 0.0, 0.12, 0.05]) axplot = plt.axes([0.81, 0.0, 0.09, 0.05]) ### ! axswch = plt.axes([0.91, 0.0, 0.09, 0.05]) axspar = plt.axes([0.14, 0.9, 0.73, 0.05]) correct_exposure = Button(axspos, 'AutoExpose') stabilize_button = Button(axstab, 'Stabilize') # pause_play = Button(axstop, 'Pause/Play') plot_snapshot = Button(axplot, 'Plot') switch_cameras = Button(axswch, 'Switch') set_exposure = Slider(axspar, 'Exposure', valmin=0.1, valmax=MAX_EXP, valinit=exp_t.magnitude) correct_exposure.on_clicked(find_exposure)
switch_cameras.on_clicked(switch_cam) set_exposure.on_changed(adjust_exposure) ## Begin Animation ## _ = animation.FuncAnimation(fig, animate, interval=100) plt.show() plt.close(fig) ########## Helper Functions ############### # noinspection PyPep8Naming def gaussian1d(x, x0, w0, A, offset): """ Returns intensity profile of 1d gaussian beam x0: x-offset w0: waist of Gaussian beam A: Amplitude offset: Global offset """ if w0 == 0: return 0 return A * np.exp(-2 * (x - x0) ** 2 / (w0 ** 2)) + offset # noinspection PyPep8Naming def gaussianarray1d(x, x0_vec, wx_vec, A_vec, offset, ntraps): """ Returns intensity profile of trap array x0_vec: 1-by-ntraps array of x-offsets of traps wx_vec: 1-by-ntraps array of waists of traps A_vec: 1-by-ntraps array of amplitudes of traps offset: global offset ntraps: Number of traps """ array = np.zeros(np.shape(x)) for k in range(ntraps): array = array + gaussian1d(x, x0_vec[k], wx_vec[k], A_vec[k], 0) return array + offset def wrapper_fit_func(x, ntraps, *args): """ Juggles parameters in order to be able to fit a list of parameters """ a, b, c = list(args[0][:ntraps]), list(args[0][ntraps:2 * ntraps]), list(args[0][2 * ntraps:3 * ntraps]) offset = args[0][-1] return gaussianarray1d(x, a, b, c, offset, ntraps) def guess_image(which_cam, image, ntraps): """ Scans the given image for the 'ntraps' number of trap intensity peaks. Then extracts the 1-dimensional gaussian profiles across the traps and returns a list of the amplitudes. """ threshes = [0.5, 0.65] ## Image Conditioning ## margin = 10 threshold = np.max(image)*threshes[which_cam] im = image.transpose() x_len = len(im) peak_locs = np.zeros(x_len) peak_vals = np.zeros(x_len) ## Trap Peak Detection ## for i in range(x_len): if i < margin or x_len - i < margin: peak_locs[i] = 0 peak_vals[i] = 0 else: peak_locs[i] = np.argmax(im[i]) peak_vals[i] = max(im[i]) ## Trap Range Detection ## first = True pos_first, pos_last = 0, 0 left_pos = 0 for i, p in enumerate(peak_vals): if p > threshold: left_pos = i elif p < threshold and left_pos != 0: if first: pos_first = (left_pos + i) // 2 first = False pos_last = (left_pos + i) // 2 left_pos = 0 ## Separation Value ## separation = (pos_last - pos_first) / ntraps # In Pixels ## Initial Guesses ## means0 = np.linspace(pos_first, pos_last, ntraps).tolist() waists0 = (separation * np.ones(ntraps) / 2).tolist() ampls0 = (max(peak_vals) * 0.7 * np.ones(ntraps)).tolist() _params0 = [means0, waists0, ampls0, [0.06]] params0 = [item for sublist in _params0 for item in sublist] xdata = np.arange(x_len) plt.figure() plt.plot(xdata, peak_vals) plt.plot(xdata, wrapper_fit_func(xdata, ntraps, params0), '--r') # Initial Guess plt.xlim((pos_first - margin, pos_last + margin)) plt.legend(["Data", "Guess", "Fit"]) plt.show(block=False) def analyze_image(which_cam, image, ntraps, iteration=0, verbose=False): """ Scans the given image for the 'ntraps' number of trap intensity peaks. Then extracts the 1-dimensional gaussian profiles across the traps and returns a list of the amplitudes. """ threshes = [0.5, 0.6] margin = 10 threshold = np.max(image) * threshes[which_cam] im = image.transpose() x_len = len(im) peak_locs = np.zeros(x_len) peak_vals = np.zeros(x_len) ## Trap Peak Detection ## for i in range(x_len): if i < margin or x_len - i < margin: peak_locs[i] = 0 peak_vals[i] = 0 else: peak_locs[i] = np.argmax(im[i]) peak_vals[i] = max(im[i]) ## Trap Range Detection ## first = True pos_first, pos_last = 0, 0 left_pos = 0 for i, p in enumerate(peak_vals): if p >
stabilize_button.on_clicked(stabilize) # pause_play.on_clicked(playback) plot_snapshot.on_clicked(snapshot)
random_line_split
gui_debugging.py
mean_power = trap_powers.mean() rel_dif = 100 * trap_powers.std() / mean_power print(f'Relative Power Difference: {rel_dif:.2f} %') if rel_dif < 0.8: print("WOW") break deltaP = [mean_power - P for P in trap_powers] dmags = [(dP / abs(dP)) * sqrt(abs(dP)) * L for dP in deltaP] mags = np.add(mags, dmags) print("Magnitudes: ", mags) break # self._update_magnitudes(mags) _ = analyze_image(im, ntraps, verbose=verbose) def _run_cam(which_cam, verbose=False): names = ['ThorCam', 'ChamberCam'] # False, True ## https://instrumental-lib.readthedocs.io/en/stable/uc480-cameras.html ## cam = instrument(names[which_cam]) ## Cam Live Stream ## cam.start_live_video(framerate=10 * u.hertz) exp_t = cam._get_exposure() ## Create Figure ## fig = plt.figure() ax1 = fig.add_subplot(1, 1, 1) ## Animation Frame ## def animate(i): if cam.wait_for_frame(): im = cam.latest_frame() ax1.clear() ax1.imshow(im) ## Button: Automatic Exposure Adjustment ## def find_exposure(event): fix_exposure(cam, set_exposure, verbose) ## Button: Intensity Feedback ## def stabilize(event): # Wrapper for Intensity Feedback function. im = cam.latest_frame() print(analyze_image(which_cam, im, 12, 1, True)) # stabilize_intensity(which_cam, cam, verbose) def snapshot(event): im = cam.latest_frame() guess_image(which_cam, im, 12) def switch_cam(event): nonlocal cam, which_cam cam.close() which_cam = not which_cam cam = instrument(names[which_cam]) cam.start_live_video(framerate=10 * u.hertz) # ## Button: Pause ## # def playback(event): # if playback.running: # spcm_dwSetParam_i32(self.hCard, SPC_M2CMD, M2CMD_CARD_STOP) # playback.running = 0 # else: # spcm_dwSetParam_i32(self.hCard, SPC_M2CMD, M2CMD_CARD_START | M2CMD_CARD_ENABLETRIGGER) # playback.running = 1 # playback.running = 1 ## Slider: Exposure ## def adjust_exposure(exp_t): cam._set_exposure(exp_t * u.milliseconds) ## Button Construction ## axspos = plt.axes([0.56, 0.0, 0.13, 0.05]) axstab = plt.axes([0.7, 0.0, 0.1, 0.05]) # axstop = plt.axes([0.81, 0.0, 0.12, 0.05]) axplot = plt.axes([0.81, 0.0, 0.09, 0.05]) ### ! axswch = plt.axes([0.91, 0.0, 0.09, 0.05]) axspar = plt.axes([0.14, 0.9, 0.73, 0.05]) correct_exposure = Button(axspos, 'AutoExpose') stabilize_button = Button(axstab, 'Stabilize') # pause_play = Button(axstop, 'Pause/Play') plot_snapshot = Button(axplot, 'Plot') switch_cameras = Button(axswch, 'Switch') set_exposure = Slider(axspar, 'Exposure', valmin=0.1, valmax=MAX_EXP, valinit=exp_t.magnitude) correct_exposure.on_clicked(find_exposure) stabilize_button.on_clicked(stabilize) # pause_play.on_clicked(playback) plot_snapshot.on_clicked(snapshot) switch_cameras.on_clicked(switch_cam) set_exposure.on_changed(adjust_exposure) ## Begin Animation ## _ = animation.FuncAnimation(fig, animate, interval=100) plt.show() plt.close(fig) ########## Helper Functions ############### # noinspection PyPep8Naming def gaussian1d(x, x0, w0, A, offset): """ Returns intensity profile of 1d gaussian beam x0: x-offset w0: waist of Gaussian beam A: Amplitude offset: Global offset """ if w0 == 0: return 0 return A * np.exp(-2 * (x - x0) ** 2 / (w0 ** 2)) + offset # noinspection PyPep8Naming def gaussianarray1d(x, x0_vec, wx_vec, A_vec, offset, ntraps): """ Returns intensity profile of trap array x0_vec: 1-by-ntraps array of x-offsets of traps wx_vec: 1-by-ntraps array of waists of traps A_vec: 1-by-ntraps array of amplitudes of traps offset: global offset ntraps: Number of traps """ array = np.zeros(np.shape(x)) for k in range(ntraps): array = array + gaussian1d(x, x0_vec[k], wx_vec[k], A_vec[k], 0) return array + offset def wrapper_fit_func(x, ntraps, *args): """ Juggles parameters in order to be able to fit a list of parameters """ a, b, c = list(args[0][:ntraps]), list(args[0][ntraps:2 * ntraps]), list(args[0][2 * ntraps:3 * ntraps]) offset = args[0][-1] return gaussianarray1d(x, a, b, c, offset, ntraps) def guess_image(which_cam, image, ntraps): """ Scans the given image for the 'ntraps' number of trap intensity peaks. Then extracts the 1-dimensional gaussian profiles across the traps and returns a list of the amplitudes. """ threshes = [0.5, 0.65] ## Image Conditioning ## margin = 10 threshold = np.max(image)*threshes[which_cam] im = image.transpose() x_len = len(im) peak_locs = np.zeros(x_len) peak_vals = np.zeros(x_len) ## Trap Peak Detection ## for i in range(x_len): if i < margin or x_len - i < margin: peak_locs[i] = 0 peak_vals[i] = 0 else: peak_locs[i] = np.argmax(im[i]) peak_vals[i] = max(im[i]) ## Trap Range Detection ## first = True pos_first, pos_last = 0, 0 left_pos = 0 for i, p in enumerate(peak_vals): if p > threshold: left_pos = i elif p < threshold and left_pos != 0: if first: pos_first = (left_pos + i) // 2 first = False pos_last = (left_pos + i) // 2 left_pos = 0 ## Separation Value ## separation = (pos_last - pos_first) / ntraps # In Pixels ## Initial Guesses ## means0 = np.linspace(pos_first, pos_last, ntraps).tolist() waists0 = (separation * np.ones(ntraps) / 2).tolist() ampls0 = (max(peak_vals) * 0.7 * np.ones(ntraps)).tolist() _params0 = [means0, waists0, ampls0, [0.06]] params0 = [item for sublist in _params0 for item in sublist] xdata = np.arange(x_len) plt.figure() plt.plot(xdata, peak_vals) plt.plot(xdata, wrapper_fit_func(xdata, ntraps, params0), '--r') # Initial Guess plt.xlim((pos_first - margin,
""" Given a UC480 camera object (instrumental module) and a number indicating the number of trap objects, applies an iterative image analysis to individual trap adjustment in order to achieve a nearly homogeneous intensity profile across traps. """ L = 0.5 # Correction Rate mags = np.ones(12) ### ! ntraps = len(mags) iteration = 0 while iteration < 5: iteration += 1 print("Iteration ", iteration) im = cam.latest_frame() try: trap_powers = analyze_image(which_cam, im, ntraps, iteration, verbose) except (AttributeError, ValueError) as e: print("No Bueno, error occurred during image analysis:\n", e) break
identifier_body
operations.go
w", err) } op.store.session, err = session.NewStore() if err != nil { return nil, fmt.Errorf("failed to create web auth protocol session store: %w", err) } op.store.users, err = user.NewStore(config.Storage.Storage) if err != nil { return nil, fmt.Errorf("failed to open users store: %w", err) } return op, nil } // GetRESTHandlers get all controller API handler available for this service. func (o *Operation) GetRESTHandlers() []common.Handler { return []common.Handler{ common.NewHTTPHandler(registerBeginPath, http.MethodGet, o.beginRegistration), common.NewHTTPHandler(registerFinishPath, http.MethodPost, o.finishRegistration), common.NewHTTPHandler(loginBeginPath, http.MethodGet, o.beginLogin), common.NewHTTPHandler(loginFinishPath, http.MethodPost, o.finishLogin), } } func (o *Operation) beginRegistration(w http.ResponseWriter, r *http.Request) { logger.Debugf("handling device registration: %s", r.URL.String()) userData, canProceed := o.getUserData(w, r, userSubCookieName) if !canProceed { return } device := NewDevice(userData) webAuthnUser := protocol.UserEntity{ ID: device.WebAuthnID(), DisplayName: device.WebAuthnDisplayName(), CredentialEntity: protocol.CredentialEntity{ Name: device.WebAuthnName(), }, } registerOptions := func(credCreationOpts *protocol.PublicKeyCredentialCreationOptions) { credCreationOpts.User = webAuthnUser credCreationOpts.CredentialExcludeList = device.CredentialExcludeList() credCreationOpts.Attestation = protocol.PreferDirectAttestation } // generate PublicKeyCredentialCreationOptions, session data credentialParams, sessionData, err := o.webauthn.BeginRegistration( device, registerOptions, ) if err != nil { common.WriteErrorResponsef(w, logger, http.StatusInternalServerError, "failed to begin registration %s", err.Error()) return } // store session data as marshaled JSON err = o.store.session.SaveWebauthnSession(userData.Sub, sessionData, r, w) if err != nil { common.WriteErrorResponsef(w, logger, http.StatusInternalServerError, "failed to save web auth session %s", err.Error()) return } jsonResponse(w, credentialParams, http.StatusOK) logger.Debugf("Registration begins") } func (o *Operation) finishRegistration(w http.ResponseWriter, r *http.Request) { // nolint:funlen // not clean to split logger.Debugf("handling finish device registration: %s", r.URL.String()) userData, canProceed := o.getUserData(w, r, userSubCookieName) if !canProceed { return } sessionData, err := o.store.session.GetWebauthnSession(userData.Sub, r) if err != nil { common.WriteErrorResponsef(w, logger, http.StatusInternalServerError, "failed to get web auth session: %s", err.Error()) return } device := NewDevice(userData) // unfold webauthn.FinishRegistration, to access parsedResponse parsedResponse, err := protocol.ParseCredentialCreationResponse(r) if err != nil { common.WriteErrorResponsef(w, logger, http.StatusInternalServerError, "failed to finish registration: parsing ccr: %#v", err) return } credential, err := o.webauthn.CreateCredential(device, sessionData, parsedResponse) if err != nil { common.WriteErrorResponsef(w, logger, http.StatusInternalServerError, "failed to finish registration: cred: %#v", err) return } deviceCerts, ok := parsedResponse.Response.AttestationObject.AttStatement["x5c"].([]interface{}) if ok && len(deviceCerts) > 0 { err = o.requestDeviceValidation(r.Context(), userData.Sub, string(credential.Authenticator.AAGUID), deviceCerts) if err != nil { common.WriteErrorResponsef(w, logger, http.StatusInternalServerError, "failed to finish registration: %#v", err) return } } else { logger.Warnf("credential attestation of format '%s' has no certificates", parsedResponse.Response.AttestationObject.Format) } device.AddCredential(*credential) err = o.saveDeviceInfo(device) if err != nil { common.WriteErrorResponsef(w, logger, http.StatusInternalServerError, "failed to save device info: %s", err.Error()) return } o.saveCookie(w, r, userData.Sub, deviceCookieName) jsonResponse(w, credential, http.StatusOK) logger.Debugf("Registration success") } func (o *Operation) requestDeviceValidation(ctx context.Context, userSub, aaguid string, certs []interface{}) error { if len(certs) == 0 { return fmt.Errorf("missing certs") } var certPemList []string for _, certInterface := range certs { cert, ok := certInterface.([]byte) if !ok { return fmt.Errorf("can't cast certificate data to []byte") } certPemList = append(certPemList, string(pem.EncodeToMemory( &pem.Block{Bytes: cert, Type: "CERTIFICATE"}, ))) } postData, err := json.Marshal(&struct { X5c []string `json:"x5c"` Sub string `json:"sub"` Aaguid string `json:"aaguid"` }{ X5c: certPemList, Sub: userSub, Aaguid: aaguid, }) if err != nil { return fmt.Errorf("failed to marshal cert data: %w", err) } req, err := http.NewRequestWithContext(ctx, http.MethodPost, o.hubAuthURL+"/device", bytes.NewBuffer(postData)) if err != nil { return fmt.Errorf("failed to create request: %w", err) } _, _, err = common.SendHTTPRequest(req, o.httpClient, http.StatusOK, nil) if err != nil { return fmt.Errorf("failed response from hub-auth/device endpoint: %w", err) } return nil } func (o *Operation) beginLogin(w http.ResponseWriter, r *http.Request) { logger.Debugf("handling begin device login: %s", r.URL.String()) // get username userData, canProceed := o.getUserData(w, r, deviceCookieName) if !canProceed { return } deviceData, err := o.getDeviceInfo(userData.Sub) if err != nil { common.WriteErrorResponsef(w, logger, http.StatusBadRequest, "failed to get device data: %s", err.Error()) return } // generate PublicKeyCredentialRequestOptions, session data options, sessionData, err := o.webauthn.BeginLogin(deviceData) if err != nil { common.WriteErrorResponsef(w, logger, http.StatusInternalServerError, "failed to begin login: %s", err.Error()) return } // store session data as marshaled JSON err = o.store.session.SaveWebauthnSession(deviceData.ID, sessionData, r, w) if err != nil { common.WriteErrorResponsef(w, logger, http.StatusInternalServerError, "failed to save web auth login session: %s", err.Error()) return } logger.Debugf("Login begin success") jsonResponse(w, options, http.StatusOK) } func (o *Operation) finishLogin(w http.ResponseWriter, r *http.Request) { logger.Debugf("handling finish device login: %s", r.URL.String()) // get username userData, canProceed := o.getUserData(w, r, deviceCookieName) if !canProceed { return } deviceData, err := o.getDeviceInfo(userData.Sub) if err != nil { common.WriteErrorResponsef(w, logger, http.StatusBadRequest, "failed to get device data: %s", err.Error()) return } o.saveCookie(w, r, userData.Sub, userSubCookieName) // load the session data sessionData, err := o.store.session.GetWebauthnSession(deviceData.ID, r) if err != nil { common.WriteErrorResponsef(w, logger, http.StatusBadRequest, "failed to get web auth login session: %s", err.Error()) return } _, err = o.webauthn.FinishLogin(deviceData, sessionData, r) if err != nil { common.WriteErrorResponsef(w, logger, http.StatusBadRequest, "failed to finish login: %s", err.Error()) return } // handle successful login http.Redirect(w, r, o.walletDashboard, http.StatusFound) logger.Debugf("Login finish success") } func (o *Operation) getUserData(w http.ResponseWriter, r *http.Request, cookieName string) (userData *user.User, proceed bool) { cookieSession, err := o.store.cookies.Open(r) if err != nil { common.WriteErrorResponsef(w, logger, http.StatusInternalServerError, "failed to create or decode session cookie: %s", err.Error()) return nil, false } userSub, found := cookieSession.Get(cookieName) if !found { common.WriteErrorResponsef(w, logger, http.StatusNotFound, "missing device user session cookie") return nil, false } userData, err = o.store.users.Get(fmt.Sprintf("%v", userSub)) if err != nil
{ common.WriteErrorResponsef(w, logger, http.StatusInternalServerError, "failed to get user data %s:", err.Error()) return nil, false }
conditional_block
operations.go
", err) } op.store.users, err = user.NewStore(config.Storage.Storage) if err != nil { return nil, fmt.Errorf("failed to open users store: %w", err) } return op, nil } // GetRESTHandlers get all controller API handler available for this service. func (o *Operation) GetRESTHandlers() []common.Handler { return []common.Handler{ common.NewHTTPHandler(registerBeginPath, http.MethodGet, o.beginRegistration), common.NewHTTPHandler(registerFinishPath, http.MethodPost, o.finishRegistration), common.NewHTTPHandler(loginBeginPath, http.MethodGet, o.beginLogin), common.NewHTTPHandler(loginFinishPath, http.MethodPost, o.finishLogin), } } func (o *Operation) beginRegistration(w http.ResponseWriter, r *http.Request) { logger.Debugf("handling device registration: %s", r.URL.String()) userData, canProceed := o.getUserData(w, r, userSubCookieName) if !canProceed { return } device := NewDevice(userData) webAuthnUser := protocol.UserEntity{ ID: device.WebAuthnID(), DisplayName: device.WebAuthnDisplayName(), CredentialEntity: protocol.CredentialEntity{ Name: device.WebAuthnName(), }, } registerOptions := func(credCreationOpts *protocol.PublicKeyCredentialCreationOptions) { credCreationOpts.User = webAuthnUser credCreationOpts.CredentialExcludeList = device.CredentialExcludeList() credCreationOpts.Attestation = protocol.PreferDirectAttestation } // generate PublicKeyCredentialCreationOptions, session data credentialParams, sessionData, err := o.webauthn.BeginRegistration( device, registerOptions, ) if err != nil { common.WriteErrorResponsef(w, logger, http.StatusInternalServerError, "failed to begin registration %s", err.Error()) return } // store session data as marshaled JSON err = o.store.session.SaveWebauthnSession(userData.Sub, sessionData, r, w) if err != nil { common.WriteErrorResponsef(w, logger, http.StatusInternalServerError, "failed to save web auth session %s", err.Error()) return } jsonResponse(w, credentialParams, http.StatusOK) logger.Debugf("Registration begins") } func (o *Operation) finishRegistration(w http.ResponseWriter, r *http.Request) { // nolint:funlen // not clean to split logger.Debugf("handling finish device registration: %s", r.URL.String()) userData, canProceed := o.getUserData(w, r, userSubCookieName) if !canProceed { return } sessionData, err := o.store.session.GetWebauthnSession(userData.Sub, r) if err != nil { common.WriteErrorResponsef(w, logger, http.StatusInternalServerError, "failed to get web auth session: %s", err.Error()) return } device := NewDevice(userData) // unfold webauthn.FinishRegistration, to access parsedResponse parsedResponse, err := protocol.ParseCredentialCreationResponse(r) if err != nil { common.WriteErrorResponsef(w, logger, http.StatusInternalServerError, "failed to finish registration: parsing ccr: %#v", err) return } credential, err := o.webauthn.CreateCredential(device, sessionData, parsedResponse) if err != nil { common.WriteErrorResponsef(w, logger, http.StatusInternalServerError, "failed to finish registration: cred: %#v", err) return } deviceCerts, ok := parsedResponse.Response.AttestationObject.AttStatement["x5c"].([]interface{}) if ok && len(deviceCerts) > 0 { err = o.requestDeviceValidation(r.Context(), userData.Sub, string(credential.Authenticator.AAGUID), deviceCerts) if err != nil { common.WriteErrorResponsef(w, logger, http.StatusInternalServerError, "failed to finish registration: %#v", err) return } } else { logger.Warnf("credential attestation of format '%s' has no certificates", parsedResponse.Response.AttestationObject.Format) } device.AddCredential(*credential) err = o.saveDeviceInfo(device) if err != nil { common.WriteErrorResponsef(w, logger, http.StatusInternalServerError, "failed to save device info: %s", err.Error()) return } o.saveCookie(w, r, userData.Sub, deviceCookieName) jsonResponse(w, credential, http.StatusOK) logger.Debugf("Registration success") } func (o *Operation) requestDeviceValidation(ctx context.Context, userSub, aaguid string, certs []interface{}) error { if len(certs) == 0 { return fmt.Errorf("missing certs") } var certPemList []string for _, certInterface := range certs { cert, ok := certInterface.([]byte) if !ok { return fmt.Errorf("can't cast certificate data to []byte") } certPemList = append(certPemList, string(pem.EncodeToMemory( &pem.Block{Bytes: cert, Type: "CERTIFICATE"}, ))) } postData, err := json.Marshal(&struct { X5c []string `json:"x5c"` Sub string `json:"sub"` Aaguid string `json:"aaguid"` }{ X5c: certPemList, Sub: userSub, Aaguid: aaguid, }) if err != nil { return fmt.Errorf("failed to marshal cert data: %w", err) } req, err := http.NewRequestWithContext(ctx, http.MethodPost, o.hubAuthURL+"/device", bytes.NewBuffer(postData)) if err != nil { return fmt.Errorf("failed to create request: %w", err) } _, _, err = common.SendHTTPRequest(req, o.httpClient, http.StatusOK, nil) if err != nil { return fmt.Errorf("failed response from hub-auth/device endpoint: %w", err) } return nil } func (o *Operation) beginLogin(w http.ResponseWriter, r *http.Request) { logger.Debugf("handling begin device login: %s", r.URL.String()) // get username userData, canProceed := o.getUserData(w, r, deviceCookieName) if !canProceed { return } deviceData, err := o.getDeviceInfo(userData.Sub) if err != nil { common.WriteErrorResponsef(w, logger, http.StatusBadRequest, "failed to get device data: %s", err.Error()) return } // generate PublicKeyCredentialRequestOptions, session data options, sessionData, err := o.webauthn.BeginLogin(deviceData) if err != nil { common.WriteErrorResponsef(w, logger, http.StatusInternalServerError, "failed to begin login: %s", err.Error()) return } // store session data as marshaled JSON err = o.store.session.SaveWebauthnSession(deviceData.ID, sessionData, r, w) if err != nil { common.WriteErrorResponsef(w, logger, http.StatusInternalServerError, "failed to save web auth login session: %s", err.Error()) return } logger.Debugf("Login begin success") jsonResponse(w, options, http.StatusOK) } func (o *Operation) finishLogin(w http.ResponseWriter, r *http.Request) { logger.Debugf("handling finish device login: %s", r.URL.String()) // get username userData, canProceed := o.getUserData(w, r, deviceCookieName) if !canProceed { return } deviceData, err := o.getDeviceInfo(userData.Sub) if err != nil { common.WriteErrorResponsef(w, logger, http.StatusBadRequest, "failed to get device data: %s", err.Error()) return } o.saveCookie(w, r, userData.Sub, userSubCookieName) // load the session data sessionData, err := o.store.session.GetWebauthnSession(deviceData.ID, r) if err != nil { common.WriteErrorResponsef(w, logger, http.StatusBadRequest, "failed to get web auth login session: %s", err.Error()) return } _, err = o.webauthn.FinishLogin(deviceData, sessionData, r) if err != nil { common.WriteErrorResponsef(w, logger, http.StatusBadRequest, "failed to finish login: %s", err.Error()) return } // handle successful login http.Redirect(w, r, o.walletDashboard, http.StatusFound) logger.Debugf("Login finish success") } func (o *Operation) getUserData(w http.ResponseWriter, r *http.Request, cookieName string) (userData *user.User, proceed bool) { cookieSession, err := o.store.cookies.Open(r) if err != nil { common.WriteErrorResponsef(w, logger, http.StatusInternalServerError, "failed to create or decode session cookie: %s", err.Error()) return nil, false } userSub, found := cookieSession.Get(cookieName) if !found { common.WriteErrorResponsef(w, logger, http.StatusNotFound, "missing device user session cookie") return nil, false } userData, err = o.store.users.Get(fmt.Sprintf("%v", userSub)) if err != nil { common.WriteErrorResponsef(w, logger, http.StatusInternalServerError, "failed to get user data %s:", err.Error()) return nil, false } displayName := strings.Split(fmt.Sprintf("%v", userSub), "@")[0] userData.FamilyName = displayName return userData, true } func (o *Operation)
saveCookie
identifier_name
operations.go
authn.WebAuthn HubAuthURL string } // StorageConfig holds storage config. type StorageConfig struct { Storage ariesstorage.Provider SessionStore ariesstorage.Provider } type stores struct { users *user.Store cookies cookie.Store storage ariesstorage.Store session *session.Store } type httpClient interface { Do(req *http.Request) (*http.Response, error) } // Operation implements OIDC operations. type Operation struct { store *stores walletDashboard string tlsConfig *tls.Config httpClient httpClient webauthn *webauthn.WebAuthn hubAuthURL string } // New returns a new Operation. func New(config *Config) (*Operation, error)
op.store.session, err = session.NewStore() if err != nil { return nil, fmt.Errorf("failed to create web auth protocol session store: %w", err) } op.store.users, err = user.NewStore(config.Storage.Storage) if err != nil { return nil, fmt.Errorf("failed to open users store: %w", err) } return op, nil } // GetRESTHandlers get all controller API handler available for this service. func (o *Operation) GetRESTHandlers() []common.Handler { return []common.Handler{ common.NewHTTPHandler(registerBeginPath, http.MethodGet, o.beginRegistration), common.NewHTTPHandler(registerFinishPath, http.MethodPost, o.finishRegistration), common.NewHTTPHandler(loginBeginPath, http.MethodGet, o.beginLogin), common.NewHTTPHandler(loginFinishPath, http.MethodPost, o.finishLogin), } } func (o *Operation) beginRegistration(w http.ResponseWriter, r *http.Request) { logger.Debugf("handling device registration: %s", r.URL.String()) userData, canProceed := o.getUserData(w, r, userSubCookieName) if !canProceed { return } device := NewDevice(userData) webAuthnUser := protocol.UserEntity{ ID: device.WebAuthnID(), DisplayName: device.WebAuthnDisplayName(), CredentialEntity: protocol.CredentialEntity{ Name: device.WebAuthnName(), }, } registerOptions := func(credCreationOpts *protocol.PublicKeyCredentialCreationOptions) { credCreationOpts.User = webAuthnUser credCreationOpts.CredentialExcludeList = device.CredentialExcludeList() credCreationOpts.Attestation = protocol.PreferDirectAttestation } // generate PublicKeyCredentialCreationOptions, session data credentialParams, sessionData, err := o.webauthn.BeginRegistration( device, registerOptions, ) if err != nil { common.WriteErrorResponsef(w, logger, http.StatusInternalServerError, "failed to begin registration %s", err.Error()) return } // store session data as marshaled JSON err = o.store.session.SaveWebauthnSession(userData.Sub, sessionData, r, w) if err != nil { common.WriteErrorResponsef(w, logger, http.StatusInternalServerError, "failed to save web auth session %s", err.Error()) return } jsonResponse(w, credentialParams, http.StatusOK) logger.Debugf("Registration begins") } func (o *Operation) finishRegistration(w http.ResponseWriter, r *http.Request) { // nolint:funlen // not clean to split logger.Debugf("handling finish device registration: %s", r.URL.String()) userData, canProceed := o.getUserData(w, r, userSubCookieName) if !canProceed { return } sessionData, err := o.store.session.GetWebauthnSession(userData.Sub, r) if err != nil { common.WriteErrorResponsef(w, logger, http.StatusInternalServerError, "failed to get web auth session: %s", err.Error()) return } device := NewDevice(userData) // unfold webauthn.FinishRegistration, to access parsedResponse parsedResponse, err := protocol.ParseCredentialCreationResponse(r) if err != nil { common.WriteErrorResponsef(w, logger, http.StatusInternalServerError, "failed to finish registration: parsing ccr: %#v", err) return } credential, err := o.webauthn.CreateCredential(device, sessionData, parsedResponse) if err != nil { common.WriteErrorResponsef(w, logger, http.StatusInternalServerError, "failed to finish registration: cred: %#v", err) return } deviceCerts, ok := parsedResponse.Response.AttestationObject.AttStatement["x5c"].([]interface{}) if ok && len(deviceCerts) > 0 { err = o.requestDeviceValidation(r.Context(), userData.Sub, string(credential.Authenticator.AAGUID), deviceCerts) if err != nil { common.WriteErrorResponsef(w, logger, http.StatusInternalServerError, "failed to finish registration: %#v", err) return } } else { logger.Warnf("credential attestation of format '%s' has no certificates", parsedResponse.Response.AttestationObject.Format) } device.AddCredential(*credential) err = o.saveDeviceInfo(device) if err != nil { common.WriteErrorResponsef(w, logger, http.StatusInternalServerError, "failed to save device info: %s", err.Error()) return } o.saveCookie(w, r, userData.Sub, deviceCookieName) jsonResponse(w, credential, http.StatusOK) logger.Debugf("Registration success") } func (o *Operation) requestDeviceValidation(ctx context.Context, userSub, aaguid string, certs []interface{}) error { if len(certs) == 0 { return fmt.Errorf("missing certs") } var certPemList []string for _, certInterface := range certs { cert, ok := certInterface.([]byte) if !ok { return fmt.Errorf("can't cast certificate data to []byte") } certPemList = append(certPemList, string(pem.EncodeToMemory( &pem.Block{Bytes: cert, Type: "CERTIFICATE"}, ))) } postData, err := json.Marshal(&struct { X5c []string `json:"x5c"` Sub string `json:"sub"` Aaguid string `json:"aaguid"` }{ X5c: certPemList, Sub: userSub, Aaguid: aaguid, }) if err != nil { return fmt.Errorf("failed to marshal cert data: %w", err) } req, err := http.NewRequestWithContext(ctx, http.MethodPost, o.hubAuthURL+"/device", bytes.NewBuffer(postData)) if err != nil { return fmt.Errorf("failed to create request: %w", err) } _, _, err = common.SendHTTPRequest(req, o.httpClient, http.StatusOK, nil) if err != nil { return fmt.Errorf("failed response from hub-auth/device endpoint: %w", err) } return nil } func (o *Operation) beginLogin(w http.ResponseWriter, r *http.Request) { logger.Debugf("handling begin device login: %s", r.URL.String()) // get username userData, canProceed := o.getUserData(w, r, deviceCookieName) if !canProceed { return } deviceData, err := o.getDeviceInfo(userData.Sub) if err != nil { common.WriteErrorResponsef(w, logger, http.StatusBadRequest, "failed to get device data: %s", err.Error()) return } // generate PublicKeyCredentialRequestOptions, session data options, sessionData, err := o.webauthn.BeginLogin(deviceData) if err != nil { common.WriteErrorResponsef(w, logger, http.StatusInternalServerError, "failed to begin login: %s", err.Error()) return } // store session data as marshaled JSON err = o.store.session.SaveWebauthnSession(deviceData.ID, sessionData, r, w) if err != nil { common.WriteErrorResponsef(w, logger, http.StatusInternalServerError, "failed to save web auth login session: %s", err.Error()) return } logger.Debugf("Login begin success") jsonResponse(w, options, http.StatusOK) } func (o *Operation) finishLogin(w http.ResponseWriter, r *http.Request) { logger.Debugf("handling finish device login: %s", r.URL.String()) // get username userData, canProceed := o.getUserData(w, r, deviceCookieName) if !canProceed { return } deviceData, err := o.getDeviceInfo(userData.Sub) if err != nil { common.WriteErrorResponsef(w, logger, http.StatusBadRequest, "failed to get device data: %s", err.Error()) return } o.saveCookie(w, r, userData.Sub, userSubCookieName) // load the session data sessionData, err := o.store.session.GetWebauthn
{ op := &Operation{ store: &stores{ cookies: cookie.NewStore(config.Cookie), }, tlsConfig: config.TLSConfig, httpClient: &http.Client{Transport: &http.Transport{TLSClientConfig: config.TLSConfig}}, webauthn: config.Webauthn, walletDashboard: config.WalletDashboard, hubAuthURL: config.HubAuthURL, } var err error protocol.RegisterAttestationFormat("apple", ValidateAppleAttestation) op.store.storage, err = store.Open(config.Storage.Storage, deviceStoreName) if err != nil { return nil, fmt.Errorf("failed to open store: %w", err) }
identifier_body
operations.go
webauthn.WebAuthn HubAuthURL string } // StorageConfig holds storage config. type StorageConfig struct { Storage ariesstorage.Provider SessionStore ariesstorage.Provider } type stores struct { users *user.Store cookies cookie.Store storage ariesstorage.Store session *session.Store } type httpClient interface { Do(req *http.Request) (*http.Response, error) } // Operation implements OIDC operations. type Operation struct { store *stores walletDashboard string tlsConfig *tls.Config httpClient httpClient webauthn *webauthn.WebAuthn hubAuthURL string } // New returns a new Operation. func New(config *Config) (*Operation, error) { op := &Operation{ store: &stores{ cookies: cookie.NewStore(config.Cookie), }, tlsConfig: config.TLSConfig, httpClient: &http.Client{Transport: &http.Transport{TLSClientConfig: config.TLSConfig}}, webauthn: config.Webauthn, walletDashboard: config.WalletDashboard, hubAuthURL: config.HubAuthURL, } var err error protocol.RegisterAttestationFormat("apple", ValidateAppleAttestation) op.store.storage, err = store.Open(config.Storage.Storage, deviceStoreName) if err != nil { return nil, fmt.Errorf("failed to open store: %w", err) } op.store.session, err = session.NewStore() if err != nil { return nil, fmt.Errorf("failed to create web auth protocol session store: %w", err) } op.store.users, err = user.NewStore(config.Storage.Storage) if err != nil { return nil, fmt.Errorf("failed to open users store: %w", err) } return op, nil } // GetRESTHandlers get all controller API handler available for this service. func (o *Operation) GetRESTHandlers() []common.Handler { return []common.Handler{ common.NewHTTPHandler(registerBeginPath, http.MethodGet, o.beginRegistration), common.NewHTTPHandler(registerFinishPath, http.MethodPost, o.finishRegistration), common.NewHTTPHandler(loginBeginPath, http.MethodGet, o.beginLogin), common.NewHTTPHandler(loginFinishPath, http.MethodPost, o.finishLogin), } } func (o *Operation) beginRegistration(w http.ResponseWriter, r *http.Request) { logger.Debugf("handling device registration: %s", r.URL.String()) userData, canProceed := o.getUserData(w, r, userSubCookieName) if !canProceed { return } device := NewDevice(userData) webAuthnUser := protocol.UserEntity{
}, } registerOptions := func(credCreationOpts *protocol.PublicKeyCredentialCreationOptions) { credCreationOpts.User = webAuthnUser credCreationOpts.CredentialExcludeList = device.CredentialExcludeList() credCreationOpts.Attestation = protocol.PreferDirectAttestation } // generate PublicKeyCredentialCreationOptions, session data credentialParams, sessionData, err := o.webauthn.BeginRegistration( device, registerOptions, ) if err != nil { common.WriteErrorResponsef(w, logger, http.StatusInternalServerError, "failed to begin registration %s", err.Error()) return } // store session data as marshaled JSON err = o.store.session.SaveWebauthnSession(userData.Sub, sessionData, r, w) if err != nil { common.WriteErrorResponsef(w, logger, http.StatusInternalServerError, "failed to save web auth session %s", err.Error()) return } jsonResponse(w, credentialParams, http.StatusOK) logger.Debugf("Registration begins") } func (o *Operation) finishRegistration(w http.ResponseWriter, r *http.Request) { // nolint:funlen // not clean to split logger.Debugf("handling finish device registration: %s", r.URL.String()) userData, canProceed := o.getUserData(w, r, userSubCookieName) if !canProceed { return } sessionData, err := o.store.session.GetWebauthnSession(userData.Sub, r) if err != nil { common.WriteErrorResponsef(w, logger, http.StatusInternalServerError, "failed to get web auth session: %s", err.Error()) return } device := NewDevice(userData) // unfold webauthn.FinishRegistration, to access parsedResponse parsedResponse, err := protocol.ParseCredentialCreationResponse(r) if err != nil { common.WriteErrorResponsef(w, logger, http.StatusInternalServerError, "failed to finish registration: parsing ccr: %#v", err) return } credential, err := o.webauthn.CreateCredential(device, sessionData, parsedResponse) if err != nil { common.WriteErrorResponsef(w, logger, http.StatusInternalServerError, "failed to finish registration: cred: %#v", err) return } deviceCerts, ok := parsedResponse.Response.AttestationObject.AttStatement["x5c"].([]interface{}) if ok && len(deviceCerts) > 0 { err = o.requestDeviceValidation(r.Context(), userData.Sub, string(credential.Authenticator.AAGUID), deviceCerts) if err != nil { common.WriteErrorResponsef(w, logger, http.StatusInternalServerError, "failed to finish registration: %#v", err) return } } else { logger.Warnf("credential attestation of format '%s' has no certificates", parsedResponse.Response.AttestationObject.Format) } device.AddCredential(*credential) err = o.saveDeviceInfo(device) if err != nil { common.WriteErrorResponsef(w, logger, http.StatusInternalServerError, "failed to save device info: %s", err.Error()) return } o.saveCookie(w, r, userData.Sub, deviceCookieName) jsonResponse(w, credential, http.StatusOK) logger.Debugf("Registration success") } func (o *Operation) requestDeviceValidation(ctx context.Context, userSub, aaguid string, certs []interface{}) error { if len(certs) == 0 { return fmt.Errorf("missing certs") } var certPemList []string for _, certInterface := range certs { cert, ok := certInterface.([]byte) if !ok { return fmt.Errorf("can't cast certificate data to []byte") } certPemList = append(certPemList, string(pem.EncodeToMemory( &pem.Block{Bytes: cert, Type: "CERTIFICATE"}, ))) } postData, err := json.Marshal(&struct { X5c []string `json:"x5c"` Sub string `json:"sub"` Aaguid string `json:"aaguid"` }{ X5c: certPemList, Sub: userSub, Aaguid: aaguid, }) if err != nil { return fmt.Errorf("failed to marshal cert data: %w", err) } req, err := http.NewRequestWithContext(ctx, http.MethodPost, o.hubAuthURL+"/device", bytes.NewBuffer(postData)) if err != nil { return fmt.Errorf("failed to create request: %w", err) } _, _, err = common.SendHTTPRequest(req, o.httpClient, http.StatusOK, nil) if err != nil { return fmt.Errorf("failed response from hub-auth/device endpoint: %w", err) } return nil } func (o *Operation) beginLogin(w http.ResponseWriter, r *http.Request) { logger.Debugf("handling begin device login: %s", r.URL.String()) // get username userData, canProceed := o.getUserData(w, r, deviceCookieName) if !canProceed { return } deviceData, err := o.getDeviceInfo(userData.Sub) if err != nil { common.WriteErrorResponsef(w, logger, http.StatusBadRequest, "failed to get device data: %s", err.Error()) return } // generate PublicKeyCredentialRequestOptions, session data options, sessionData, err := o.webauthn.BeginLogin(deviceData) if err != nil { common.WriteErrorResponsef(w, logger, http.StatusInternalServerError, "failed to begin login: %s", err.Error()) return } // store session data as marshaled JSON err = o.store.session.SaveWebauthnSession(deviceData.ID, sessionData, r, w) if err != nil { common.WriteErrorResponsef(w, logger, http.StatusInternalServerError, "failed to save web auth login session: %s", err.Error()) return } logger.Debugf("Login begin success") jsonResponse(w, options, http.StatusOK) } func (o *Operation) finishLogin(w http.ResponseWriter, r *http.Request) { logger.Debugf("handling finish device login: %s", r.URL.String()) // get username userData, canProceed := o.getUserData(w, r, deviceCookieName) if !canProceed { return } deviceData, err := o.getDeviceInfo(userData.Sub) if err != nil { common.WriteErrorResponsef(w, logger, http.StatusBadRequest, "failed to get device data: %s", err.Error()) return } o.saveCookie(w, r, userData.Sub, userSubCookieName) // load the session data sessionData, err := o.store.session.GetWebauthnSession
ID: device.WebAuthnID(), DisplayName: device.WebAuthnDisplayName(), CredentialEntity: protocol.CredentialEntity{ Name: device.WebAuthnName(),
random_line_split
main.go
IP: C.GoString(&pAlarmer.sDeviceIP[0]), Command: int(command)} // switch int(command) { // case COMM_ALARM_V30: // log.Println("ALARM") // i.AlarmType = int(C.getalarminfo(pAlarmInfo).dwAlarmType) // motions <- i // break // case COMM_DEV_STATUS_CHANGED: // log.Printf("COMM_DEV_STATUS_CHANGED") // break // default: // log.Printf("Unknown Alarm [0x%x] !!!", command) // } // } // //export onmessage // func onmessage(command C.int, ip *C.char, data *C.char, ln C.uint) C.int { // i := AlarmItem{IP: C.GoString(ip), Command: int(command)} // switch int(command) { // case COMM_ALARM_V30: // i.AlarmType = int(C.getalarminfo(data).dwAlarmType) // motions <- i // break // case COMM_DEV_STATUS_CHANGED: // log.Printf("COMM_DEV_STATUS_CHANGED %s %s", C.GoString(ip), C.GoString(data)) // break // default: // log.Printf("Unknown Alarm [0x%x] %s %s !!!", command, C.GoString(ip), C.GoString(data)) // } // return 1 // } func bot() { videolist := map[string]string{} done := make(chan int, 1) done <- 1 admin := touser(strconv.Itoa(*adminParam)) b, err := tb.NewBot(tb.Settings{ Token: *tkeyParam, Poller: &tb.LongPoller{Timeout: 10 * time.Second}, }) if err != nil { log.Fatal(err) return } // var menu = &tb.ReplyMarkup{ResizeReplyKeyboard: true} // var btnSettings = menu.Data("⚙", "Settings") video := func(offset int, limit int) { //var v = hiklib.MotionVideos{} //C.MotionVideos{} mm, _ := b.Send(admin, "Fetch video from camera...") //C.HListVideo(C.int(user), &v) _, v := hiklib.HikListVideo(user) b.Edit(mm, strconv.Itoa(v.Count)+" video on camera.") if v.Count > 0 { txt := "" if offset == 0 { txt = fmt.Sprintf("First %d video:\n", limit) } else { txt = fmt.Sprintf("%d video from %d :\n", limit, offset) } for i := offset - 1; i < v.Count && i < offset+limit-1; i++ { dt := time.Date(v.Videos[i].From_year, time.Month(v.Videos[i].From_month), v.Videos[i].From_day, v.Videos[i].From_hour, v.Videos[i].From_min, v.Videos[i].From_sec, 0, time.UTC) todt := time.Date(v.Videos[i].To_year, time.Month(v.Videos[i].To_month), v.Videos[i].To_day, v.Videos[i].To_hour, v.Videos[i].To_min, v.Videos[i].To_sec, 0, time.UTC) txt = txt + "<b>" + dt.Format("2006-01-02 15:04:05") + " - " + todt.Format("15:04:05") + "</b> /dl_" + v.Videos[i].Filename + " \n" videolist[v.Videos[i].Filename] = dt.Format("2006-01-02/15:04:05") } if offset+limit < v.Count { txt = txt + fmt.Sprintf("<b>Next 10 video /video_%d_%d</b>\n", offset+limit, limit) } // menu.Inline(menu.Row(btnSettings)) // b.Send(admin, txt, &tb.SendOptions{ReplyMarkup: menu, ParseMode: tb.ModeHTML}) _, err = b.Send(admin, txt, &tb.SendOptions{ParseMode: tb.ModeHTML}) if err != nil { log.Println(err) } } } snapshot := func(mareas bool) { fname := filepath.Join(*datadirParam, fmt.Sprintf("%s.jpeg", uuid.Must(uuid.NewRandom()).String())) err := hiklib.HikCaptureImage(user, dev.ByStartChan, fname) if err > -1 { caption := "" if mareas { // var ma = C.MotionAreas{} // C.HMotionArea(C.int(user), &ma) _, ma := hiklib.HikMotionArea(user) col := color.RGBA{255, 0, 0, 128} var dst *image.RGBA var b image.Rectangle f, err := os.Open(fname) if err == nil { defer f.Close() img, _, err := image.Decode(f) if err == nil { b = img.Bounds() dst = image.NewRGBA(image.Rect(0, 0, b.Dx(), b.Dy())) draw.Draw(dst, b, img, b.Min, draw.Src) } } caption = caption + fmt.Sprintf("Image size %vx%v\n", b.Dx(), b.Dy()) for i := 0; i < 8; i++ { if ma.Areas[i].W > 0 && ma.Areas[i].H > 0 { x, y, w, h := int(float32(b.Dx())*float32(ma.Areas[i].X)), int(float32(b.Dy())*float32(ma.Areas[i].Y)), int(float32(b.Dx())*float32(ma.Areas[i].W)), int(float32(b.Dy())*float32(ma.Areas[i].H)) log.Printf("Area %v x:%v y:%v [%vx%v]\n", i+1, x, y, w, h) caption = caption + fmt.Sprintf("Area %v x:%v y:%v [%vx%v]\n", i+1, x, y, w, h) if dst != nil { Rect(dst, x, y, w, h, col) } } } if dst != nil { f.Close() f, err = os.Create(fname) if err == nil { defer f.Close() opt := jpeg.Options{ Quality: 100, } err = jpeg.Encode(f, dst, &opt) } } } //p := &tb.Photo{File: tb.FromDisk(fname)} //b.SendAlbum(admin, tb.Album{p}) p := &tb.Document{File: tb.FromDisk(fname), MIME: "image/jpeg", FileName: time.Now().Format(time.RFC3339) + ".jpeg"} if caption != "" { p.Caption = caption } b.Send(admin, p) os.Remove(fname) } else { b.Send(admin, fmt.Sprintf("Error get snapshot [%d].", err)) } } // On inline button pressed (callback) // b.Handle(&btnSettings, func(c *tb.Callback) { // b.Respond(c, &tb.CallbackResponse{Text: "testttt"}) // }) b.Handle("/video", func(m *tb.Message) { <-done if m.Sender.ID == *adminParam { video(1, 10) } done <- 1 }) b.Handle("/mareas", func(m *tb.Message) { <-done if m.Sender.ID == *adminParam { snapshot(true) } done <- 1 }) b.Handle("/snap", func(m *tb.Message) { <-done if m.Sender.ID == *adminParam { snapshot(false) } done <- 1 }) b.Handle("/reboot", func(m *tb.Message) { <-done if m.Sender.ID == *adminParam { res := hiklib.HikReboot(user) if res > 0 { b.Send(m.Sender, "Rebooting! Wait 10 sec.") time.Sleep(10 * time.Second) for Login() < 1 {
b.Send(m.Sender, "Camera online.") } else { b.Send(m.Sender, fmt.Sprintf("Fail [%d].", res)) } } done <- 1 }) b.Handle(tb.OnText, func(m *tb.Message) { <-done if m.Sender.ID == *adminParam { if strings.HasPrefix(m.Text, "/dl_") { mm, _ := b.Send(admin, "Loading...") log.Println(m.Text[4:]) if filename, ok := videolist[m.Text[4:]]; ok { os.MkdirAll(filepath.Join(*datadirParam
b.Send(m.Sender, "Wait 3 sec.") time.Sleep(3 * time.Second) }
conditional_block
main.go
() string { return string(t) } // //export onmessagev30 // func onmessagev30(command C.int, pAlarmer *C.NET_DVR_ALARMER, pAlarmInfo *C.char, dwBufLen C.uint, pUserData unsafe.Pointer) { // i := AlarmItem{IP: C.GoString(&pAlarmer.sDeviceIP[0]), Command: int(command)} // switch int(command) { // case COMM_ALARM_V30: // log.Println("ALARM") // i.AlarmType = int(C.getalarminfo(pAlarmInfo).dwAlarmType) // motions <- i // break // case COMM_DEV_STATUS_CHANGED: // log.Printf("COMM_DEV_STATUS_CHANGED") // break // default: // log.Printf("Unknown Alarm [0x%x] !!!", command) // } // } // //export onmessage // func onmessage(command C.int, ip *C.char, data *C.char, ln C.uint) C.int { // i := AlarmItem{IP: C.GoString(ip), Command: int(command)} // switch int(command) { // case COMM_ALARM_V30: // i.AlarmType = int(C.getalarminfo(data).dwAlarmType) // motions <- i // break // case COMM_DEV_STATUS_CHANGED: // log.Printf("COMM_DEV_STATUS_CHANGED %s %s", C.GoString(ip), C.GoString(data)) // break // default: // log.Printf("Unknown Alarm [0x%x] %s %s !!!", command, C.GoString(ip), C.GoString(data)) // } // return 1 // } func bot() { videolist := map[string]string{} done := make(chan int, 1) done <- 1 admin := touser(strconv.Itoa(*adminParam)) b, err := tb.NewBot(tb.Settings{ Token: *tkeyParam, Poller: &tb.LongPoller{Timeout: 10 * time.Second}, }) if err != nil { log.Fatal(err) return } // var menu = &tb.ReplyMarkup{ResizeReplyKeyboard: true} // var btnSettings = menu.Data("⚙", "Settings") video := func(offset int, limit int) { //var v = hiklib.MotionVideos{} //C.MotionVideos{} mm, _ := b.Send(admin, "Fetch video from camera...") //C.HListVideo(C.int(user), &v) _, v := hiklib.HikListVideo(user) b.Edit(mm, strconv.Itoa(v.Count)+" video on camera.") if v.Count > 0 { txt := "" if offset == 0 { txt = fmt.Sprintf("First %d video:\n", limit) } else { txt = fmt.Sprintf("%d video from %d :\n", limit, offset) } for i := offset - 1; i < v.Count && i < offset+limit-1; i++ { dt := time.Date(v.Videos[i].From_year, time.Month(v.Videos[i].From_month), v.Videos[i].From_day, v.Videos[i].From_hour, v.Videos[i].From_min, v.Videos[i].From_sec, 0, time.UTC) todt := time.Date(v.Videos[i].To_year, time.Month(v.Videos[i].To_month), v.Videos[i].To_day, v.Videos[i].To_hour, v.Videos[i].To_min, v.Videos[i].To_sec, 0, time.UTC) txt = txt + "<b>" + dt.Format("2006-01-02 15:04:05") + " - " + todt.Format("15:04:05") + "</b> /dl_" + v.Videos[i].Filename + " \n" videolist[v.Videos[i].Filename] = dt.Format("2006-01-02/15:04:05") } if offset+limit < v.Count { txt = txt + fmt.Sprintf("<b>Next 10 video /video_%d_%d</b>\n", offset+limit, limit) } // menu.Inline(menu.Row(btnSettings)) // b.Send(admin, txt, &tb.SendOptions{ReplyMarkup: menu, ParseMode: tb.ModeHTML}) _, err = b.Send(admin, txt, &tb.SendOptions{ParseMode: tb.ModeHTML}) if err != nil { log.Println(err) } } } snapshot := func(mareas bool) { fname := filepath.Join(*datadirParam, fmt.Sprintf("%s.jpeg", uuid.Must(uuid.NewRandom()).String())) err := hiklib.HikCaptureImage(user, dev.ByStartChan, fname) if err > -1 { caption := "" if mareas { // var ma = C.MotionAreas{} // C.HMotionArea(C.int(user), &ma) _, ma := hiklib.HikMotionArea(user) col := color.RGBA{255, 0, 0, 128} var dst *image.RGBA var b image.Rectangle f, err := os.Open(fname) if err == nil { defer f.Close() img, _, err := image.Decode(f) if err == nil { b = img.Bounds() dst = image.NewRGBA(image.Rect(0, 0, b.Dx(), b.Dy())) draw.Draw(dst, b, img, b.Min, draw.Src) } } caption = caption + fmt.Sprintf("Image size %vx%v\n", b.Dx(), b.Dy()) for i := 0; i < 8; i++ { if ma.Areas[i].W > 0 && ma.Areas[i].H > 0 { x, y, w, h := int(float32(b.Dx())*float32(ma.Areas[i].X)), int(float32(b.Dy())*float32(ma.Areas[i].Y)), int(float32(b.Dx())*float32(ma.Areas[i].W)), int(float32(b.Dy())*float32(ma.Areas[i].H)) log.Printf("Area %v x:%v y:%v [%vx%v]\n", i+1, x, y, w, h) caption = caption + fmt.Sprintf("Area %v x:%v y:%v [%vx%v]\n", i+1, x, y, w, h) if dst != nil { Rect(dst, x, y, w, h, col) } } } if dst != nil { f.Close() f, err = os.Create(fname) if err == nil { defer f.Close() opt := jpeg.Options{ Quality: 100, } err = jpeg.Encode(f, dst, &opt) } } } //p := &tb.Photo{File: tb.FromDisk(fname)} //b.SendAlbum(admin, tb.Album{p}) p := &tb.Document{File: tb.FromDisk(fname), MIME: "image/jpeg", FileName: time.Now().Format(time.RFC3339) + ".jpeg"} if caption != "" { p.Caption = caption } b.Send(admin, p) os.Remove(fname) } else { b.Send(admin, fmt.Sprintf("Error get snapshot [%d].", err)) } } // On inline button pressed (callback) // b.Handle(&btnSettings, func(c *tb.Callback) { // b.Respond(c, &tb.CallbackResponse{Text: "testttt"}) // }) b.Handle("/video", func(m *tb.Message) { <-done if m.Sender.ID == *adminParam { video(1, 10) } done <- 1 }) b.Handle("/mareas", func(m *tb.Message) { <-done if m.Sender.ID == *adminParam { snapshot(true) } done <- 1 }) b.Handle("/snap", func(m *tb.Message) { <-done if m.Sender.ID == *adminParam { snapshot(false) } done <- 1 }) b.Handle("/reboot", func(m *tb.Message) { <-done if m.Sender.ID == *adminParam { res := hiklib.HikReboot(user) if res > 0 { b.Send(m.Sender, "Rebooting! Wait 10 sec.") time.Sleep(10 * time.Second) for Login() < 1 { b.Send(m.Sender, "Wait 3 sec.") time.Sleep(3 * time.Second) } b.Send(m.Sender, "Camera online.") } else { b.Send(m.Sender, fmt.Sprintf("Fail [%d].", res)) } } done <- 1 }) b.Handle(tb.OnText, func(m *tb.Message) { <-done if m.Sender.ID
Recipient
identifier_name
main.go
: // log.Println("ALARM") // i.AlarmType = int(C.getalarminfo(pAlarmInfo).dwAlarmType) // motions <- i // break // case COMM_DEV_STATUS_CHANGED: // log.Printf("COMM_DEV_STATUS_CHANGED") // break // default: // log.Printf("Unknown Alarm [0x%x] !!!", command) // } // } // //export onmessage // func onmessage(command C.int, ip *C.char, data *C.char, ln C.uint) C.int { // i := AlarmItem{IP: C.GoString(ip), Command: int(command)} // switch int(command) { // case COMM_ALARM_V30: // i.AlarmType = int(C.getalarminfo(data).dwAlarmType) // motions <- i // break // case COMM_DEV_STATUS_CHANGED: // log.Printf("COMM_DEV_STATUS_CHANGED %s %s", C.GoString(ip), C.GoString(data)) // break // default: // log.Printf("Unknown Alarm [0x%x] %s %s !!!", command, C.GoString(ip), C.GoString(data)) // } // return 1 // } func bot() { videolist := map[string]string{} done := make(chan int, 1) done <- 1 admin := touser(strconv.Itoa(*adminParam)) b, err := tb.NewBot(tb.Settings{ Token: *tkeyParam, Poller: &tb.LongPoller{Timeout: 10 * time.Second}, }) if err != nil { log.Fatal(err) return } // var menu = &tb.ReplyMarkup{ResizeReplyKeyboard: true} // var btnSettings = menu.Data("⚙", "Settings") video := func(offset int, limit int) { //var v = hiklib.MotionVideos{} //C.MotionVideos{} mm, _ := b.Send(admin, "Fetch video from camera...") //C.HListVideo(C.int(user), &v) _, v := hiklib.HikListVideo(user) b.Edit(mm, strconv.Itoa(v.Count)+" video on camera.") if v.Count > 0 { txt := "" if offset == 0 { txt = fmt.Sprintf("First %d video:\n", limit) } else { txt = fmt.Sprintf("%d video from %d :\n", limit, offset) } for i := offset - 1; i < v.Count && i < offset+limit-1; i++ { dt := time.Date(v.Videos[i].From_year, time.Month(v.Videos[i].From_month), v.Videos[i].From_day, v.Videos[i].From_hour, v.Videos[i].From_min, v.Videos[i].From_sec, 0, time.UTC) todt := time.Date(v.Videos[i].To_year, time.Month(v.Videos[i].To_month), v.Videos[i].To_day, v.Videos[i].To_hour, v.Videos[i].To_min, v.Videos[i].To_sec, 0, time.UTC) txt = txt + "<b>" + dt.Format("2006-01-02 15:04:05") + " - " + todt.Format("15:04:05") + "</b> /dl_" + v.Videos[i].Filename + " \n" videolist[v.Videos[i].Filename] = dt.Format("2006-01-02/15:04:05") } if offset+limit < v.Count { txt = txt + fmt.Sprintf("<b>Next 10 video /video_%d_%d</b>\n", offset+limit, limit) } // menu.Inline(menu.Row(btnSettings)) // b.Send(admin, txt, &tb.SendOptions{ReplyMarkup: menu, ParseMode: tb.ModeHTML}) _, err = b.Send(admin, txt, &tb.SendOptions{ParseMode: tb.ModeHTML}) if err != nil { log.Println(err) } } } snapshot := func(mareas bool) { fname := filepath.Join(*datadirParam, fmt.Sprintf("%s.jpeg", uuid.Must(uuid.NewRandom()).String())) err := hiklib.HikCaptureImage(user, dev.ByStartChan, fname) if err > -1 { caption := "" if mareas { // var ma = C.MotionAreas{} // C.HMotionArea(C.int(user), &ma) _, ma := hiklib.HikMotionArea(user) col := color.RGBA{255, 0, 0, 128} var dst *image.RGBA var b image.Rectangle f, err := os.Open(fname) if err == nil { defer f.Close() img, _, err := image.Decode(f) if err == nil { b = img.Bounds() dst = image.NewRGBA(image.Rect(0, 0, b.Dx(), b.Dy())) draw.Draw(dst, b, img, b.Min, draw.Src) } } caption = caption + fmt.Sprintf("Image size %vx%v\n", b.Dx(), b.Dy()) for i := 0; i < 8; i++ { if ma.Areas[i].W > 0 && ma.Areas[i].H > 0 { x, y, w, h := int(float32(b.Dx())*float32(ma.Areas[i].X)), int(float32(b.Dy())*float32(ma.Areas[i].Y)), int(float32(b.Dx())*float32(ma.Areas[i].W)), int(float32(b.Dy())*float32(ma.Areas[i].H)) log.Printf("Area %v x:%v y:%v [%vx%v]\n", i+1, x, y, w, h) caption = caption + fmt.Sprintf("Area %v x:%v y:%v [%vx%v]\n", i+1, x, y, w, h) if dst != nil { Rect(dst, x, y, w, h, col) } } } if dst != nil { f.Close() f, err = os.Create(fname) if err == nil { defer f.Close() opt := jpeg.Options{ Quality: 100, } err = jpeg.Encode(f, dst, &opt) } } } //p := &tb.Photo{File: tb.FromDisk(fname)} //b.SendAlbum(admin, tb.Album{p}) p := &tb.Document{File: tb.FromDisk(fname), MIME: "image/jpeg", FileName: time.Now().Format(time.RFC3339) + ".jpeg"} if caption != "" { p.Caption = caption } b.Send(admin, p) os.Remove(fname) } else { b.Send(admin, fmt.Sprintf("Error get snapshot [%d].", err)) } } // On inline button pressed (callback) // b.Handle(&btnSettings, func(c *tb.Callback) { // b.Respond(c, &tb.CallbackResponse{Text: "testttt"}) // }) b.Handle("/video", func(m *tb.Message) { <-done if m.Sender.ID == *adminParam { video(1, 10) } done <- 1 }) b.Handle("/mareas", func(m *tb.Message) { <-done if m.Sender.ID == *adminParam { snapshot(true) } done <- 1 }) b.Handle("/snap", func(m *tb.Message) { <-done if m.Sender.ID == *adminParam { snapshot(false) } done <- 1 }) b.Handle("/reboot", func(m *tb.Message) { <-done if m.Sender.ID == *adminParam { res := hiklib.HikReboot(user) if res > 0 { b.Send(m.Sender, "Rebooting! Wait 10 sec.") time.Sleep(10 * time.Second) for Login() < 1 { b.Send(m.Sender, "Wait 3 sec.") time.Sleep(3 * time.Second) } b.Send(m.Sender, "Camera online.") } else { b.Send(m.Sender, fmt.Sprintf("Fail [%d].", res)) } } done <- 1 }) b.Handle(tb.OnText, func(m *tb.Message) { <-done if m.Sender.ID == *adminParam { if strings.HasPrefix(m.Text, "/dl_") { mm, _ := b.Send(admin, "Loading...") log.Println(m.Text[4:])
if filename, ok := videolist[m.Text[4:]]; ok { os.MkdirAll(filepath.Join(*datadirParam, strings.Split(filename, "/")[0]), 0755) fname := filepath.Join(*datadirParam, filename+".mpeg") p := &tb.Video{}
random_line_split
main.go
_, ma := hiklib.HikMotionArea(user) col := color.RGBA{255, 0, 0, 128} var dst *image.RGBA var b image.Rectangle f, err := os.Open(fname) if err == nil { defer f.Close() img, _, err := image.Decode(f) if err == nil { b = img.Bounds() dst = image.NewRGBA(image.Rect(0, 0, b.Dx(), b.Dy())) draw.Draw(dst, b, img, b.Min, draw.Src) } } caption = caption + fmt.Sprintf("Image size %vx%v\n", b.Dx(), b.Dy()) for i := 0; i < 8; i++ { if ma.Areas[i].W > 0 && ma.Areas[i].H > 0 { x, y, w, h := int(float32(b.Dx())*float32(ma.Areas[i].X)), int(float32(b.Dy())*float32(ma.Areas[i].Y)), int(float32(b.Dx())*float32(ma.Areas[i].W)), int(float32(b.Dy())*float32(ma.Areas[i].H)) log.Printf("Area %v x:%v y:%v [%vx%v]\n", i+1, x, y, w, h) caption = caption + fmt.Sprintf("Area %v x:%v y:%v [%vx%v]\n", i+1, x, y, w, h) if dst != nil { Rect(dst, x, y, w, h, col) } } } if dst != nil { f.Close() f, err = os.Create(fname) if err == nil { defer f.Close() opt := jpeg.Options{ Quality: 100, } err = jpeg.Encode(f, dst, &opt) } } } //p := &tb.Photo{File: tb.FromDisk(fname)} //b.SendAlbum(admin, tb.Album{p}) p := &tb.Document{File: tb.FromDisk(fname), MIME: "image/jpeg", FileName: time.Now().Format(time.RFC3339) + ".jpeg"} if caption != "" { p.Caption = caption } b.Send(admin, p) os.Remove(fname) } else { b.Send(admin, fmt.Sprintf("Error get snapshot [%d].", err)) } } // On inline button pressed (callback) // b.Handle(&btnSettings, func(c *tb.Callback) { // b.Respond(c, &tb.CallbackResponse{Text: "testttt"}) // }) b.Handle("/video", func(m *tb.Message) { <-done if m.Sender.ID == *adminParam { video(1, 10) } done <- 1 }) b.Handle("/mareas", func(m *tb.Message) { <-done if m.Sender.ID == *adminParam { snapshot(true) } done <- 1 }) b.Handle("/snap", func(m *tb.Message) { <-done if m.Sender.ID == *adminParam { snapshot(false) } done <- 1 }) b.Handle("/reboot", func(m *tb.Message) { <-done if m.Sender.ID == *adminParam { res := hiklib.HikReboot(user) if res > 0 { b.Send(m.Sender, "Rebooting! Wait 10 sec.") time.Sleep(10 * time.Second) for Login() < 1 { b.Send(m.Sender, "Wait 3 sec.") time.Sleep(3 * time.Second) } b.Send(m.Sender, "Camera online.") } else { b.Send(m.Sender, fmt.Sprintf("Fail [%d].", res)) } } done <- 1 }) b.Handle(tb.OnText, func(m *tb.Message) { <-done if m.Sender.ID == *adminParam { if strings.HasPrefix(m.Text, "/dl_") { mm, _ := b.Send(admin, "Loading...") log.Println(m.Text[4:]) if filename, ok := videolist[m.Text[4:]]; ok { os.MkdirAll(filepath.Join(*datadirParam, strings.Split(filename, "/")[0]), 0755) fname := filepath.Join(*datadirParam, filename+".mpeg") p := &tb.Video{} if _, err := os.Stat(fname); os.IsNotExist(err) { opts := ffmpeg.KwArgs{ "format": "mp4", //"fs": strconv.Itoa(*previewsizeParam), "vcodec": "copy", //"libx264", "preset": "ultrafast", "acodec": "none", "movflags": "+faststart", } // C.HSaveFile(C.int(user), C.CString(m.Text[4:]), C.CString(fname)) hiklib.HikSaveFile(user, m.Text[4:], fname) b.Edit(mm, "Probing...") f, err := ffmpeg.Probe(fname) var fjson FFProbe err = json.Unmarshal([]byte(f), &fjson) if err == nil { // b.Send(admin, f) p.Width = int(fjson.Streams[0]["width"].(float64)) p.Height = int(fjson.Streams[0]["height"].(float64)) if sz, err := strconv.Atoi(fjson.Format["size"].(string)); err == nil { if sz > *previewsizeParam { if s, err := strconv.ParseFloat(fjson.Format["duration"].(string), 64); err == nil { opts["vcodec"] = "libx264" opts["b"] = strconv.Itoa(int(math.Floor(float64(*previewsizeParam)/math.Floor(s)) * 8)) p.Width = int(math.Round(float64(p.Width) / float64(*zParam))) p.Height = int(math.Round(float64(p.Height) / float64(*zParam))) opts["vf"] = fmt.Sprintf("scale=%d:%d", p.Width, p.Height) //opts["vf"] = "scale=iw/2:ih/2" log.Println("Change bitrate", opts["b"]) } } } } else { log.Println(err) } b.Edit(mm, "Transcoding ...") err = ffmpeg.Input(fname). Output(fname+".mp4", opts).OverWriteOutput(). Run() if err != nil { log.Println(err) } } else { b.Edit(mm, "Probing...") f, err := ffmpeg.Probe(fname) var fjson FFProbe err = json.Unmarshal([]byte(f), &fjson) if err == nil { p.Width = int(fjson.Streams[0]["width"].(float64)) p.Height = int(fjson.Streams[0]["height"].(float64)) } else { log.Println(err) } } b.Edit(mm, "Sending...") p.File = tb.FromDisk(fname + ".mp4") p.FileName = "video.mp4" b.Send(admin, p) b.Delete(mm) if *datadirParam == "/tmp" { os.Remove(fname) os.Remove(fname + ".mp4") } } else { b.Send(admin, "Not found.") } } else if strings.HasPrefix(m.Text, "/video_") { args := strings.Split(m.Text[7:], "_") if len(args) > 1 { offset, err := strconv.Atoi(args[0]) if err == nil { limit, err := strconv.Atoi(args[1]) if err == nil && offset > -1 && limit > 0 { video(offset, limit) } } } } } done <- 1 }) go func() { for { i := <-motions if i.AlarmType == 3 { snapshot(false) } else { log.Println(i) } } }() b.Send(admin, "Bot restart!") b.Start() } func Login() int {
// user = C.HLogin(C.CString(*ipParam), C.CString(*userParam), C.CString(*passParam), &dev) user, dev = hiklib.HikLogin(*ipParam, *userParam, *passParam) if int(user) > -1 { if *x1Param { hiklib.HikOnAlarmV30(user, *alarmParam, func(item hiklib.AlarmItem) { motions <- item }) } else { hiklib.HikOnAlarm(user, *alarmParam, func(item hiklib.AlarmItem) { motions <- item }) } return int(user) } else { return int(user) } }
identifier_body
sync_keys.py
2_ALPHABET[((data[idx + 2] & 0x0F) << 1) | ((data[idx + 3] & 0x80) >> 7)] result += ZBASE32_ALPHABET[(data[idx + 3] & 0x7C) >> 2] if idx + 4 == len(data): result += ZBASE32_ALPHABET[(data[idx + 3] & 0x03) << 3] break result += ZBASE32_ALPHABET[((data[idx + 3] & 0x03) << 3) | ((data[idx + 4] & 0xE0) >> 5)] result += ZBASE32_ALPHABET[data[idx + 4] & 0x1F] assert len(result) == (len(data) * 8 + 4) // 5 return result def zbase32_decode(text: str) -> bytes: """Decode some data using the z-base-32 encoding""" result = bytearray(len(text) * 5 // 8) cur_byte = 0 cur_numbits = 0 idx = 0 for character in text: value = ZBASE32_ALPHABET_REV[character] cur_byte = (cur_byte << 5) | value cur_numbits += 5 if cur_numbits >= 8: cur_numbits -= 8 result[idx] = cur_byte >> cur_numbits idx += 1 cur_byte &= (1 << cur_numbits) - 1 return bytes(result) def get_wkd_advanced_url(email: str) -> str: """Craft an URL for WKD advanced method""" local, domain = email.split("@", 1) domain = domain.lower() local_sha1 = hashlib.sha1(local.lower().encode("ascii")).digest() local_b32 = zbase32_encode(local_sha1) params = urllib.parse.urlencode({"l": local}) return f"https://openpgpkey.{domain}/.well-known/openpgpkey/{domain}/hu/{local_b32}?{params}" def get_wkd_direct_url(email: str) -> str: """Craft an URL for WKD direct method""" local, domain = email.split("@", 1) domain = domain.lower() local_sha1 = hashlib.sha1(local.lower().encode("ascii")).digest() local_b32 = zbase32_encode(local_sha1) params = urllib.parse.urlencode({"l": local}) return f"https://{domain}/.well-known/openpgpkey/hu/{local_b32}?{params}" def self_check() -> None: """Verify that the algorithm computing WKD URLs work""" assert len(ZBASE32_ALPHABET) == 32 # Test vector from https://github.com/matusf/z-base-32/blob/0.1.2/src/lib.rs assert zbase32_encode(b"asdasd") == "cf3seamuco" assert zbase32_decode("cf3seamuco") == b"asdasd" # Test vector from https://www.uriports.com/blog/setting-up-openpgp-web-key-directory/ # assert zbase32_encode(hashlib.sha1(b"yourmail").digest()) == "hacabazoakmnagxwmkjerb9yehuwehbm" # -> this hash is wrong, and I don't know what username gives the SHA1 # e61980e2f0c2962c19f45a928207e0472744702b # Test vector from https://metacode.biz/openpgp/web-key-directory assert zbase32_encode(hashlib.sha1(b"test-wkd").digest()) == "4hg7tescnttreaouu4z1izeuuyibwww1" # Test vector from https://datatracker.ietf.org/doc/draft-koch-openpgp-webkey-service/ assert ( get_wkd_advanced_url("Joe.Doe@Example.ORG") == "https://openpgpkey.example.org/.well-known/openpgpkey/example.org/hu/iy9q119eutrkn8s1mk4r39qejnbu3n5q?l=Joe.Doe" # noqa ) assert ( get_wkd_direct_url("Joe.Doe@Example.ORG") == "https://example.org/.well-known/openpgpkey/hu/iy9q119eutrkn8s1mk4r39qejnbu3n5q?l=Joe.Doe" ) # Test vector from https://wiki.gnupg.org/WKD assert ( get_wkd_direct_url("bernhard.reiter@intevation.de") == "https://intevation.de/.well-known/openpgpkey/hu/it5sewh54rxz33fwmr8u6dy4bbz8itz4?l=bernhard.reiter" ) def get_pgp_key_id(raw_key: bytes) -> str: """Get the identifier of a key, using GnuPG""" # Flush stdout and stderr to prevent interleaving messages from a subprocess sys.stdout.flush() sys.stderr.flush() with tempfile.TemporaryDirectory(prefix="gnupghome") as tmpdir: # Create an empty public keyring to avoid a GnuPG message with (Path(tmpdir) / "pubring.kbx").open("wb"): pass output = subprocess.check_output( ("gpg", "--list-packets"), input=raw_key, env={ "GNUPGHOME": tmpdir, "HOME": tmpdir, }, ) keyid_index = output.index(b"keyid: ") + 7 keyid_end_index = output.index(b"\n", keyid_index) key_id = output[keyid_index:keyid_end_index].decode("ascii") assert len(key_id) == 16 assert all(c in "0123456789ABCDEF" for c in key_id) return key_id def gpg_recv_key(key_id: str) -> bytes: """Receive a key using GnuPG using Ubuntu keyserver https://keyserver.ubuntu.com/""" # Flush stdout and stderr to prevent interleaving messages from a subprocess sys.stdout.flush() sys.stderr.flush() with tempfile.TemporaryDirectory(prefix="gnupghome") as tmpdir: # Create an empty public keyring to avoid a GnuPG message with (Path(tmpdir) / "pubring.kbx").open("wb"): pass with (Path(tmpdir) / "trustdb.gpg").open("wb"): pass # Retry several times, as sometimes the command fails with: # gpg: keyserver receive failed: No data try_count = 0 while 1: try: subprocess.check_output( ("gpg", "--keyserver", "hkps://keyserver.ubuntu.com", "--recv-keys", key_id), input=b"", env={ "GNUPGHOME": tmpdir, "HOME": tmpdir, }, ) break except subprocess.CalledProcessError: if try_count >= 10: raise print(f"Receiving key {key_id} failed [{try_count}], retrying...") time.sleep(1) try_count += 1 raw_key = subprocess.check_output( ("gpg", "--export", key_id), env={ "GNUPGHOME": tmpdir, "HOME": tmpdir, }, ) return raw_key def sync_keys(keys_path: Path) -> None: """Sync all the keys and refresh the given file""" file_lines = [] with keys_path.open("r") as fkeys: for line in fkeys: line = line.strip() if not line or line.startswith("#"): # Keep comments and empty lines file_lines.append(line) continue fields = line.split(" ") if len(fields) < 2: raise ValueError(f"Unexpected line: {line!r}") current_key_id = fields[0] email = fields[1] raw_key = None wkd_url = None key_comment = None if "@" in email:
email = email.lower() # Download the key using WKD wkd_url = get_wkd_advanced_url(email) try: with urllib.request.urlopen(wkd_url) as response: raw_key = response.read() except urllib.error.URLError: pass else: print(f"Downloaded key for {email} from {wkd_url}") key_comment = wkd_url # Try the direct method when the advanced one failed # Ignore domains which have issues in their configuration if raw_key is None and not email.endswith("@att.net"): wkd_url = get_wkd_direct_url(email) raw_key = None try: with urllib.request.urlopen(wkd_url) as response:
conditional_block
sync_keys.py
range(8): crc <<= 1 if (crc & 0x1000000) != 0: crc ^= 0x1864CFB assert 0 <= crc <= 0xFFFFFF return crc def opgp_crc24_b64(data: bytes) -> str: """Computes the CRC24 used by OpenPGP Message Format, encoded in base64""" crc = opgp_crc24(data) return "=" + base64.b64encode(crc.to_bytes(3, "big")).decode("ascii") def unarmor_gpg(armored: Union[bytes, str]) -> bytes: if isinstance(armored, str): lines = armored.splitlines() else: lines = armored.decode("ascii").splitlines() if lines[0] != "-----BEGIN PGP PUBLIC KEY BLOCK-----": raise ValueError(f"unexpected first line {lines[0]!r}") if lines[-1] != "-----END PGP PUBLIC KEY BLOCK-----": raise ValueError(f"unexpected last line {lines[0]!r}") first_empty_line = lines.index("") data = base64.b64decode("".join(lines[first_empty_line + 1:-2])) computed_checksum = opgp_crc24_b64(data) if lines[-2] != computed_checksum: raise ValueError(f"unexpected checksum {lines[-2]!r}, expected {computed_checksum}") return data def zbase32_encode(data: bytes) -> str: """Encode some data using the z-base-32 encoding This encoding is specified for ZRTP protocol in https://www.rfc-editor.org/rfc/rfc6189.html#section-5.1.6 and used an alphabet described as: This base32 encoding scheme differs from RFC 4648, and was designed (by Bryce Wilcox-O'Hearn) to represent bit sequences in a form that is convenient for human users to manipulate with minimal ambiguity. The unusually permuted character ordering was designed for other applications that use bit sequences that do not end on quintet boundaries. This hash is used by WKD and can be computed by GnuPG: gpg --with-wkd-hash -k yourmail@example.org """ result = "" for idx in range(0, len(data), 5): result += ZBASE32_ALPHABET[(data[idx] & 0xF8) >> 3] if idx + 1 == len(data): result += ZBASE32_ALPHABET[(data[idx] & 0x07) << 2] break result += ZBASE32_ALPHABET[((data[idx] & 0x07) << 2) | ((data[idx + 1] & 0xC0) >> 6)] result += ZBASE32_ALPHABET[(data[idx + 1] & 0x3E) >> 1] if idx + 2 == len(data): result += ZBASE32_ALPHABET[(data[idx + 1] & 0x01) << 4] break result += ZBASE32_ALPHABET[((data[idx + 1] & 0x01) << 4) | ((data[idx + 2] & 0xF0) >> 4)] if idx + 3 == len(data): result += ZBASE32_ALPHABET[(data[idx + 2] & 0x0F) << 1] break result += ZBASE32_ALPHABET[((data[idx + 2] & 0x0F) << 1) | ((data[idx + 3] & 0x80) >> 7)] result += ZBASE32_ALPHABET[(data[idx + 3] & 0x7C) >> 2] if idx + 4 == len(data): result += ZBASE32_ALPHABET[(data[idx + 3] & 0x03) << 3] break result += ZBASE32_ALPHABET[((data[idx + 3] & 0x03) << 3) | ((data[idx + 4] & 0xE0) >> 5)] result += ZBASE32_ALPHABET[data[idx + 4] & 0x1F] assert len(result) == (len(data) * 8 + 4) // 5 return result def zbase32_decode(text: str) -> bytes: """Decode some data using the z-base-32 encoding""" result = bytearray(len(text) * 5 // 8) cur_byte = 0 cur_numbits = 0 idx = 0 for character in text: value = ZBASE32_ALPHABET_REV[character] cur_byte = (cur_byte << 5) | value cur_numbits += 5 if cur_numbits >= 8: cur_numbits -= 8 result[idx] = cur_byte >> cur_numbits idx += 1 cur_byte &= (1 << cur_numbits) - 1 return bytes(result) def get_wkd_advanced_url(email: str) -> str: """Craft an URL for WKD advanced method""" local, domain = email.split("@", 1) domain = domain.lower() local_sha1 = hashlib.sha1(local.lower().encode("ascii")).digest() local_b32 = zbase32_encode(local_sha1) params = urllib.parse.urlencode({"l": local}) return f"https://openpgpkey.{domain}/.well-known/openpgpkey/{domain}/hu/{local_b32}?{params}" def get_wkd_direct_url(email: str) -> str: """Craft an URL for WKD direct method""" local, domain = email.split("@", 1) domain = domain.lower() local_sha1 = hashlib.sha1(local.lower().encode("ascii")).digest() local_b32 = zbase32_encode(local_sha1) params = urllib.parse.urlencode({"l": local}) return f"https://{domain}/.well-known/openpgpkey/hu/{local_b32}?{params}" def
() -> None: """Verify that the algorithm computing WKD URLs work""" assert len(ZBASE32_ALPHABET) == 32 # Test vector from https://github.com/matusf/z-base-32/blob/0.1.2/src/lib.rs assert zbase32_encode(b"asdasd") == "cf3seamuco" assert zbase32_decode("cf3seamuco") == b"asdasd" # Test vector from https://www.uriports.com/blog/setting-up-openpgp-web-key-directory/ # assert zbase32_encode(hashlib.sha1(b"yourmail").digest()) == "hacabazoakmnagxwmkjerb9yehuwehbm" # -> this hash is wrong, and I don't know what username gives the SHA1 # e61980e2f0c2962c19f45a928207e0472744702b # Test vector from https://metacode.biz/openpgp/web-key-directory assert zbase32_encode(hashlib.sha1(b"test-wkd").digest()) == "4hg7tescnttreaouu4z1izeuuyibwww1" # Test vector from https://datatracker.ietf.org/doc/draft-koch-openpgp-webkey-service/ assert ( get_wkd_advanced_url("Joe.Doe@Example.ORG") == "https://openpgpkey.example.org/.well-known/openpgpkey/example.org/hu/iy9q119eutrkn8s1mk4r39qejnbu3n5q?l=Joe.Doe" # noqa ) assert ( get_wkd_direct_url("Joe.Doe@Example.ORG") == "https://example.org/.well-known/openpgpkey/hu/iy9q119eutrkn8s1mk4r39qejnbu3n5q?l=Joe.Doe" ) # Test vector from https://wiki.gnupg.org/WKD assert ( get_wkd_direct_url("bernhard.reiter@intevation.de") == "https://intevation.de/.well-known/openpgpkey/hu/it5sewh54rxz33fwmr8u6dy4bbz8itz4?l=bernhard.reiter" ) def get_pgp_key_id(raw_key: bytes) -> str: """Get the identifier of a key, using GnuPG""" # Flush stdout and stderr to prevent interleaving messages from a subprocess sys.stdout.flush() sys.stderr.flush() with tempfile.TemporaryDirectory(prefix="gnupghome") as tmpdir: # Create an empty public keyring to avoid a GnuPG message
self_check
identifier_name
sync_keys.py
1] & 0x3E) >> 1] if idx + 2 == len(data): result += ZBASE32_ALPHABET[(data[idx + 1] & 0x01) << 4] break result += ZBASE32_ALPHABET[((data[idx + 1] & 0x01) << 4) | ((data[idx + 2] & 0xF0) >> 4)] if idx + 3 == len(data): result += ZBASE32_ALPHABET[(data[idx + 2] & 0x0F) << 1] break result += ZBASE32_ALPHABET[((data[idx + 2] & 0x0F) << 1) | ((data[idx + 3] & 0x80) >> 7)] result += ZBASE32_ALPHABET[(data[idx + 3] & 0x7C) >> 2] if idx + 4 == len(data): result += ZBASE32_ALPHABET[(data[idx + 3] & 0x03) << 3] break result += ZBASE32_ALPHABET[((data[idx + 3] & 0x03) << 3) | ((data[idx + 4] & 0xE0) >> 5)] result += ZBASE32_ALPHABET[data[idx + 4] & 0x1F] assert len(result) == (len(data) * 8 + 4) // 5 return result def zbase32_decode(text: str) -> bytes: """Decode some data using the z-base-32 encoding""" result = bytearray(len(text) * 5 // 8) cur_byte = 0 cur_numbits = 0 idx = 0 for character in text: value = ZBASE32_ALPHABET_REV[character] cur_byte = (cur_byte << 5) | value cur_numbits += 5 if cur_numbits >= 8: cur_numbits -= 8 result[idx] = cur_byte >> cur_numbits idx += 1 cur_byte &= (1 << cur_numbits) - 1 return bytes(result) def get_wkd_advanced_url(email: str) -> str: """Craft an URL for WKD advanced method""" local, domain = email.split("@", 1) domain = domain.lower() local_sha1 = hashlib.sha1(local.lower().encode("ascii")).digest() local_b32 = zbase32_encode(local_sha1) params = urllib.parse.urlencode({"l": local}) return f"https://openpgpkey.{domain}/.well-known/openpgpkey/{domain}/hu/{local_b32}?{params}" def get_wkd_direct_url(email: str) -> str: """Craft an URL for WKD direct method""" local, domain = email.split("@", 1) domain = domain.lower() local_sha1 = hashlib.sha1(local.lower().encode("ascii")).digest() local_b32 = zbase32_encode(local_sha1) params = urllib.parse.urlencode({"l": local}) return f"https://{domain}/.well-known/openpgpkey/hu/{local_b32}?{params}" def self_check() -> None: """Verify that the algorithm computing WKD URLs work""" assert len(ZBASE32_ALPHABET) == 32 # Test vector from https://github.com/matusf/z-base-32/blob/0.1.2/src/lib.rs assert zbase32_encode(b"asdasd") == "cf3seamuco" assert zbase32_decode("cf3seamuco") == b"asdasd" # Test vector from https://www.uriports.com/blog/setting-up-openpgp-web-key-directory/ # assert zbase32_encode(hashlib.sha1(b"yourmail").digest()) == "hacabazoakmnagxwmkjerb9yehuwehbm" # -> this hash is wrong, and I don't know what username gives the SHA1 # e61980e2f0c2962c19f45a928207e0472744702b # Test vector from https://metacode.biz/openpgp/web-key-directory assert zbase32_encode(hashlib.sha1(b"test-wkd").digest()) == "4hg7tescnttreaouu4z1izeuuyibwww1" # Test vector from https://datatracker.ietf.org/doc/draft-koch-openpgp-webkey-service/ assert ( get_wkd_advanced_url("Joe.Doe@Example.ORG") == "https://openpgpkey.example.org/.well-known/openpgpkey/example.org/hu/iy9q119eutrkn8s1mk4r39qejnbu3n5q?l=Joe.Doe" # noqa ) assert ( get_wkd_direct_url("Joe.Doe@Example.ORG") == "https://example.org/.well-known/openpgpkey/hu/iy9q119eutrkn8s1mk4r39qejnbu3n5q?l=Joe.Doe" ) # Test vector from https://wiki.gnupg.org/WKD assert ( get_wkd_direct_url("bernhard.reiter@intevation.de") == "https://intevation.de/.well-known/openpgpkey/hu/it5sewh54rxz33fwmr8u6dy4bbz8itz4?l=bernhard.reiter" ) def get_pgp_key_id(raw_key: bytes) -> str: """Get the identifier of a key, using GnuPG""" # Flush stdout and stderr to prevent interleaving messages from a subprocess sys.stdout.flush() sys.stderr.flush() with tempfile.TemporaryDirectory(prefix="gnupghome") as tmpdir: # Create an empty public keyring to avoid a GnuPG message with (Path(tmpdir) / "pubring.kbx").open("wb"): pass output = subprocess.check_output( ("gpg", "--list-packets"), input=raw_key, env={ "GNUPGHOME": tmpdir, "HOME": tmpdir, }, ) keyid_index = output.index(b"keyid: ") + 7 keyid_end_index = output.index(b"\n", keyid_index) key_id = output[keyid_index:keyid_end_index].decode("ascii") assert len(key_id) == 16 assert all(c in "0123456789ABCDEF" for c in key_id) return key_id def gpg_recv_key(key_id: str) -> bytes: """Receive a key using GnuPG using Ubuntu keyserver https://keyserver.ubuntu.com/""" # Flush stdout and stderr to prevent interleaving messages from a subprocess sys.stdout.flush() sys.stderr.flush() with tempfile.TemporaryDirectory(prefix="gnupghome") as tmpdir: # Create an empty public keyring to avoid a GnuPG message with (Path(tmpdir) / "pubring.kbx").open("wb"): pass with (Path(tmpdir) / "trustdb.gpg").open("wb"): pass # Retry several times, as sometimes the command fails with: # gpg: keyserver receive failed: No data try_count = 0 while 1: try: subprocess.check_output( ("gpg", "--keyserver", "hkps://keyserver.ubuntu.com", "--recv-keys", key_id), input=b"", env={ "GNUPGHOME": tmpdir, "HOME": tmpdir, }, ) break except subprocess.CalledProcessError: if try_count >= 10: raise print(f"Receiving key {key_id} failed [{try_count}], retrying...") time.sleep(1) try_count += 1 raw_key = subprocess.check_output( ("gpg", "--export", key_id), env={ "GNUPGHOME": tmpdir, "HOME": tmpdir, }, ) return raw_key def sync_keys(keys_path: Path) -> None:
"""Sync all the keys and refresh the given file""" file_lines = [] with keys_path.open("r") as fkeys: for line in fkeys: line = line.strip() if not line or line.startswith("#"): # Keep comments and empty lines file_lines.append(line) continue fields = line.split(" ") if len(fields) < 2: raise ValueError(f"Unexpected line: {line!r}") current_key_id = fields[0] email = fields[1] raw_key = None wkd_url = None key_comment = None if "@" in email: email = email.lower()
identifier_body
sync_keys.py
) -> str: """Craft an URL for WKD direct method""" local, domain = email.split("@", 1) domain = domain.lower() local_sha1 = hashlib.sha1(local.lower().encode("ascii")).digest() local_b32 = zbase32_encode(local_sha1) params = urllib.parse.urlencode({"l": local}) return f"https://{domain}/.well-known/openpgpkey/hu/{local_b32}?{params}" def self_check() -> None: """Verify that the algorithm computing WKD URLs work""" assert len(ZBASE32_ALPHABET) == 32 # Test vector from https://github.com/matusf/z-base-32/blob/0.1.2/src/lib.rs assert zbase32_encode(b"asdasd") == "cf3seamuco" assert zbase32_decode("cf3seamuco") == b"asdasd" # Test vector from https://www.uriports.com/blog/setting-up-openpgp-web-key-directory/ # assert zbase32_encode(hashlib.sha1(b"yourmail").digest()) == "hacabazoakmnagxwmkjerb9yehuwehbm" # -> this hash is wrong, and I don't know what username gives the SHA1 # e61980e2f0c2962c19f45a928207e0472744702b # Test vector from https://metacode.biz/openpgp/web-key-directory assert zbase32_encode(hashlib.sha1(b"test-wkd").digest()) == "4hg7tescnttreaouu4z1izeuuyibwww1" # Test vector from https://datatracker.ietf.org/doc/draft-koch-openpgp-webkey-service/ assert ( get_wkd_advanced_url("Joe.Doe@Example.ORG") == "https://openpgpkey.example.org/.well-known/openpgpkey/example.org/hu/iy9q119eutrkn8s1mk4r39qejnbu3n5q?l=Joe.Doe" # noqa ) assert ( get_wkd_direct_url("Joe.Doe@Example.ORG") == "https://example.org/.well-known/openpgpkey/hu/iy9q119eutrkn8s1mk4r39qejnbu3n5q?l=Joe.Doe" ) # Test vector from https://wiki.gnupg.org/WKD assert ( get_wkd_direct_url("bernhard.reiter@intevation.de") == "https://intevation.de/.well-known/openpgpkey/hu/it5sewh54rxz33fwmr8u6dy4bbz8itz4?l=bernhard.reiter" ) def get_pgp_key_id(raw_key: bytes) -> str: """Get the identifier of a key, using GnuPG""" # Flush stdout and stderr to prevent interleaving messages from a subprocess sys.stdout.flush() sys.stderr.flush() with tempfile.TemporaryDirectory(prefix="gnupghome") as tmpdir: # Create an empty public keyring to avoid a GnuPG message with (Path(tmpdir) / "pubring.kbx").open("wb"): pass output = subprocess.check_output( ("gpg", "--list-packets"), input=raw_key, env={ "GNUPGHOME": tmpdir, "HOME": tmpdir, }, ) keyid_index = output.index(b"keyid: ") + 7 keyid_end_index = output.index(b"\n", keyid_index) key_id = output[keyid_index:keyid_end_index].decode("ascii") assert len(key_id) == 16 assert all(c in "0123456789ABCDEF" for c in key_id) return key_id def gpg_recv_key(key_id: str) -> bytes: """Receive a key using GnuPG using Ubuntu keyserver https://keyserver.ubuntu.com/""" # Flush stdout and stderr to prevent interleaving messages from a subprocess sys.stdout.flush() sys.stderr.flush() with tempfile.TemporaryDirectory(prefix="gnupghome") as tmpdir: # Create an empty public keyring to avoid a GnuPG message with (Path(tmpdir) / "pubring.kbx").open("wb"): pass with (Path(tmpdir) / "trustdb.gpg").open("wb"): pass # Retry several times, as sometimes the command fails with: # gpg: keyserver receive failed: No data try_count = 0 while 1: try: subprocess.check_output( ("gpg", "--keyserver", "hkps://keyserver.ubuntu.com", "--recv-keys", key_id), input=b"", env={ "GNUPGHOME": tmpdir, "HOME": tmpdir, }, ) break except subprocess.CalledProcessError: if try_count >= 10: raise print(f"Receiving key {key_id} failed [{try_count}], retrying...") time.sleep(1) try_count += 1 raw_key = subprocess.check_output( ("gpg", "--export", key_id), env={ "GNUPGHOME": tmpdir, "HOME": tmpdir, }, ) return raw_key def sync_keys(keys_path: Path) -> None: """Sync all the keys and refresh the given file""" file_lines = [] with keys_path.open("r") as fkeys: for line in fkeys: line = line.strip() if not line or line.startswith("#"): # Keep comments and empty lines file_lines.append(line) continue fields = line.split(" ") if len(fields) < 2: raise ValueError(f"Unexpected line: {line!r}") current_key_id = fields[0] email = fields[1] raw_key = None wkd_url = None key_comment = None if "@" in email: email = email.lower() # Download the key using WKD wkd_url = get_wkd_advanced_url(email) try: with urllib.request.urlopen(wkd_url) as response: raw_key = response.read() except urllib.error.URLError: pass else: print(f"Downloaded key for {email} from {wkd_url}") key_comment = wkd_url # Try the direct method when the advanced one failed # Ignore domains which have issues in their configuration if raw_key is None and not email.endswith("@att.net"): wkd_url = get_wkd_direct_url(email) raw_key = None try: with urllib.request.urlopen(wkd_url) as response: raw_key = response.read() except urllib.error.URLError: pass else: print(f"Downloaded key for {email} from {wkd_url}") key_comment = wkd_url for url in fields[2:]: # Check URL, and only keep the first valid key with urllib.request.urlopen(url) as response: armored_key = response.read() try: new_raw_key = unarmor_gpg(armored_key) except ValueError as exc: raise ValueError(f"Error in {url!r}: {exc}") if new_raw_key == b"": print(f"Downloaded empty key from {url}") continue if raw_key is None: raw_key = new_raw_key key_comment = url print(f"Downloaded key from {url}") # Try using GnuPG directly if raw_key is None: raw_key = gpg_recv_key(current_key_id) key_comment = "received using GnuPG" # Save the key using the key ID key_id = get_pgp_key_id(raw_key) file_name = email.replace("@", "_").replace("+", "_") + "_" + key_id + ".asc" assert re.match( r"^[A-Za-z][-0-9A-Za-z._]+$", file_name ), f"Unexpected characters in file name {file_name!r}" print(f"Saving key for {email!r} in {'all_keys/' + file_name!r}") b64_key = base64.b64encode(raw_key).decode("ascii") with (ALL_KEYS_PATH / file_name).open("w") as fkey: print("-----BEGIN PGP PUBLIC KEY BLOCK-----", file=fkey) print(f"Comment: {key_comment}", file=fkey) print("", file=fkey) for offset in range(0, len(b64_key), 64): print(b64_key[offset:offset + 64], file=fkey) print(opgp_crc24_b64(raw_key), file=fkey)
print("-----END PGP PUBLIC KEY BLOCK-----", file=fkey) # Write the key ID in the file new_line = f"0x{key_id} {email}" if len(fields) > 2:
random_line_split
10_msaa.rs
&model_vertices) .unwrap(); // Create MVP uniform. let uniform_handle = renderer .load_uniform_buffer(DefaultForwardShaderLayout::default()) .unwrap(); // Load textures and create image. let model_texture_file = if use_glb { image::open(MODEL_TEXTURE_FILE_NAME_GLB).unwrap() } else { image::open(MODEL_TEXTURE_FILE_NAME_OBJ).unwrap() }; let mip_levels = get_mip_levels(model_texture_file.dimensions()); let model_texture = renderer .load_image_with_staging_initialization( model_texture_file, MagnificationMinificationFilter::Linear, MagnificationMinificationFilter::Linear, TextureAddressMode::ClampToEdge, TextureAddressMode::ClampToEdge, TextureAddressMode::ClampToEdge, mip_levels, ) .unwrap(); let mut drawable_object_builder = DrawableObject::builder(&renderer) .uniform_buffer(&uniform_handle) .vertex_buffer(&model_buffer) .texture_image(&model_texture); if model_index_buffer.is_some() { drawable_object_builder = drawable_object_builder.index_buffer(model_index_buffer.as_ref().unwrap()); } let drawable_object = drawable_object_builder.build().unwrap(); let start_time = Instant::now(); let mut last_frame_time = start_time; let mut frame_number = 0; let mut fps_average = 0f32; let mut camera_height = -0.5f32; // Run the loop. event_loop.run_return(move |event, _, control_flow| { // By default continuously run this event loop, even if the OS hasn't // distributed an event, that way we will draw as fast as possible. *control_flow = ControlFlow::Poll; match event { Event::MainEventsCleared => { // All the main events to process are done we can do "work" now (game // engine state update etc.) let now = Instant::now(); let time_since_start_secs = ((now - start_time).as_millis() as f32) / 1000f32; if show_fps { let time_since_last_frame_secs = ((now - last_frame_time).as_nanos() as f32) / 1e9f32; let fps = 1f32 / time_since_last_frame_secs; if frame_number == 0 { fps_average = 0f32; } else { fps_average = ((frame_number as f32 * fps_average) + fps) / (frame_number as f32 + 1f32); } frame_number += 1; info!("Frame Period: {}", time_since_last_frame_secs); info!("FPS: {}", fps); info!("FPS averaged: {}", fps_average); last_frame_time = now; } // Rise to max height then gently go back down. let camera_rate = 0.25f32; let min_camera_height = -0.5f32; let camera_range = 2f32; camera_height = (camera_rate * time_since_start_secs) % (2.0f32 * camera_range) + min_camera_height; if camera_height >= (camera_range + min_camera_height) { camera_height = (2.0f32 * (camera_range + min_camera_height)) - camera_height; } let rotation = (std::f32::consts::PI + std::f32::consts::PI * time_since_start_secs / 8f32) % (2f32 * std::f32::consts::PI); update_uniforms( &renderer, &drawable_object, uv::Vec3::new(0f32, -1f32, -1.5f32), rotation, camera_height, false, ar, ) .unwrap(); renderer.draw(&drawable_object).unwrap(); // At the end of work request redraw. window.request_redraw(); } Event::RedrawRequested(_) => { // Redraw requested, this is called after MainEventsCleared. renderer.frame().unwrap_or_else(|err| { match err { SarektError::SwapchainOutOfDate | SarektError::SuboptimalSwapchain => { // Handle window resize etc. warn!("Tried to render without processing window resize event!"); let PhysicalSize { width, height } = window.inner_size(); renderer .recreate_swapchain(width, height) .expect("Error recreating swapchain"); } e => panic!("Frame had an unrecoverable error! {}", e), } }); } Event::WindowEvent { window_id, event } => { main_loop_window_event(&event, &window_id, control_flow, &mut renderer, &mut ar) .expect("Error processing window event."); } Event::LoopDestroyed => { // Explicitly call exit so resources are cleaned up. std::process::exit(0); } _ => (), } }); } /// Handles all winit window specific events. fn main_loop_window_event( event: &WindowEvent, _id: &WindowId, control_flow: &mut winit::event_loop::ControlFlow, renderer: &mut VulkanRenderer, ar: &mut f32, ) -> SarektResult<()> { match event { WindowEvent::CloseRequested => { // When the window system requests a close, signal to winit that we'd like to // close the window. info!("Exiting due to close request event from window system..."); *control_flow = ControlFlow::Exit; } WindowEvent::KeyboardInput { input, .. } => { // When the keyboard input is a press on the escape key, exit and print the // line. if let (Some(VirtualKeyCode::Escape), ElementState::Pressed) = (input.virtual_keycode, input.state) { info!("Exiting due to escape press..."); *control_flow = ControlFlow::Exit } } WindowEvent::Resized(size) => { // If the size is 0, minimization or something like that happened so I // toggle drawing. info!("Window resized, recreating renderer swapchain..."); let enabled = !(size.height == 0 && size.width == 0); if enabled { *ar = size.width as f32 / size.height as f32; } renderer.set_rendering_enabled(enabled); return renderer.recreate_swapchain(size.width, size.height); } _ => (), } Ok(()) } fn update_uniforms( renderer: &VulkanRenderer, object: &DrawableObject<VulkanRenderer, DefaultForwardShaderLayout>, position: uv::Vec3, rotation: f32, camera_height: f32, enable_colors: bool, ar: f32, ) -> SarektResult<()> { // Pi radians per second around the y axis. let total_rotation = uv::Mat4::from_rotation_y(rotation) * uv::Mat4::from_rotation_x(-std::f32::consts::PI / 2f32); let model_matrix = uv::Mat4::from_translation(position) * total_rotation; let view_matrix = uv::Mat4::look_at( /* eye= */ uv::Vec3::new(0.0f32, camera_height, 0.0f32), /* at= */ position, /* up= */ uv::Vec3::unit_y(), ); // TODO BACKENDS this proj should be conditional on backend. let perspective_matrix = uv::projection::rh_yup::perspective_vk(std::f32::consts::PI / 2f32, ar, 0.1f32, 10f32); let uniform = DefaultForwardShaderLayout::new( perspective_matrix * view_matrix * model_matrix, enable_colors, /* enable_texture_mixing= */ true, ); object.set_uniform(renderer, &uniform) } /// For now only use the first object in the obj file. /// Returns (vertices, vertex_indicies, texture_coordinate indices) fn load_obj_models(obj_file_path: &str) -> (Vec<DefaultForwardShaderVertex>, Option<Vec<u32>>) { let mut model_file = File::open(obj_file_path).unwrap(); let mut model_file_text = String::new(); model_file.read_to_string(&mut model_file_text).unwrap(); let obj_set = obj::obj::parse(&model_file_text).unwrap(); if obj_set.objects.len() != 1 { panic!( "The model you attempted to load has more than one object in it, implying it is a scene, if \ you wish to use it as a single model, modify the application code to ignore that or join \ your meshes into a single model" ); } info!("Loaded model {}", OBJ_MODEL_FILE_NAME); let mut vertices: Vec<DefaultForwardShaderVertex> = Vec::new(); let mut indices: Vec<u32> = Vec::new(); // Map of inserted (obj_vertex_index, obj_texture_index) to index in the // vertices array im building. let mut inserted_indices: HashMap<(usize, usize), usize> = HashMap::new(); let model_vertices = &obj_set.objects[0].vertices;
random_line_split
10_msaa.rs
_time = Instant::now(); let mut last_frame_time = start_time; let mut frame_number = 0; let mut fps_average = 0f32; let mut camera_height = -0.5f32; // Run the loop. event_loop.run_return(move |event, _, control_flow| { // By default continuously run this event loop, even if the OS hasn't // distributed an event, that way we will draw as fast as possible. *control_flow = ControlFlow::Poll; match event { Event::MainEventsCleared => { // All the main events to process are done we can do "work" now (game // engine state update etc.) let now = Instant::now(); let time_since_start_secs = ((now - start_time).as_millis() as f32) / 1000f32; if show_fps { let time_since_last_frame_secs = ((now - last_frame_time).as_nanos() as f32) / 1e9f32; let fps = 1f32 / time_since_last_frame_secs; if frame_number == 0 { fps_average = 0f32; } else { fps_average = ((frame_number as f32 * fps_average) + fps) / (frame_number as f32 + 1f32); } frame_number += 1; info!("Frame Period: {}", time_since_last_frame_secs); info!("FPS: {}", fps); info!("FPS averaged: {}", fps_average); last_frame_time = now; } // Rise to max height then gently go back down. let camera_rate = 0.25f32; let min_camera_height = -0.5f32; let camera_range = 2f32; camera_height = (camera_rate * time_since_start_secs) % (2.0f32 * camera_range) + min_camera_height; if camera_height >= (camera_range + min_camera_height) { camera_height = (2.0f32 * (camera_range + min_camera_height)) - camera_height; } let rotation = (std::f32::consts::PI + std::f32::consts::PI * time_since_start_secs / 8f32) % (2f32 * std::f32::consts::PI); update_uniforms( &renderer, &drawable_object, uv::Vec3::new(0f32, -1f32, -1.5f32), rotation, camera_height, false, ar, ) .unwrap(); renderer.draw(&drawable_object).unwrap(); // At the end of work request redraw. window.request_redraw(); } Event::RedrawRequested(_) => { // Redraw requested, this is called after MainEventsCleared. renderer.frame().unwrap_or_else(|err| { match err { SarektError::SwapchainOutOfDate | SarektError::SuboptimalSwapchain => { // Handle window resize etc. warn!("Tried to render without processing window resize event!"); let PhysicalSize { width, height } = window.inner_size(); renderer .recreate_swapchain(width, height) .expect("Error recreating swapchain"); } e => panic!("Frame had an unrecoverable error! {}", e), } }); } Event::WindowEvent { window_id, event } => { main_loop_window_event(&event, &window_id, control_flow, &mut renderer, &mut ar) .expect("Error processing window event."); } Event::LoopDestroyed => { // Explicitly call exit so resources are cleaned up. std::process::exit(0); } _ => (), } }); } /// Handles all winit window specific events. fn main_loop_window_event( event: &WindowEvent, _id: &WindowId, control_flow: &mut winit::event_loop::ControlFlow, renderer: &mut VulkanRenderer, ar: &mut f32, ) -> SarektResult<()> { match event { WindowEvent::CloseRequested => { // When the window system requests a close, signal to winit that we'd like to // close the window. info!("Exiting due to close request event from window system..."); *control_flow = ControlFlow::Exit; } WindowEvent::KeyboardInput { input, .. } => { // When the keyboard input is a press on the escape key, exit and print the // line. if let (Some(VirtualKeyCode::Escape), ElementState::Pressed) = (input.virtual_keycode, input.state) { info!("Exiting due to escape press..."); *control_flow = ControlFlow::Exit } } WindowEvent::Resized(size) => { // If the size is 0, minimization or something like that happened so I // toggle drawing. info!("Window resized, recreating renderer swapchain..."); let enabled = !(size.height == 0 && size.width == 0); if enabled { *ar = size.width as f32 / size.height as f32; } renderer.set_rendering_enabled(enabled); return renderer.recreate_swapchain(size.width, size.height); } _ => (), } Ok(()) } fn update_uniforms( renderer: &VulkanRenderer, object: &DrawableObject<VulkanRenderer, DefaultForwardShaderLayout>, position: uv::Vec3, rotation: f32, camera_height: f32, enable_colors: bool, ar: f32, ) -> SarektResult<()> { // Pi radians per second around the y axis. let total_rotation = uv::Mat4::from_rotation_y(rotation) * uv::Mat4::from_rotation_x(-std::f32::consts::PI / 2f32); let model_matrix = uv::Mat4::from_translation(position) * total_rotation; let view_matrix = uv::Mat4::look_at( /* eye= */ uv::Vec3::new(0.0f32, camera_height, 0.0f32), /* at= */ position, /* up= */ uv::Vec3::unit_y(), ); // TODO BACKENDS this proj should be conditional on backend. let perspective_matrix = uv::projection::rh_yup::perspective_vk(std::f32::consts::PI / 2f32, ar, 0.1f32, 10f32); let uniform = DefaultForwardShaderLayout::new( perspective_matrix * view_matrix * model_matrix, enable_colors, /* enable_texture_mixing= */ true, ); object.set_uniform(renderer, &uniform) } /// For now only use the first object in the obj file. /// Returns (vertices, vertex_indicies, texture_coordinate indices) fn load_obj_models(obj_file_path: &str) -> (Vec<DefaultForwardShaderVertex>, Option<Vec<u32>>) { let mut model_file = File::open(obj_file_path).unwrap(); let mut model_file_text = String::new(); model_file.read_to_string(&mut model_file_text).unwrap(); let obj_set = obj::obj::parse(&model_file_text).unwrap(); if obj_set.objects.len() != 1 { panic!( "The model you attempted to load has more than one object in it, implying it is a scene, if \ you wish to use it as a single model, modify the application code to ignore that or join \ your meshes into a single model" ); } info!("Loaded model {}", OBJ_MODEL_FILE_NAME); let mut vertices: Vec<DefaultForwardShaderVertex> = Vec::new(); let mut indices: Vec<u32> = Vec::new(); // Map of inserted (obj_vertex_index, obj_texture_index) to index in the // vertices array im building. let mut inserted_indices: HashMap<(usize, usize), usize> = HashMap::new(); let model_vertices = &obj_set.objects[0].vertices; for geo in obj_set.objects[0].geometry.iter() { // For every set of geometry (regardless of material for now). for shape in geo.shapes.iter() { // For every face/shape in the set of geometry. match shape.primitive { obj::obj::Primitive::Triangle(x, y, z) =>
{ for &vert in [x, y, z].iter() { // We're only building a buffer of indices and vertices which contain position // and tex coord. let index_key = (vert.0, vert.1.unwrap()); if let Some(&vtx_index) = inserted_indices.get(&index_key) { // Already loaded this (vertex index, texture index) combo, just add it to the // index buffer. indices.push(vtx_index as _); continue; } // This is a new unique vertex (where a vertex is both a position and it's // texture coordinate) so add it to the vertex buffer and the index buffer. let current_vertex = model_vertices[vert.0]; let vertex_as_float = [ current_vertex.x as f32, current_vertex.y as f32, current_vertex.z as f32, ];
conditional_block
10_msaa.rs
() { simple_logger::init_with_level(Level::Info).unwrap(); main_loop(); } /// Takes full control of the executing thread and runs the event loop for it. fn main_loop() { let args: Vec<String> = std::env::args().collect(); let show_fps = args.contains(&"fps".to_owned()); let use_glb = args.contains(&"glb".to_owned()); let msaa_level = if args.contains(&"4x".to_owned()) { 4u8 } else if args.contains(&"8x".to_owned()) { 8u8 } else if args.contains(&"noaa".to_owned()) { 1u8 } else { 2u8 }; info!("MSAA {}x", msaa_level); info!("Show FPS: {}", show_fps); info!("Use GLTF Model Type: {}", use_glb); info!("Running main loop..."); let mut ar = WIDTH as f32 / HEIGHT as f32; // Build Window. let mut event_loop = EventLoop::new(); let window = Arc::new( WindowBuilder::new() .with_inner_size(LogicalSize::new(WIDTH, HEIGHT)) .build(&event_loop) .unwrap(), ); // Build Renderer. let config = Config::builder() .requested_width(WIDTH) .requested_height(HEIGHT) .msaa_config(MsaaConfig::new( msaa_level.try_into().unwrap(), Some(0.2f32), )) .build() .unwrap(); let mut renderer = VulkanRenderer::new(window.clone(), config).unwrap(); // Create Vertex Resources. let (model_vertices, model_indices) = if use_glb { load_glb_model(GLB_MODEL_FILE_NAME) } else { load_obj_models(OBJ_MODEL_FILE_NAME) }; info!("Model file loaded"); let model_index_buffer = model_indices.map(|mi| { renderer .load_buffer(BufferType::Index(IndexBufferElemSize::UInt32), &mi) .unwrap() }); let model_buffer = renderer .load_buffer(BufferType::Vertex, &model_vertices) .unwrap(); // Create MVP uniform. let uniform_handle = renderer .load_uniform_buffer(DefaultForwardShaderLayout::default()) .unwrap(); // Load textures and create image. let model_texture_file = if use_glb { image::open(MODEL_TEXTURE_FILE_NAME_GLB).unwrap() } else { image::open(MODEL_TEXTURE_FILE_NAME_OBJ).unwrap() }; let mip_levels = get_mip_levels(model_texture_file.dimensions()); let model_texture = renderer .load_image_with_staging_initialization( model_texture_file, MagnificationMinificationFilter::Linear, MagnificationMinificationFilter::Linear, TextureAddressMode::ClampToEdge, TextureAddressMode::ClampToEdge, TextureAddressMode::ClampToEdge, mip_levels, ) .unwrap(); let mut drawable_object_builder = DrawableObject::builder(&renderer) .uniform_buffer(&uniform_handle) .vertex_buffer(&model_buffer) .texture_image(&model_texture); if model_index_buffer.is_some() { drawable_object_builder = drawable_object_builder.index_buffer(model_index_buffer.as_ref().unwrap()); } let drawable_object = drawable_object_builder.build().unwrap(); let start_time = Instant::now(); let mut last_frame_time = start_time; let mut frame_number = 0; let mut fps_average = 0f32; let mut camera_height = -0.5f32; // Run the loop. event_loop.run_return(move |event, _, control_flow| { // By default continuously run this event loop, even if the OS hasn't // distributed an event, that way we will draw as fast as possible. *control_flow = ControlFlow::Poll; match event { Event::MainEventsCleared => { // All the main events to process are done we can do "work" now (game // engine state update etc.) let now = Instant::now(); let time_since_start_secs = ((now - start_time).as_millis() as f32) / 1000f32; if show_fps { let time_since_last_frame_secs = ((now - last_frame_time).as_nanos() as f32) / 1e9f32; let fps = 1f32 / time_since_last_frame_secs; if frame_number == 0 { fps_average = 0f32; } else { fps_average = ((frame_number as f32 * fps_average) + fps) / (frame_number as f32 + 1f32); } frame_number += 1; info!("Frame Period: {}", time_since_last_frame_secs); info!("FPS: {}", fps); info!("FPS averaged: {}", fps_average); last_frame_time = now; } // Rise to max height then gently go back down. let camera_rate = 0.25f32; let min_camera_height = -0.5f32; let camera_range = 2f32; camera_height = (camera_rate * time_since_start_secs) % (2.0f32 * camera_range) + min_camera_height; if camera_height >= (camera_range + min_camera_height) { camera_height = (2.0f32 * (camera_range + min_camera_height)) - camera_height; } let rotation = (std::f32::consts::PI + std::f32::consts::PI * time_since_start_secs / 8f32) % (2f32 * std::f32::consts::PI); update_uniforms( &renderer, &drawable_object, uv::Vec3::new(0f32, -1f32, -1.5f32), rotation, camera_height, false, ar, ) .unwrap(); renderer.draw(&drawable_object).unwrap(); // At the end of work request redraw. window.request_redraw(); } Event::RedrawRequested(_) => { // Redraw requested, this is called after MainEventsCleared. renderer.frame().unwrap_or_else(|err| { match err { SarektError::SwapchainOutOfDate | SarektError::SuboptimalSwapchain => { // Handle window resize etc. warn!("Tried to render without processing window resize event!"); let PhysicalSize { width, height } = window.inner_size(); renderer .recreate_swapchain(width, height) .expect("Error recreating swapchain"); } e => panic!("Frame had an unrecoverable error! {}", e), } }); } Event::WindowEvent { window_id, event } => { main_loop_window_event(&event, &window_id, control_flow, &mut renderer, &mut ar) .expect("Error processing window event."); } Event::LoopDestroyed => { // Explicitly call exit so resources are cleaned up. std::process::exit(0); } _ => (), } }); } /// Handles all winit window specific events. fn main_loop_window_event( event: &WindowEvent, _id: &WindowId, control_flow: &mut winit::event_loop::ControlFlow, renderer: &mut VulkanRenderer, ar: &mut f32, ) -> SarektResult<()> { match event { WindowEvent::CloseRequested => { // When the window system requests a close, signal to winit that we'd like to // close the window. info!("Exiting due to close request event from window system..."); *control_flow = ControlFlow::Exit; } WindowEvent::KeyboardInput { input, .. } => { // When the keyboard input is a press on the escape key, exit and print the // line. if let (Some(VirtualKeyCode::Escape), ElementState::Pressed) = (input.virtual_keycode, input.state) { info!("Exiting due to escape press..."); *control_flow = ControlFlow::Exit } } WindowEvent::Resized(size) => { // If the size is 0, minimization or something like that happened so I // toggle drawing. info!("Window resized, recreating renderer swapchain..."); let enabled = !(size.height == 0 && size.width == 0); if enabled { *ar = size.width as f32 / size.height as f32; } renderer.set_rendering_enabled(enabled); return renderer.recreate_swapchain(size.width, size.height); } _ => (), } Ok(()) } fn update_uniforms( renderer: &VulkanRenderer, object: &DrawableObject<VulkanRenderer, DefaultForwardShaderLayout>, position: uv::Vec3, rotation: f32, camera_height: f32, enable_colors: bool, ar: f32, ) -> SarektResult<()> { // Pi radians per second around the y axis. let total_rotation = uv::Mat4::from_rotation_y(rotation) * uv::Mat4::from_rotation_x(-std::f32::consts::PI / 2
main
identifier_name
10_msaa.rs
let mut ar = WIDTH as f32 / HEIGHT as f32; // Build Window. let mut event_loop = EventLoop::new(); let window = Arc::new( WindowBuilder::new() .with_inner_size(LogicalSize::new(WIDTH, HEIGHT)) .build(&event_loop) .unwrap(), ); // Build Renderer. let config = Config::builder() .requested_width(WIDTH) .requested_height(HEIGHT) .msaa_config(MsaaConfig::new( msaa_level.try_into().unwrap(), Some(0.2f32), )) .build() .unwrap(); let mut renderer = VulkanRenderer::new(window.clone(), config).unwrap(); // Create Vertex Resources. let (model_vertices, model_indices) = if use_glb { load_glb_model(GLB_MODEL_FILE_NAME) } else { load_obj_models(OBJ_MODEL_FILE_NAME) }; info!("Model file loaded"); let model_index_buffer = model_indices.map(|mi| { renderer .load_buffer(BufferType::Index(IndexBufferElemSize::UInt32), &mi) .unwrap() }); let model_buffer = renderer .load_buffer(BufferType::Vertex, &model_vertices) .unwrap(); // Create MVP uniform. let uniform_handle = renderer .load_uniform_buffer(DefaultForwardShaderLayout::default()) .unwrap(); // Load textures and create image. let model_texture_file = if use_glb { image::open(MODEL_TEXTURE_FILE_NAME_GLB).unwrap() } else { image::open(MODEL_TEXTURE_FILE_NAME_OBJ).unwrap() }; let mip_levels = get_mip_levels(model_texture_file.dimensions()); let model_texture = renderer .load_image_with_staging_initialization( model_texture_file, MagnificationMinificationFilter::Linear, MagnificationMinificationFilter::Linear, TextureAddressMode::ClampToEdge, TextureAddressMode::ClampToEdge, TextureAddressMode::ClampToEdge, mip_levels, ) .unwrap(); let mut drawable_object_builder = DrawableObject::builder(&renderer) .uniform_buffer(&uniform_handle) .vertex_buffer(&model_buffer) .texture_image(&model_texture); if model_index_buffer.is_some() { drawable_object_builder = drawable_object_builder.index_buffer(model_index_buffer.as_ref().unwrap()); } let drawable_object = drawable_object_builder.build().unwrap(); let start_time = Instant::now(); let mut last_frame_time = start_time; let mut frame_number = 0; let mut fps_average = 0f32; let mut camera_height = -0.5f32; // Run the loop. event_loop.run_return(move |event, _, control_flow| { // By default continuously run this event loop, even if the OS hasn't // distributed an event, that way we will draw as fast as possible. *control_flow = ControlFlow::Poll; match event { Event::MainEventsCleared => { // All the main events to process are done we can do "work" now (game // engine state update etc.) let now = Instant::now(); let time_since_start_secs = ((now - start_time).as_millis() as f32) / 1000f32; if show_fps { let time_since_last_frame_secs = ((now - last_frame_time).as_nanos() as f32) / 1e9f32; let fps = 1f32 / time_since_last_frame_secs; if frame_number == 0 { fps_average = 0f32; } else { fps_average = ((frame_number as f32 * fps_average) + fps) / (frame_number as f32 + 1f32); } frame_number += 1; info!("Frame Period: {}", time_since_last_frame_secs); info!("FPS: {}", fps); info!("FPS averaged: {}", fps_average); last_frame_time = now; } // Rise to max height then gently go back down. let camera_rate = 0.25f32; let min_camera_height = -0.5f32; let camera_range = 2f32; camera_height = (camera_rate * time_since_start_secs) % (2.0f32 * camera_range) + min_camera_height; if camera_height >= (camera_range + min_camera_height) { camera_height = (2.0f32 * (camera_range + min_camera_height)) - camera_height; } let rotation = (std::f32::consts::PI + std::f32::consts::PI * time_since_start_secs / 8f32) % (2f32 * std::f32::consts::PI); update_uniforms( &renderer, &drawable_object, uv::Vec3::new(0f32, -1f32, -1.5f32), rotation, camera_height, false, ar, ) .unwrap(); renderer.draw(&drawable_object).unwrap(); // At the end of work request redraw. window.request_redraw(); } Event::RedrawRequested(_) => { // Redraw requested, this is called after MainEventsCleared. renderer.frame().unwrap_or_else(|err| { match err { SarektError::SwapchainOutOfDate | SarektError::SuboptimalSwapchain => { // Handle window resize etc. warn!("Tried to render without processing window resize event!"); let PhysicalSize { width, height } = window.inner_size(); renderer .recreate_swapchain(width, height) .expect("Error recreating swapchain"); } e => panic!("Frame had an unrecoverable error! {}", e), } }); } Event::WindowEvent { window_id, event } => { main_loop_window_event(&event, &window_id, control_flow, &mut renderer, &mut ar) .expect("Error processing window event."); } Event::LoopDestroyed => { // Explicitly call exit so resources are cleaned up. std::process::exit(0); } _ => (), } }); } /// Handles all winit window specific events. fn main_loop_window_event( event: &WindowEvent, _id: &WindowId, control_flow: &mut winit::event_loop::ControlFlow, renderer: &mut VulkanRenderer, ar: &mut f32, ) -> SarektResult<()> { match event { WindowEvent::CloseRequested => { // When the window system requests a close, signal to winit that we'd like to // close the window. info!("Exiting due to close request event from window system..."); *control_flow = ControlFlow::Exit; } WindowEvent::KeyboardInput { input, .. } => { // When the keyboard input is a press on the escape key, exit and print the // line. if let (Some(VirtualKeyCode::Escape), ElementState::Pressed) = (input.virtual_keycode, input.state) { info!("Exiting due to escape press..."); *control_flow = ControlFlow::Exit } } WindowEvent::Resized(size) => { // If the size is 0, minimization or something like that happened so I // toggle drawing. info!("Window resized, recreating renderer swapchain..."); let enabled = !(size.height == 0 && size.width == 0); if enabled { *ar = size.width as f32 / size.height as f32; } renderer.set_rendering_enabled(enabled); return renderer.recreate_swapchain(size.width, size.height); } _ => (), } Ok(()) } fn update_uniforms( renderer: &VulkanRenderer, object: &DrawableObject<VulkanRenderer, DefaultForwardShaderLayout>, position: uv::Vec3, rotation: f32, camera_height: f32, enable_colors: bool, ar: f32, ) -> SarektResult<()> { // Pi radians per second around the y axis. let total_rotation = uv::Mat4::from_rotation_y(rotation) * uv::Mat4::from_rotation_x(-std::f32::consts::PI / 2f32); let model_matrix = uv::Mat4::from_translation(position) * total_rotation; let view_matrix = uv::Mat4::look_at( /* eye= */ uv
{ let args: Vec<String> = std::env::args().collect(); let show_fps = args.contains(&"fps".to_owned()); let use_glb = args.contains(&"glb".to_owned()); let msaa_level = if args.contains(&"4x".to_owned()) { 4u8 } else if args.contains(&"8x".to_owned()) { 8u8 } else if args.contains(&"noaa".to_owned()) { 1u8 } else { 2u8 }; info!("MSAA {}x", msaa_level); info!("Show FPS: {}", show_fps); info!("Use GLTF Model Type: {}", use_glb); info!("Running main loop...");
identifier_body
QueueWorkerService.ts
console.log('bullqueue paused') }) .on('resumed', function (job) { // The queue has been resumed. console.log('bullqueue resumed') }) .on('cleaned', function (jobs, type) { // Old jobs have been cleaned from the queue. `jobs` is an array of cleaned // jobs, and `type` is the type of jobs cleaned. console.log('bullqueue cleaned', type) }); const queueIndex = this.queueTopics.length - 1; queue.process((job,done) => { const MaxDoneTime = 3 * 60 * 1000; // 最多执行30分钟 this.doneInComeCall(job, queueIndex, MaxDoneTime) .then(res => { this.logger.debug('doneInComeCall res:', res); done(null,res); }) .catch(err => { this.logger.error('doneInComeCall error:', err); done(err); }) }); return queue; } else { return this.getQueueByName(qNameTopic); } } getQueueByName(name: string): Queue { try { const index = this.queueTopics.indexOf(name); if (index > -1) { return this.queues[index]; } else { return null; } } catch (ex) { this.logger.error('getQueueByName error:', ex); } } async readyCacheBullQueue() { try { const keys: string[] = await this.bullQueueClient.keys('bullQueueCache*'); this.logger.debug('readyCacheBullQueue:', keys.join(',')); for (let i = 0; i < keys.length; i++) { const key = keys[i] || 'bullQueueCache'; const works = key.split('::'); if (works.length === 4) { this.add(works[2], works[3]); } } return Promise.resolve(); } catch (ex) { this.logger.debug('readyCacheBullQueue error ', ex); return Promise.reject(ex); } } setCacheBullKey(name) { this.logger.debug('Cache Bull Key : ', name); this.bullQueueClient.set(`bullQueueCache::${name}`, 1) .then() .catch(err => { this.logger.error('bullQueue set cache key error:', name); }) } /** * @description * shuffle算法,类似摸牌 * arr为原数组,cards为乱序结果数组 * random取一个index,取arr中这个元素,放入cards,同时移除arr中这个元素。 * @param originalArray 原始数组 * @return {Array} */ shuffle(originalArray: string[]) { const mixedArray = []; const copyArray = originalArray.slice(0); // 防止改变原来的参数,数组是引用传递 while (copyArray.length > 0) { //generate a random index of the original array const randomIndex = Math.random() * copyArray.length; //push the random element into the mixed one, at the same time, delete the original element mixedArray.push(copyArray[randomIndex]); copyArray.splice(randomIndex, 1); } return mixedArray; } async randomStrategy({ members, tenantId }) { try { const lockedMembers = await this.getLockedMembers({ tenantId }); const unlockedMember = members.filter(x => { return lockedMembers.indexOf(String(x)) === -1 }); const membersRandom = this.shuffle(unlockedMember); const member = this.cycleFind({ members: membersRandom, tenantId }); return Promise.resolve(member); } catch (ex) { this.logger.error('randomStrategy error:', ex); return Promise.reject(ex); } } /** * @description * 按队列中坐席的顺序从上到下一次查找可用的坐席 * 为什么要按members坐席的顺序来,因为实际应用中,这个顺序通常可以用来代表坐席的等级,一般技能卓越 * 的优秀坐席应该放在前面,以为客户提供最好的服务! * */ async topDownStrategy({ members, tenantId }) { try { const lockedMembers = await this.getLockedMembers({ tenantId }); const unlockedMember = members.filter(x => { return lockedMembers.indexOf(String(x)) === -1 }); const member = this.cycleFind({ members: unlockedMember, tenantId }); return Promise.resolve(member); } catch (ex) { this.logger.error('topDownStrategy error:', ex); return Promise.reject(ex); } } async cycleFind({ members, tenantId }) { try { let finded = null; for (let i = 0; i < members.length; i++) { //redis锁定成员 //检查是否可用 const member = `${members[i]}`; finded = await this.pbxExtensionController.checkAgentCanDail(tenantId, member); // this.logger.debug('find a member:', finded); if (finded) break; } return Promise.resolve(finded); } catch (ex) { this.logger.error('cycleFind error:', ex); return Promise.reject(ex); } } async roundRobinStrategy({ members, tenantId, queueNumber }) { try { const finds = await this.pbxAgentController.getRoundRobinAgents(tenantId, queueNumber); let newArray = []; if (finds && finds.length) { const agent = finds[0]; if (agent && agent.agentNumber) { const index = members.indexOf(Number(agent.agentNumber)); if (index > -1) { const firstArray = members.slice(index + 1); const lastArray = members.slice(0, index + 1); newArray = firstArray.concat(lastArray); } else { this.logger.info('AgentNumber Is Not Found In Queue Members!'); newArray = members.slice(0); } } else { this.logger.info('Agent Is Not Found In pbx_agents!'); newArray = members.slice(0); } } else { newArray = members.slice(0); } const lockedMembers = await this.getLockedMembers({ tenantId }); const unlockedMember = newArray.filter(x => { return lockedMembers.indexOf(String(x)) === -1 }); // this.logger.info('=====roundRobinStrategy newArray=====', members, newArray, unlockedMember); const member = this.cycleFind({ members: unlockedMember, tenantId }); return Promise.resolve(member); } catch (ex) { this.logger.error(ex); return Promise.reject(ex); } } async getLockedMembers({ tenantId }) { try { const regKey = `esl::queue::member::locked::${tenantId}::*`; const keys = await this.redLockClient.keys(regKey); const members = []; keys.forEach(key => { const items = key.split('::'); members.push(items[items.length - 1]) }) return Promise.resolve(members); } catch (ex) { return Promise.reject(ex); } } async lockMember({ tenantId, member }) { try { const key = `esl::queue::member::locked::${tenantId}::${member}`; //await redisQC.setnx(key, member); // await redisQC.expire(key, 60); const ttl = 3 * 1000; const lock = await this.redlock.lock(key, ttl); return Promise.resolve(lock); } catch (ex) { return Promise.reject(ex); } } async doneInComeCall(args: Job, queueIndex, timeout2: number) { try { const argData = args.data; const { queue, tenantId, callId, timeout } = argData; this.logger.info(`doneInComeCall ${tenantId}[${callId}]`); const { members, queueNumber } = queue; const eventName = `stopFindAgent::${tenantId}::${callId}`; this.logger.debug('doneInComeCall', eventName); let eslSendStop = false; let maxTimeOut = false; this.e
ventService.once(eventName, (data) => { this.logger.info(`ESL Send Stop Job:${data.jobId} ,My Job Is: ${args.id}`); if (data.jobId === args.id) { eslSendStop = true; } }) const startTime = new Date().getTime(); let isMyTurn = false; let pubData = null; if (Array.isArray(members) && members.length > 0) { while (!eslSendStop && !isMyTurn) { const activeJobs = await this.queues[queueIndex].getActive(); this.logger.info(`activeJobs:${activeJobs.length}`); isMyTurn = true; for (let k = 0; k < activeJobs.length; k++) { const actJob = activeJobs[k]; const actJobOpts = actJob.opts; const actJobId = actJob.id; const { priority, timestamp } = actJobOpts; console.log(`myJob:[${args.id},${args.opts.priority},${args.opts.timestamp}],compareJob:[${actJobId},${priority},${timestamp}]`);
identifier_body
QueueWorkerService.ts
60 * 1000; // 最多执行30分钟 this.doneInComeCall(job, queueIndex, MaxDoneTime) .then(res => { this.logger.debug('doneInComeCall res:', res); done(null,res); }) .catch(err => { this.logger.error('doneInComeCall error:', err); done(err); }) }); return queue; } else { return this.getQueueByName(qNameTopic); } } getQueueByName(name: string): Queue { try { const index = this.queueTopics.indexOf(name); if (index > -1) { return this.queues[index]; } else { return null; } } catch (ex) { this.logger.error('getQueueByName error:', ex); } } async readyCacheBullQueue() { try { const keys: string[] = await this.bullQueueClient.keys('bullQueueCache*'); this.logger.debug('readyCacheBullQueue:', keys.join(',')); for (let i = 0; i < keys.length; i++) { const key = keys[i] || 'bullQueueCache'; const works = key.split('::'); if (works.length === 4) { this.add(works[2], works[3]); } } return Promise.resolve(); } catch (ex) { this.logger.debug('readyCacheBullQueue error ', ex); return Promise.reject(ex); } } setCacheBullKey(name) { this.logger.debug('Cache Bull Key : ', name); this.bullQueueClient.set(`bullQueueCache::${name}`, 1) .then() .catch(err => { this.logger.error('bullQueue set cache key error:', name); }) } /** * @description * shuffle算法,类似摸牌 * arr为原数组,cards为乱序结果数组 * random取一个index,取arr中这个元素,放入cards,同时移除arr中这个元素。 * @param originalArray 原始数组 * @return {Array} */ shuffle(originalArray: string[]) { const mixedArray = []; const copyArray = originalArray.slice(0); // 防止改变原来的参数,数组是引用传递 while (copyArray.length > 0) { //generate a random index of the original array const randomIndex = Math.random() * copyArray.length; //push the random element into the mixed one, at the same time, delete the original element mixedArray.push(copyArray[randomIndex]); copyArray.splice(randomIndex, 1); } return mixedArray; } async randomStrategy({ members, tenantId }) { try { const lockedMembers = await this.getLockedMembers({ tenantId }); const unlockedMember = members.filter(x => { return lockedMembers.indexOf(String(x)) === -1 }); const membersRandom = this.shuffle(unlockedMember); const member = this.cycleFind({ members: membersRandom, tenantId }); return Promise.resolve(member); } catch (ex) { this.logger.error('randomStrategy error:', ex); return Promise.reject(ex); } } /** * @description * 按队列中坐席的顺序从上到下一次查找可用的坐席 * 为什么要按members坐席的顺序来,因为实际应用中,这个顺序通常可以用来代表坐席的等级,一般技能卓越 * 的优秀坐席应该放在前面,以为客户提供最好的服务! * */ async topDownStrategy({ members, tenantId }) { try { const lockedMembers = await this.getLockedMembers({ tenantId }); const unlockedMember = members.filter(x => { return lockedMembers.indexOf(String(x)) === -1 }); const member = this.cycleFind({ members: unlockedMember, tenantId }); return Promise.resolve(member); } catch (ex) { this.logger.error('topDownStrategy error:', ex); return Promise.reject(ex); } } async cycleFind({ members, tenantId }) { try { let finded = null; for (let i = 0; i < members.length; i++) { //redis锁定成员 //检查是否可用 const member = `${members[i]}`; finded = await this.pbxExtensionController.checkAgentCanDail(tenantId, member); // this.logger.debug('find a member:', finded); if (finded) break; } return Promise.resolve(finded); } catch (ex) { this.logger.error('cycleFind error:', ex); return Promise.reject(ex); } } async roundRobinStrategy({ members, tenantId, queueNumber }) { try { const finds = await this.pbxAgentController.getRoundRobinAgents(tenantId, queueNumber); let newArray = []; if (finds && finds.length) { const agent = finds[0]; if (agent && agent.agentNumber) { const index = members.indexOf(Number(agent.agentNumber)); if (index > -1) { const firstArray = members.slice(index + 1); const lastArray = members.slice(0, index + 1); newArray = firstArray.concat(lastArray); } else { this.logger.info('AgentNumber Is Not Found In Queue Members!'); newArray = members.slice(0); } } else { this.logger.info('Agent Is Not Found In pbx_agents!'); newArray = members.slice(0); } } else { newArray = members.slice(0); } const lockedMembers = await this.getLockedMembers({ tenantId }); const unlockedMember = newArray.filter(x => { return lockedMembers.indexOf(String(x)) === -1 }); // this.logger.info('=====roundRobinStrategy newArray=====', members, newArray, unlockedMember); const member = this.cycleFind({ members: unlockedMember, tenantId }); return Promise.resolve(member); } catch (ex) { this.logger.error(ex); return Promise.reject(ex); } } async getLockedMembers({ tenantId }) { try { const regKey = `esl::queue::member::locked::${tenantId}::*`; const keys = await this.redLockClient.keys(regKey); const members = []; keys.forEach(key => { const items = key.split('::'); members.push(items[items.length - 1]) }) return Promise.resolve(members); } catch (ex) { return Promise.reject(ex); } } async lockMember({ tenantId, member }) { try { const key = `esl::queue::member::locked::${tenantId}::${member}`; //await redisQC.setnx(key, member); // await redisQC.expire(key, 60); const ttl = 3 * 1000; const lock = await this.redlock.lock(key, ttl); return Promise.resolve(lock); } catch (ex) { return Promise.reject(ex); } } async doneInComeCall(args: Job, queueIndex, timeout2: number) { try { const argData = args.data; const { queue, tenantId, callId, timeout } = argData; this.logger.info(`doneInComeCall ${tenantId}[${callId}]`); const { members, queueNumber } = queue; const eventName = `stopFindAgent::${tenantId}::${callId}`; this.logger.debug('doneInComeCall', eventName); let eslSendStop = false; let maxTimeOut = false; this.eventService.once(eventName, (data) => { this.logger.info(`ESL Send Stop Job:${data.jobId} ,My Job Is: ${args.id}`); if (data.jobId === args.id) { eslSendStop = true; } }) const startTime = new Date().getTime(); let isMyTurn = false; let pubData = null; if (Array.isArray(members) && members.length > 0) { while (!eslSendStop && !isMyTurn) { const activeJobs = await this.queues[queueIndex].getActive(); this.logger.info(`activeJobs:${activeJobs.length}`); isMyTurn = true; for (let k = 0; k < activeJobs.length; k++) { const actJob = activeJobs[k]; const actJobOpts = actJob.opts; const actJobId = actJob.id; const { priority, time
stamp } = actJobOpts; console.log(`myJob:[${args.id},${args.opts.priority},${args.opts.timestamp}],compareJob:[${actJobId},${priority},${timestamp}]`); if (priority < args.opts.priority) { this.logger.info('=============存在优先级比我高的============='); isMyTurn = false; await this.wait(1 * 1000); break; } else if (priority === args.opts.priority && timestamp < args.opts.timestamp) { this.logger.info('=============和我优先级别一样,但是比我进的早!==========='); isMyTurn = false; await this.wait(1 * 1000); break; } else { await this.wait(1 * 1000); } } }
conditional_block
QueueWorkerService.ts
如果已经存在,返回 */ add(tenantId: string, queueNumber: string): Queue { const qNameTopic = `esl_q_queue::${tenantId}::${queueNumber}`; if (this.queueTopics.indexOf(qNameTopic) < 0) { const queue = new BullQueue(qNameTopic, this.queueOptions); this.queueTopics.push(qNameTopic) this.queues.push(queue); this.setCacheBullKey(qNameTopic); queue .on('error', function (error) { // An error occured. console.log('bullqueue', error) }) .on('active', function (job, jobPromise) { // A job has started. You can use `jobPromise.cancel()`` to abort it. console.log('bullqueue active', job.id, new Date()) }) .on('stalled', function (job) { // A job has been marked as stalled. This is useful for debugging job // workers that crash or pause the event loop. console.log('bullqueue stalled', job.id, new Date()) }) .on('progress', function (job, progress) { // A job's progress was updated! console.log('bullqueue progress', job.id, new Date()) }) .on('global:progress', function (jobId, progress) { console.log(`Job ${jobId} is ${progress * 100}% ready!`); }) .on('completed', function (job, result) { // A job successfully completed with a `result`. console.log('in queueworker bullqueue completed', job.id, result) }) .on('failed', function (job, err) { // A job failed with reason `err`! console.log('bullqueue failed', job.id) // seneca.act({ role: 'pubsub', path: 'queue_job_fail', data: JSON.stringify({ id:job.id, data:job.data }) }, (err, rsp) => { // console.log('bullqueue failed pubsub',err,rsp) // }) }) .on('paused', function () { // The queue has been paused. console.log('bullqueue paused') }) .on('resumed', function (job) { // The queue has been resumed. console.log('bullqueue resumed') }) .on('cleaned', function (jobs, type) { // Old jobs have been cleaned from the queue. `jobs` is an array of cleaned // jobs, and `type` is the type of jobs cleaned. console.log('bullqueue cleaned', type) }); const queueIndex = this.queueTopics.length - 1; queue.process((job,done) => { const MaxDoneTime = 3 * 60 * 1000; // 最多执行30分钟 this.doneInComeCall(job, queueIndex, MaxDoneTime) .then(res => { this.logger.debug('doneInComeCall res:', res); done(null,res); }) .catch(err => { this.logger.error('doneInComeCall error:', err); done(err); }) }); return queue; } else { return this.getQueueByName(qNameTopic); } } getQueueByName(name: string): Queue { try { const index = this.queueTopics.indexOf(name); if (index > -1) { return this.queues[index]; } else { return null; } } catch (ex) { this.logger.error('getQueueByName error:', ex); } } async readyCacheBullQueue() { try { const keys: string[] = await this.bullQueueClient.keys('bullQueueCache*'); this.logger.debug('readyCacheBullQueue:', keys.join(',')); for (let i = 0; i < keys.length; i++) { const key = keys[i] || 'bullQueueCache'; const works = key.split('::'); if (works.length === 4) { this.add(works[2], works[3]); } } return Promise.resolve(); } catch (ex) { this.logger.debug('readyCacheBullQueue error ', ex); return Promise.reject(ex); } } setCacheBullKey(name) { this.logger.debug('Cache Bull Key : ', name); this.bullQueueClient.set(`bullQueueCache::${name}`, 1) .then() .catch(err => { this.logger.error('bullQueue set cache key error:', name); }) } /** * @description * shuffle算法,类似摸牌 * arr为原数组,cards为乱序结果数组 * random取一个index,取arr中这个元素,放入cards,同时移除arr中这个元素。 * @param originalArray 原始数组 * @return {Array} */ shuffle(originalArray: string[]) { const mixedArray = []; const copyArray = originalArray.slice(0); // 防止改变原来的参数,数组是引用传递 while (copyArray.length > 0) { //generate a random index of the original array const randomIndex = Math.random() * copyArray.length; //push the random element into the mixed one, at the same time, delete the original element mixedArray.push(copyArray[randomIndex]); copyArray.splice(randomIndex, 1); } return mixedArray; } async randomStrategy({ members, tenantId }) { try { const lockedMembers = await this.getLockedMembers({ tenantId }); const unlockedMember = members.filter(x => { return lockedMembers.indexOf(String(x)) === -1 }); const membersRandom = this.shuffle(unlockedMember); const member = this.cycleFind({ members: membersRandom, tenantId }); return Promise.resolve(member); } catch (ex) { this.logger.error('randomStrategy error:', ex); return Promise.reject(ex); } } /** * @description * 按队列中坐席的顺序从上到下一次查找可用的坐席 * 为什么要按members坐席的顺序来,因为实际应用中,这个顺序通常可以用来代表坐席的等级,一般技能卓越 * 的优秀坐席应该放在前面,以为客户提供最好的服务! * */ async topDownStrategy({ members, tenantId }) { try { const lockedMembers = await this.getLockedMembers({ tenantId }); const unlockedMember = members.filter(x => { return lockedMembers.indexOf(String(x)) === -1 }); const member = this.cycleFind({ members: unlockedMember, tenantId }); return Promise.resolve(member); } catch (ex) { this.logger.error('topDownStrategy error:', ex); return Promise.reject(ex); } } async cycleFind({ members, tenantId }) { try { let finded = null; for (let i = 0; i < members.length; i++) { //redis锁定成员 //检查是否可用 const member = `${members[i]}`; finded = await this.pbxExtensionController.checkAgentCanDail(tenantId, member); // this.logger.debug('find a member:', finded); if (finded) break; } return Promise.resolve(finded); } catch (ex) { this.logger.error('cycleFind error:', ex); return Promise.reject(ex); } } async roundRobinStrategy({ members, tenantId, queueNumber }) { try { const finds = await this.pbxAgentController.getRoundRobinAgents(tenantId, queueNumber); let newArray = []; if (finds && finds.length) { const agent = finds[0]; if (agent && agent.agentNumber) { const index = members.indexOf(Number(agent.agentNumber)); if (index > -1) { const firstArray = members.slice(index + 1); const lastArray = members.slice(0, index + 1); newArray = firstArray.concat(lastArray); } else { this.logger.info('AgentNumber Is Not Found In Queue Members!'); newArray = members.slice(0); } } else { this.logger.info('Agent Is Not Found In pbx_agents!'); newArray = members.slice(0); } } else { newArray = members.slice(0); } const lockedMembers = await this.getLockedMembers({ tenantId }); const unlockedMember = newArray.filter(x => { return lockedMembers.indexOf(String(x)) === -1 }); // this.logger.info('=====roundRobinStrategy newArray=====', members, newArray, unlockedMember); const member = this.cycleFind({ members: unlockedMember, tenantId }); return Promise.resolve(member); } catch (ex) { this.logger.error(ex); return Promise.reject(ex); } } async getLockedMembers({ tenantId }) { try { const regKey = `esl::queue::member::locked::${tenantId}::*`; const keys = await this.redLockClient.keys(regKey); const members = []; keys.forEach(key => { const items = key.split('::'); members.push(items[items.length - 1]) }) return Promise.resolve(members); } catch (ex) { return Promise
.reject(ex);
identifier_name
QueueWorkerService.ts
import { PBXExtensionController } from '../controllers/pbx_extension'; import { PBXAgentController } from '../controllers/pbx_agent'; import { UserEventController } from '../controllers/userEvent'; import { RedisOptions, Redis } from 'ioredis'; import Redlock = require('redlock'); import { Lock } from 'redlock'; @Injectable() /** * @description 用bull实现队列排队,以每一个组合的每一个队列为一个排队队列,包括插队等功能 */ export class QueueWorkerService { private queueTopics: string[]; private queues: Queue[]; // 此处算法可以改进 private childInjector: Injector; private queueOptions: QueueOptions; private redlock: Redlock; private redLockClient: Redis; private bullQueueClient: Redis; private redisService: RedisService; private eventService: EventService; private pbxAgentController: PBXAgentController; private pbxExtensionController: PBXExtensionController; constructor(private injector: Injector, private logger: LoggerService, private config: ConfigService) { this.createChildInjector(); this.eventService = this.injector.get(EventService); this.pbxAgentController = this.childInjector.get(PBXAgentController); this.pbxExtensionController = this.childInjector.get(PBXExtensionController); } async init() { try { this.redisService = this.injector.get(RedisService); this.queueOptions = { redis: { host: this.config.getConfig().redis.host, port: this.config.getConfig().redis.port, password: this.config.getConfig().redis.password ? this.config.getConfig().redis.password : null, db: 10, }, prefix: 'esl_bull' } this.redLockClient = this.redisService.getClientByName('RedLock'); this.bullQueueClient = this.redisService.getClientByName('BullQueue'); this.redlock = new Redlock( // you should have one client for each redis node // in your cluster [this.redLockClient], { // the expected clock drift; for more details driftFactor: 0.01, // time in ms // the max number of times Redlock will attempt to lock a resource before erroring retryCount: 10, // the time in ms between attempts retryDelay: 200, // time in ms // the max time in ms randomly added to retries // to improve performance under high contention // see https://www.awsarchitectureblog.com/2015/03/backoff.html retryJitter: 200 // time in ms }); this.redlock.on('clientError', function (err) { this.logger.error('A redis error has occurred:', err); }) this.queueTopics = []; this.queues = [] } catch (ex) { } } /** * * @param topic 以租户及队列名称组合的唯一的队列名,如果已经存在,返回 */ add(tenantId: string, queueNumber: string): Queue { const qNameTopic = `esl_q_queue::${tenantId}::${queueNumber}`; if (this.queueTopics.indexOf(qNameTopic) < 0) { const queue = new BullQueue(qNameTopic, this.queueOptions); this.queueTopics.push(qNameTopic) this.queues.push(queue); this.setCacheBullKey(qNameTopic); queue .on('error', function (error) { // An error occured. console.log('bullqueue', error) }) .on('active', function (job, jobPromise) { // A job has started. You can use `jobPromise.cancel()`` to abort it. console.log('bullqueue active', job.id, new Date()) }) .on('stalled', function (job) { // A job has been marked as stalled. This is useful for debugging job // workers that crash or pause the event loop. console.log('bullqueue stalled', job.id, new Date()) }) .on('progress', function (job, progress) { // A job's progress was updated! console.log('bullqueue progress', job.id, new Date()) }) .on('global:progress', function (jobId, progress) { console.log(`Job ${jobId} is ${progress * 100}% ready!`); }) .on('completed', function (job, result) { // A job successfully completed with a `result`. console.log('in queueworker bullqueue completed', job.id, result) }) .on('failed', function (job, err) { // A job failed with reason `err`! console.log('bullqueue failed', job.id) // seneca.act({ role: 'pubsub', path: 'queue_job_fail', data: JSON.stringify({ id:job.id, data:job.data }) }, (err, rsp) => { // console.log('bullqueue failed pubsub',err,rsp) // }) }) .on('paused', function () { // The queue has been paused. console.log('bullqueue paused') }) .on('resumed', function (job) { // The queue has been resumed. console.log('bullqueue resumed') }) .on('cleaned', function (jobs, type) { // Old jobs have been cleaned from the queue. `jobs` is an array of cleaned // jobs, and `type` is the type of jobs cleaned. console.log('bullqueue cleaned', type) }); const queueIndex = this.queueTopics.length - 1; queue.process((job,done) => { const MaxDoneTime = 3 * 60 * 1000; // 最多执行30分钟 this.doneInComeCall(job, queueIndex, MaxDoneTime) .then(res => { this.logger.debug('doneInComeCall res:', res); done(null,res); }) .catch(err => { this.logger.error('doneInComeCall error:', err); done(err); }) }); return queue; } else { return this.getQueueByName(qNameTopic); } } getQueueByName(name: string): Queue { try { const index = this.queueTopics.indexOf(name); if (index > -1) { return this.queues[index]; } else { return null; } } catch (ex) { this.logger.error('getQueueByName error:', ex); } } async readyCacheBullQueue() { try { const keys: string[] = await this.bullQueueClient.keys('bullQueueCache*'); this.logger.debug('readyCacheBullQueue:', keys.join(',')); for (let i = 0; i < keys.length; i++) { const key = keys[i] || 'bullQueueCache'; const works = key.split('::'); if (works.length === 4) { this.add(works[2], works[3]); } } return Promise.resolve(); } catch (ex) { this.logger.debug('readyCacheBullQueue error ', ex); return Promise.reject(ex); } } setCacheBullKey(name) { this.logger.debug('Cache Bull Key : ', name); this.bullQueueClient.set(`bullQueueCache::${name}`, 1) .then() .catch(err => { this.logger.error('bullQueue set cache key error:', name); }) } /** * @description * shuffle算法,类似摸牌 * arr为原数组,cards为乱序结果数组 * random取一个index,取arr中这个元素,放入cards,同时移除arr中这个元素。 * @param originalArray 原始数组 * @return {Array} */ shuffle(originalArray: string[]) { const mixedArray = []; const copyArray = originalArray.slice(0); // 防止改变原来的参数,数组是引用传递 while (copyArray.length > 0) { //generate a random index of the original array const randomIndex = Math.random() * copyArray.length; //push the random element into the mixed one, at the same time, delete the original element mixedArray.push(copyArray[randomIndex]); copyArray.splice(randomIndex, 1); } return mixedArray; } async randomStrategy({ members, tenantId }) { try { const lockedMembers = await this.getLockedMembers({ tenantId }); const unlockedMember = members.filter(x => { return lockedMembers.indexOf(String(x)) === -1 }); const membersRandom = this.shuffle(unlockedMember); const member = this.cycleFind({ members: membersRandom, tenantId }); return Promise.resolve(member); } catch (ex) { this.logger.error('randomStrategy error:', ex); return Promise.reject(ex); } } /** * @description * 按队列中坐席的顺序从上到下一次查找可用的坐席 * 为什么要按members坐席的顺序来,因为实际应用中,这个顺序通常可以用来代表坐
import { EventService } from '../service/EventService'; import { TenantController } from '../controllers/tenant';
random_line_split
op.rs
impl<F: Float> DummyOp<F> { #[allow(dead_code)] pub(crate) fn new() -> Self { DummyOp { phantom: PhantomData, } } } impl<F: Float> Op<F> for DummyOp<F> { fn compute(&self, _: &mut ComputeContext<F>) -> Result<(), OpError> { Ok(()) } fn grad(&self, _: &mut GradientContext<F>) {} } /// Wrapper for NdArrayView/NdArrayViewMut which is fed to `Op::compute` /// /// Used in `Op::ComputeContext`. pub(crate) enum OpInput<'v, T: Float> { NonVariable(Option<NdArrayView<'v, T>>), RdOnlyVariable(Option<NdArrayView<'v, T>>), RdWrVariable(Option<NdArrayViewMut<'v, T>>), } /// `Op::compute`'s output #[derive(Clone)] pub(crate) enum OpOutput<T: Float> { Owned(NdArray<T>), View(RawNdArrayView<T>), } impl<'view, T: Float> OpInput<'view, T> { #[inline] /// Make a read-only input array pub fn new_non_variable(x: NdArrayView<'view, T>) -> Self { NonVariable(Some(x)) } #[inline] /// Make a read-only input array pub fn new_rdonly_variable(x: NdArrayView<'view, T>) -> Self { OpInput::RdOnlyVariable(Some(x)) } #[inline] /// Make a read/write input array pub fn new_rdwr_variable(x: NdArrayViewMut<'view, T>) -> Self { OpInput::RdWrVariable(Some(x)) } } /// Context of an `Op`'s computation phase. /// /// # Example /// /// ``` /// use autograd as ag; /// /// // Implementing `Op` trait for `Sigmoid`. /// struct Sigmoid; /// /// impl<T: ag::Float> ag::op::Op<T> for Sigmoid { /// fn compute( /// &self, /// ctx: &mut ag::op::ComputeContext<T>, /// ) -> Result<(), ag::op::OpError> { /// // Getting the first input array. /// let x: &ag::NdArrayView<_> = &ctx.input(0); /// let half = T::from(0.5).unwrap(); /// let y = x.mapv(move |a| ((a * half).tanh() * half) + half); /// // Put the computed result. /// ctx.append_output(y); /// Ok(()) /// } /// /// fn grad(&self, ctx: &mut ag::op::GradientContext<T>) { /* ... */ } /// } /// ``` pub struct ComputeContext<'v, T: Float> { // Input arrays xs: SmallVec<OpInput<'v, T>>, // Output arrays pub(crate) ys: SmallVec<OpOutput<T>>, } impl<'graph, 'view, T: Float> ComputeContext<'view, T> { #[inline] pub(crate) fn new(xs: SmallVec<OpInput<'view, T>>) -> Self { ComputeContext { xs, ys: SmallVec::new(), } } /// Grabs the `i` th input array as a *read-only* array view. /// /// Calling `input(i)` more than once causes panic. #[inline] pub fn input(&mut self, i: usize) -> NdArrayView<'view, T> { let x = match self.xs.get_mut(i) { Some(x) => x, None => panic!("Bad op impl: input index out of range."), }; match x { NonVariable(ref mut a) => match a.take() { Some(ret) => ret, None => panic!( "Bad op impl: input({})/input_mut({}) cannot be called twice", i, i ), }, OpInput::RdOnlyVariable(ref mut a) => match a.take() { Some(ret) => ret, None => panic!( "Bad op impl: input({})/input_mut({}) cannot be called twice", i, i ), }, OpInput::RdWrVariable(_) => { panic!( "Bad op impl: cannot perform mutable borrowing for input({}). Use input_mut() instead.", i ); } } } /// Grabs the `i` th input array as a *read-write* array view. /// /// Calling `input_mut(i)` more than once causes panic. #[inline] pub fn input_mut(&mut self, i: usize) -> NdArrayViewMut<'view, T> { let x = match self.xs.get_mut(i) { Some(x) => x, None => panic!("Bad op impl: {}'s input doesn't exist.", i), }; match x { OpInput::RdWrVariable(ref mut a) => match a.take() { Some(ret) => ret, None => panic!( "Bad op impl: input({})/input_mut({}) cannot be called twice", i, i ), }, _ => { panic!( "Bad op impl: cannot perform mutable borrowing for input({})", i ); } } } /// Appends an `ndarray::ArrayView` to the back of the output list of the current op. /// /// NOTE: Implementor of `Op::compute` must not forget to call `append_*` as many as the number of its output in `Op::compute`, otherwise panic occurs. #[inline] pub fn append_output_view(&mut self, y: NdArrayView<'view, T>) { self.append_output_view_raw(y.raw_view()); } /// Appends an `ndarray::ArrayView` to the back of the output list of the current op. /// /// NOTE: Implementor of `Op::compute` must not forget to call `append_*` as many as the number of its output in `Op::compute`, otherwise panic occurs. #[inline] pub(crate) fn append_output_view_raw(&mut self, y: RawNdArrayView<T>) { let mut contains_variable_input= false; for x in &self.xs { match x { NonVariable(_) => {}, _ => contains_variable_input = true } } if contains_variable_input { // copy it beforehand to avoid use-after-free self.ys.push(OpOutput::Owned(unsafe { y.deref_into_view().to_owned() })); } else { self.ys.push(OpOutput::View(y)); } } #[inline] pub fn append_empty_output(&mut self) { self.ys.push(OpOutput::Owned(NdArray::zeros( crate::ndarray::IxDyn(&[]), ))); } /// Appends an ndarray to the back of the output list of the current op. /// /// NOTE: Implementor of `Op::compute` must not forget to call `append_*` as many as the number of its output in `Op::compute`, otherwise panic occurs. #[inline] pub fn append_output(&mut self, y: NdArray<T>) { self.ys.push(OpOutput::Owned(y)); } /// Returns a number of input arrays. #[inline] pub fn num_inputs(&self) -> usize { self.xs.len() } } /// Context of an `Op`'s gradient propagation phase. /// /// This is passed to an `Op` through `Op::grad`. /// `Op::grad` should provide the gradients of its inputs by calling `GradientContext::append_input_grad`. /// /// Use `graph()` to access `Graph` object for tensor computations. /// /// ``` /// use autograd as ag; /// use ag::tensor_ops as T; /// /// struct Sigmoid; /// /// impl<F: ag::Float> ag::op::Op<F> for Sigmoid { /// fn compute(&self, ctx: &mut ag::op::ComputeContext<F>) -> Result<(), ag::op::OpError> { /// /* ... */ /// Ok(()) /// } /// /// fn grad(&self, ctx: &mut ag::op::GradientContext<F>) { /// // gradient of the input of Sigmoid /// let gy = ctx.output_grad(); /// // output tensor /// let y = ctx.output(); /// // `Tensor` computations /// let gx = gy * (y - T::square(y)); /// // Propagates input's gradient. /// ctx.append_input_grad(Some(gx)); /// } /// } /// ``` pub struct GradientContext<'graph, T: Float> { gy: Tensor<'graph, T>, y: Tensor<'graph, T>, graph: &'graph crate::graph::Graph<T>, gxs: SmallVec<Option<Tensor<'graph, T>>>, } impl<'graph, T: Float> GradientContext<'graph, T> { #[inline] pub(crate) fn new( gy: Tensor<'graph, T>, y: Tensor<'graph, T>, graph: &'graph crate::graph::Graph<T>, ) -> Self
{ GradientContext { gy, y, graph, gxs: SmallVec::new(), } }
identifier_body
op.rs
//! .build(Sigmoid) //! } //! ``` //! use std::any::type_name; use std::fmt; use std::marker::PhantomData; use std::mem; use crate::ndarray_ext::{NdArrayView, NdArrayViewMut, RawNdArrayView}; use crate::smallvec::SmallVec as RawSmallVec; use crate::tensor::Tensor; use crate::{Float, NdArray}; use crate::op::OpInput::NonVariable; pub(crate) const DEFAULT_NUM_EDGES: usize = 2; pub(crate) type SmallVec<T> = RawSmallVec<[T; DEFAULT_NUM_EDGES]>; /// Error in `Op`'s computation. #[derive(Clone, Debug, PartialEq)] pub enum
{ NdArrayError(String, ndarray::ShapeError), IncompatibleShape(String), TypeUnsupported(String), InvalidDims(String), OutOfBounds(String), } impl std::error::Error for OpError {} impl fmt::Display for OpError { fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { match self { OpError::NdArrayError(pref, e) => write!(f, "{}: ", pref).and_then(|()| e.fmt(f)), OpError::IncompatibleShape(s) => write!(f, "{}: ", s), OpError::TypeUnsupported(s) => write!(f, "{}: ", s), OpError::InvalidDims(s) => write!(f, "{}: ", s), OpError::OutOfBounds(s) => write!(f, "{}: ", s), } } } /// Trait for tensor operations. `Tensor` structs wrap this. pub trait Op<F: Float> { /// Name of this op fn name(&self) -> &'static str { type_name::<Self>() } /// Runs this op with `ComputeContext`. fn compute(&self, ctx: &mut ComputeContext<F>) -> Result<(), OpError>; /// Returns gradients for input nodes by use of output's gradients etc. fn grad(&self, ctx: &mut GradientContext<F>); } pub(crate) struct DummyOp<F: Float> { pub phantom: PhantomData<F>, } impl<F: Float> DummyOp<F> { #[allow(dead_code)] pub(crate) fn new() -> Self { DummyOp { phantom: PhantomData, } } } impl<F: Float> Op<F> for DummyOp<F> { fn compute(&self, _: &mut ComputeContext<F>) -> Result<(), OpError> { Ok(()) } fn grad(&self, _: &mut GradientContext<F>) {} } /// Wrapper for NdArrayView/NdArrayViewMut which is fed to `Op::compute` /// /// Used in `Op::ComputeContext`. pub(crate) enum OpInput<'v, T: Float> { NonVariable(Option<NdArrayView<'v, T>>), RdOnlyVariable(Option<NdArrayView<'v, T>>), RdWrVariable(Option<NdArrayViewMut<'v, T>>), } /// `Op::compute`'s output #[derive(Clone)] pub(crate) enum OpOutput<T: Float> { Owned(NdArray<T>), View(RawNdArrayView<T>), } impl<'view, T: Float> OpInput<'view, T> { #[inline] /// Make a read-only input array pub fn new_non_variable(x: NdArrayView<'view, T>) -> Self { NonVariable(Some(x)) } #[inline] /// Make a read-only input array pub fn new_rdonly_variable(x: NdArrayView<'view, T>) -> Self { OpInput::RdOnlyVariable(Some(x)) } #[inline] /// Make a read/write input array pub fn new_rdwr_variable(x: NdArrayViewMut<'view, T>) -> Self { OpInput::RdWrVariable(Some(x)) } } /// Context of an `Op`'s computation phase. /// /// # Example /// /// ``` /// use autograd as ag; /// /// // Implementing `Op` trait for `Sigmoid`. /// struct Sigmoid; /// /// impl<T: ag::Float> ag::op::Op<T> for Sigmoid { /// fn compute( /// &self, /// ctx: &mut ag::op::ComputeContext<T>, /// ) -> Result<(), ag::op::OpError> { /// // Getting the first input array. /// let x: &ag::NdArrayView<_> = &ctx.input(0); /// let half = T::from(0.5).unwrap(); /// let y = x.mapv(move |a| ((a * half).tanh() * half) + half); /// // Put the computed result. /// ctx.append_output(y); /// Ok(()) /// } /// /// fn grad(&self, ctx: &mut ag::op::GradientContext<T>) { /* ... */ } /// } /// ``` pub struct ComputeContext<'v, T: Float> { // Input arrays xs: SmallVec<OpInput<'v, T>>, // Output arrays pub(crate) ys: SmallVec<OpOutput<T>>, } impl<'graph, 'view, T: Float> ComputeContext<'view, T> { #[inline] pub(crate) fn new(xs: SmallVec<OpInput<'view, T>>) -> Self { ComputeContext { xs, ys: SmallVec::new(), } } /// Grabs the `i` th input array as a *read-only* array view. /// /// Calling `input(i)` more than once causes panic. #[inline] pub fn input(&mut self, i: usize) -> NdArrayView<'view, T> { let x = match self.xs.get_mut(i) { Some(x) => x, None => panic!("Bad op impl: input index out of range."), }; match x { NonVariable(ref mut a) => match a.take() { Some(ret) => ret, None => panic!( "Bad op impl: input({})/input_mut({}) cannot be called twice", i, i ), }, OpInput::RdOnlyVariable(ref mut a) => match a.take() { Some(ret) => ret, None => panic!( "Bad op impl: input({})/input_mut({}) cannot be called twice", i, i ), }, OpInput::RdWrVariable(_) => { panic!( "Bad op impl: cannot perform mutable borrowing for input({}). Use input_mut() instead.", i ); } } } /// Grabs the `i` th input array as a *read-write* array view. /// /// Calling `input_mut(i)` more than once causes panic. #[inline] pub fn input_mut(&mut self, i: usize) -> NdArrayViewMut<'view, T> { let x = match self.xs.get_mut(i) { Some(x) => x, None => panic!("Bad op impl: {}'s input doesn't exist.", i), }; match x { OpInput::RdWrVariable(ref mut a) => match a.take() { Some(ret) => ret, None => panic!( "Bad op impl: input({})/input_mut({}) cannot be called twice", i, i ), }, _ => { panic!( "Bad op impl: cannot perform mutable borrowing for input({})", i ); } } } /// Appends an `ndarray::ArrayView` to the back of the output list of the current op. /// /// NOTE: Implementor of `Op::compute` must not forget to call `append_*` as many as the number of its output in `Op::compute`, otherwise panic occurs. #[inline] pub fn append_output_view(&mut self, y: NdArrayView<'view, T>) { self.append_output_view_raw(y.raw_view()); } /// Appends an `ndarray::ArrayView` to the back of the output list of the current op. /// /// NOTE: Implementor of `Op::compute` must not forget to call `append_*` as many as the number of its output in `Op::compute`, otherwise panic occurs. #[inline] pub(crate) fn append_output_view_raw(&mut self, y: RawNdArrayView<T>) { let mut contains_variable_input= false; for x in &self.xs { match x { NonVariable(_) => {}, _ => contains_variable_input = true } } if contains_variable_input { // copy it beforehand to avoid use-after-free self.ys.push(OpOutput::Owned(unsafe { y.deref_into_view().to_owned() })); } else { self.ys.push(OpOutput::View(y)); } } #[inline] pub fn append_empty_output(&mut self) { self.ys.push(OpOutput::Owned(NdArray::zeros( crate::ndarray::IxDyn(&[]), ))); } /// Appends an ndarray to the back of the output list of the current op. /// /// NOTE: Implementor of `Op::compute` must not forget to call `append_*` as many as the number of its output in `Op::compute`, otherwise panic occurs. #[inline]
OpError
identifier_name
op.rs
//! .build(Sigmoid) //! } //! ``` //! use std::any::type_name; use std::fmt; use std::marker::PhantomData; use std::mem; use crate::ndarray_ext::{NdArrayView, NdArrayViewMut, RawNdArrayView}; use crate::smallvec::SmallVec as RawSmallVec; use crate::tensor::Tensor; use crate::{Float, NdArray}; use crate::op::OpInput::NonVariable; pub(crate) const DEFAULT_NUM_EDGES: usize = 2; pub(crate) type SmallVec<T> = RawSmallVec<[T; DEFAULT_NUM_EDGES]>; /// Error in `Op`'s computation. #[derive(Clone, Debug, PartialEq)] pub enum OpError { NdArrayError(String, ndarray::ShapeError), IncompatibleShape(String), TypeUnsupported(String), InvalidDims(String), OutOfBounds(String), } impl std::error::Error for OpError {} impl fmt::Display for OpError { fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { match self { OpError::NdArrayError(pref, e) => write!(f, "{}: ", pref).and_then(|()| e.fmt(f)), OpError::IncompatibleShape(s) => write!(f, "{}: ", s), OpError::TypeUnsupported(s) => write!(f, "{}: ", s), OpError::InvalidDims(s) => write!(f, "{}: ", s), OpError::OutOfBounds(s) => write!(f, "{}: ", s), } } } /// Trait for tensor operations. `Tensor` structs wrap this. pub trait Op<F: Float> { /// Name of this op fn name(&self) -> &'static str { type_name::<Self>() } /// Runs this op with `ComputeContext`. fn compute(&self, ctx: &mut ComputeContext<F>) -> Result<(), OpError>; /// Returns gradients for input nodes by use of output's gradients etc. fn grad(&self, ctx: &mut GradientContext<F>); } pub(crate) struct DummyOp<F: Float> { pub phantom: PhantomData<F>, } impl<F: Float> DummyOp<F> { #[allow(dead_code)] pub(crate) fn new() -> Self { DummyOp { phantom: PhantomData, } } } impl<F: Float> Op<F> for DummyOp<F> { fn compute(&self, _: &mut ComputeContext<F>) -> Result<(), OpError> { Ok(()) } fn grad(&self, _: &mut GradientContext<F>) {} } /// Wrapper for NdArrayView/NdArrayViewMut which is fed to `Op::compute` /// /// Used in `Op::ComputeContext`. pub(crate) enum OpInput<'v, T: Float> { NonVariable(Option<NdArrayView<'v, T>>), RdOnlyVariable(Option<NdArrayView<'v, T>>), RdWrVariable(Option<NdArrayViewMut<'v, T>>), } /// `Op::compute`'s output #[derive(Clone)] pub(crate) enum OpOutput<T: Float> { Owned(NdArray<T>), View(RawNdArrayView<T>), } impl<'view, T: Float> OpInput<'view, T> { #[inline] /// Make a read-only input array pub fn new_non_variable(x: NdArrayView<'view, T>) -> Self { NonVariable(Some(x)) } #[inline] /// Make a read-only input array pub fn new_rdonly_variable(x: NdArrayView<'view, T>) -> Self { OpInput::RdOnlyVariable(Some(x)) } #[inline] /// Make a read/write input array pub fn new_rdwr_variable(x: NdArrayViewMut<'view, T>) -> Self { OpInput::RdWrVariable(Some(x)) } } /// Context of an `Op`'s computation phase. /// /// # Example /// /// ``` /// use autograd as ag; /// /// // Implementing `Op` trait for `Sigmoid`. /// struct Sigmoid; /// /// impl<T: ag::Float> ag::op::Op<T> for Sigmoid { /// fn compute( /// &self, /// ctx: &mut ag::op::ComputeContext<T>, /// ) -> Result<(), ag::op::OpError> { /// // Getting the first input array. /// let x: &ag::NdArrayView<_> = &ctx.input(0); /// let half = T::from(0.5).unwrap(); /// let y = x.mapv(move |a| ((a * half).tanh() * half) + half); /// // Put the computed result. /// ctx.append_output(y); /// Ok(()) /// } /// /// fn grad(&self, ctx: &mut ag::op::GradientContext<T>) { /* ... */ } /// } /// ``` pub struct ComputeContext<'v, T: Float> { // Input arrays xs: SmallVec<OpInput<'v, T>>, // Output arrays pub(crate) ys: SmallVec<OpOutput<T>>, } impl<'graph, 'view, T: Float> ComputeContext<'view, T> { #[inline] pub(crate) fn new(xs: SmallVec<OpInput<'view, T>>) -> Self { ComputeContext { xs, ys: SmallVec::new(), } } /// Grabs the `i` th input array as a *read-only* array view. /// /// Calling `input(i)` more than once causes panic. #[inline] pub fn input(&mut self, i: usize) -> NdArrayView<'view, T> { let x = match self.xs.get_mut(i) { Some(x) => x, None => panic!("Bad op impl: input index out of range."), }; match x { NonVariable(ref mut a) => match a.take() { Some(ret) => ret, None => panic!( "Bad op impl: input({})/input_mut({}) cannot be called twice", i, i ), }, OpInput::RdOnlyVariable(ref mut a) => match a.take() { Some(ret) => ret,
i, i ), }, OpInput::RdWrVariable(_) => { panic!( "Bad op impl: cannot perform mutable borrowing for input({}). Use input_mut() instead.", i ); } } } /// Grabs the `i` th input array as a *read-write* array view. /// /// Calling `input_mut(i)` more than once causes panic. #[inline] pub fn input_mut(&mut self, i: usize) -> NdArrayViewMut<'view, T> { let x = match self.xs.get_mut(i) { Some(x) => x, None => panic!("Bad op impl: {}'s input doesn't exist.", i), }; match x { OpInput::RdWrVariable(ref mut a) => match a.take() { Some(ret) => ret, None => panic!( "Bad op impl: input({})/input_mut({}) cannot be called twice", i, i ), }, _ => { panic!( "Bad op impl: cannot perform mutable borrowing for input({})", i ); } } } /// Appends an `ndarray::ArrayView` to the back of the output list of the current op. /// /// NOTE: Implementor of `Op::compute` must not forget to call `append_*` as many as the number of its output in `Op::compute`, otherwise panic occurs. #[inline] pub fn append_output_view(&mut self, y: NdArrayView<'view, T>) { self.append_output_view_raw(y.raw_view()); } /// Appends an `ndarray::ArrayView` to the back of the output list of the current op. /// /// NOTE: Implementor of `Op::compute` must not forget to call `append_*` as many as the number of its output in `Op::compute`, otherwise panic occurs. #[inline] pub(crate) fn append_output_view_raw(&mut self, y: RawNdArrayView<T>) { let mut contains_variable_input= false; for x in &self.xs { match x { NonVariable(_) => {}, _ => contains_variable_input = true } } if contains_variable_input { // copy it beforehand to avoid use-after-free self.ys.push(OpOutput::Owned(unsafe { y.deref_into_view().to_owned() })); } else { self.ys.push(OpOutput::View(y)); } } #[inline] pub fn append_empty_output(&mut self) { self.ys.push(OpOutput::Owned(NdArray::zeros( crate::ndarray::IxDyn(&[]), ))); } /// Appends an ndarray to the back of the output list of the current op. /// /// NOTE: Implementor of `Op::compute` must not forget to call `append_*` as many as the number of its output in `Op::compute`, otherwise panic occurs. #[inline] pub
None => panic!( "Bad op impl: input({})/input_mut({}) cannot be called twice",
random_line_split
de.sc.portal.env.dlg.mongo.js
rendererOptions: { showDataLabels: true, dataLabels: 'value', sliceMargin: 2, startAngle: -90, innerDiameter: 5, ringMargin: 5, shadow: true, showMarker: true, markerOptions: { show: true }, pointLabels: { show: true, location: 's' }, smooth: true, animation: { show: true, speed: 2500 }, shadowDepth:11, numberRows:3, numberColumns:3, //dataLabelThreshold:0, dataLabels : 'percent' //‘label’, ‘value’, ‘percent’ } }, highlighter: { show: false, sizeAdjust: 7.5 }, cursor: { show: false, zoom: true, showTooltip: true }, legend: { renderer: $.jqplot.DonutLegendRenderer, //renderer: $.jqplot.EnhancedLegendRenderer, numberRows:3, numberColumns:3, show: true, placement: 'outsideGrid', location: 'e' }, grid: { background: 'white' } }; var data = new Array(); { var dbs = jsondata["databases"]; var sum=0; for(var i=0;i<dbs.length;i++) { var db = dbs[i]; var sizeOnDisk = de.sc.portal.env.dlg.mongo.getSizeAsReadableString(db["sizeOnDisk"]); var size = db["empty"] ? " (is empty)" : " (size: "+sizeOnDisk+")"; var thisdata = new Array(); thisdata.push(db["name"] + (db["empty"] ? " (is empty)" : "")); thisdata.push(db["sizeOnDisk"]); thisdata.push(size); data.push(thisdata); sum += db["sizeOnDisk"]; } options.title = options.title + " (" +de.sc.portal.env.dlg.mongo.getSizeAsReadableString(sum)+")"; } de.sc.portal.env.dlg.mongo.plot = jQuery.jqplot ('mongodbschart', [data], options); $('div.mongodbschart-container').resizable({delay: 20}); $('div.mongodbschart-container').bind('resize', function(event, ui) { de.sc.portal.env.dlg.mongo.plot.replot(); }); $('#mongodbschart').bind('jqplotDataHighlight', function (ev, seriesIndex, pointIndex, data) { $('#mongodbschartinfo').html(data[0]+" "+data[2]); var mouseX = ev.pageX; var mouseY = ev.pageY; var o = $("#dlg_mongo").offset(); mouseX -= o.left; mouseY -= o.top; var cssObj = { position : 'absolute', 'font-weight' : 'bold', left : mouseX + 'px', top : mouseY + 'px', color:'black' }; $('#mongodbschartinfo').css(cssObj); }); $('#mongodbschart').bind('jqplotDataUnhighlight', function (ev, seriesIndex, pointIndex, data) { $('#mongodbschartinfo').html(""); }); $(".jqplot-table-legend", "#dlg_mongo").css("text-align","left").css("background-color","white").css("font-size","9px"); $("table.jqplot-table-legend", "#dlg_mongo").css("top","0px"); }); } return true; }; de.sc.portal.env.dlg.mongo.isLastTabActivated = function(ui) { var next = ui.newTab.next().size(); if (next > 0)return false; $("#cols, #fields","#dlg_mongo_tabs").html(" "); var select = $("#dbs","#dlg_mongo_tabs"); select.next().show(); var url = de.sc.portal.env.dlg.mongo.mu.d()+"listDatabases"; $.ajax ({ url: ((window.location.href.indexOf("http") === 0) ? ("/cgi-bin/get.cgi?"+escape(url)) : "_listDatabases"), dataType:"json" }) .success(function(data) { var dbs = data["databases"]; if(select.size() < 1) { de.sc.log("cisLastTabActivated:\n!select"+ui.newTab.html(), de.sc.ERR); return; } select.html("<option></option>"); for(var i=0;i<dbs.length;i++) { var db = dbs[i]; var sizeOnDisk = de.sc.portal.env.dlg.mongo.getSizeAsReadableString(db["sizeOnDisk"]); var size = db["empty"] ? " (is empty)" : " (size: "+sizeOnDisk+")"; var o = new Option(db["name"]+size, db["name"], false, false); select[0][select[0].options.length] = o; } select.next().hide(); }) .error(function(xhr, statusmsg, err) { de.sc.log("\ncisLastTabActivated:\nstatus:'"+statusmsg+"'\nerror:'"+err+"'\nurl: '"+url+"'\nthis:'"+window.location.href+"'", de.sc.WRN); }); return true; }; de.sc.portal.env.dlg.mongo.getSizeAsReadableString = function(sizeAsNumber) { if((sizeAsNumber === null) || (sizeAsNumber === 0) || (sizeAsNumber === ""))return sizeAsNumber; if(sizeAsNumber < 1024) return sizeAsNumber + " bytes"; if(sizeAsNumber < 1024*1024) return (sizeAsNumber / 1024).toFixed(2) + " kb"; if(sizeAsNumber < 1024*1024*1024) return (sizeAsNumber / ( 1024*1024)).toFixed(2) + " mb"; if(sizeAsNumber < 1024*1024*1024*1024) return (sizeAsNumber / ( 1024*1024*1024)).toFixed(2) + " gb"; if(sizeAsNumber < 1024*1024*1024*1024*1024) return (sizeAsNumber / ( 1024*1024*1024*1024)).toFixed(2) + " tb"; de.sc.log("\ngetSizeAsReadableString:\nunexpected nr:"+sizeAsNumber, de.sc.WRN); return sizeAsNumber; }; de.sc.portal.env.dlg.mongo.onSelectCols = function(eSelect) { $("#fields","#dlg_mongo_tabs").html(" "); var db = $("#dbs","#dlg_mongo_tabs")[0].options[$("#dbs","#dlg_mongo_tabs")[0].selectedIndex].value; if(de.sc.isempty(db))return; var col = eSelect.options[eSelect.selectedIndex].value; if(de.sc.isempty(col))return; var select = $("#fields","#dlg_mongo_tabs"); select.next().show(); var url = de.sc.portal.env.dlg.mongo.mu.d()+db+"/"+col+"/?limit=1"; $.ajax ({ url: ((window.location.href.indexOf("http") === 0) ? ("/cgi-bin/get_with_params.cgi?"+escape(url)) : "_items"), dataType:"json" }) .success(function(data) { var item = data["rows"][0]; if (select.size() < 1) { de.sc.log("onSelectCols:\n!select", de.sc.ERR); return; } select.html("<option></option>"); for(var attrName in item) { if(attrName.indexOf('_') === 0)continue; //var attrValue = item[attrName]; var o = new Option(attrName, attrName, false, false); select[0][select[0].options.length] = o; } select.next().hide(); }) .error(function(xhr, statusmsg, err) { select.next().hide(); de.sc.log("\nonSelectCols:\nstatus:'"+statusmsg+"'\nerror:'"+err+"'\nurl: '"+url+"'\nthis:'"+window.location.href+"'", de.sc.WRN); }); }; de.sc.portal.env.dlg.mongo.onSearch = function(searchB) { var db = $("#dbs" ,"#dlg_mongo_tabs")[0].options[$("#dbs" ,"#dlg_mongo_tabs")[0].selectedIndex].value; var col = $("#cols" ,"#dlg_mongo_tabs")[0].options[$("#cols" ,"#dlg_mongo_tabs")[0].selectedIndex].value; var fld = $("#fields","#dlg_mongo_tabs")[0].options[$("#fields","#dlg_mongo_tabs")[0].selectedIndex].value; if(de.sc.isempty(db) || de.sc
random_line_split
mapfile.rs
State::Base } else { TileState::Generated } }, ) } pub(crate) fn read_texture( &self, device: &wgpu::Device, encoder: &mut wgpu::CommandEncoder, name: &str, ) -> Result<wgpu::Texture, Error> { let desc = self.lookup_texture(name)?.unwrap(); let texture = device.create_texture(&wgpu::TextureDescriptor { size: wgpu::Extent3d { width: desc.width, height: desc.height, depth: desc.depth }, format: desc.format.to_wgpu(), mip_level_count: 1, sample_count: 1, dimension: if desc.depth == 1 { wgpu::TextureDimension::D2 } else { wgpu::TextureDimension::D3 }, usage: wgpu::TextureUsage::COPY_SRC | wgpu::TextureUsage::COPY_DST | wgpu::TextureUsage::SAMPLED | wgpu::TextureUsage::STORAGE, label: None, }); let (width, height) = (desc.width as usize, (desc.height * desc.depth) as usize); assert_eq!(width % desc.format.block_size() as usize, 0); assert_eq!(height % desc.format.block_size() as usize, 0); let (width, height) = (width / desc.format.block_size() as usize, height / desc.format.block_size() as usize); let row_bytes = width * desc.format.bytes_per_block(); let row_pitch = (row_bytes + 255) & !255; let data = if desc.format == TextureFormat::RGBA8 { image::open(TERRA_DIRECTORY.join(format!("{}.bmp", name)))?.to_rgba().into_vec() } else { fs::read(TERRA_DIRECTORY.join(format!("{}.raw", name)))? }; let buffer = device.create_buffer(&wgpu::BufferDescriptor { size: (row_pitch * height) as u64, usage: wgpu::BufferUsage::MAP_WRITE | wgpu::BufferUsage::COPY_SRC, label: None, mapped_at_creation: true, }); let mut buffer_view = buffer.slice(..).get_mapped_range_mut(); for row in 0..height { buffer_view[row * row_pitch..][..row_bytes] .copy_from_slice(&data[row * row_bytes..][..row_bytes]); } drop(buffer_view); buffer.unmap(); encoder.copy_buffer_to_texture( wgpu::BufferCopyView { buffer: &buffer, layout: wgpu::TextureDataLayout { offset: 0, bytes_per_row: row_pitch as u32, rows_per_image: height as u32 / desc.depth, }, }, wgpu::TextureCopyView { texture: &texture, mip_level: 0, origin: wgpu::Origin3d { x: 0, y: 0, z: 0 }, }, wgpu::Extent3d { width: width as u32, height: height as u32 / desc.depth, depth: desc.depth, }, ); Ok(texture) } pub(crate) fn write_texture( &self, name: &str, desc: TextureDescriptor, data: &[u8], ) -> Result<(), Error> { self.update_texture(name, desc)?; if desc.format == TextureFormat::RGBA8 { let filename = TERRA_DIRECTORY.join(format!("{}.bmp", name)); Ok(image::save_buffer_with_format( &filename, data, desc.width, desc.height * desc.depth, image::ColorType::Rgba8, image::ImageFormat::Bmp, )?) } else { let filename = TERRA_DIRECTORY.join(format!("{}.raw", name)); Ok(fs::write(&filename, data)?) } } pub(crate) fn reload_texture(&self, name: &str) -> bool { let desc = self.lookup_texture(name); if let Ok(Some(desc)) = desc { if desc.format == TextureFormat::RGBA8 { TERRA_DIRECTORY.join(format!("{}.bmp", name)).exists() } else { TERRA_DIRECTORY.join(format!("{}.raw", name)).exists() } } else { false } } pub(crate) fn layers(&self) -> &VecMap<LayerParams> { &self.layers } pub(crate) fn tile_name(layer: LayerType, node: VNode) -> PathBuf { let face = match node.face() { 0 => "0E", 1 => "180E", 2 => "90E", 3 => "90W", 4 => "N", 5 => "S", _ => unreachable!(), }; let (layer, ext) = match layer { LayerType::Displacements => ("displacements", "raw"), LayerType::Albedo => ("albedo", "bmp"), LayerType::Roughness => ("roughness", "raw"), LayerType::Normals => ("normals", "raw"), LayerType::Heightmaps => ("heightmaps", "raw.sz"), }; TERRA_DIRECTORY.join(&format!( "tiles/{}_{}_{}_{}x{}.{}", layer, node.level(), face, node.x(), node.y(), ext )) } pub(crate) fn reload_tile_state( &self, layer: LayerType, node: VNode, base: bool, ) -> Result<TileState, Error> { let filename = Self::tile_name(layer, node); let meta = self.lookup_tile_meta(layer, node); let exists = filename.exists(); let target_state = if base && exists { TileState::Base } else if base { TileState::MissingBase } else if exists { TileState::Generated } else { TileState::Missing }; if let Ok(Some(TileMeta { state, .. })) = meta { if state == target_state { return Ok(state); } } let new_meta = TileMeta { state: target_state, crc32: 0 }; self.update_tile_meta(layer, node, new_meta)?; Ok(target_state) } // pub(crate) fn set_missing( // &self, // layer: LayerType, // node: VNode, // base: bool, // ) -> Result<(), Error> { // let state = if base { TileState::MissingBase } else { TileState::Missing }; // self.update_tile_meta(layer, node, TileMeta { crc32: 0, state }) // } pub(crate) fn clear_generated(&mut self, layer: LayerType) -> Result<(), Error> { self.scan_tile_meta(layer, |node, meta| { if let TileState::Generated = meta.state { self.remove_tile_meta(layer, node)?; } Ok(()) }) } pub(crate) fn get_missing_base(&self, layer: LayerType) -> Result<Vec<VNode>, Error> { let mut missing = Vec::new(); self.scan_tile_meta(layer, |node, meta| { if let TileState::MissingBase = meta.state { missing.push(node); } Ok(()) })?; Ok(missing) } // // These functions use the database. // fn lookup_tile_meta(&self, layer: LayerType, node: VNode) -> Result<Option<TileMeta>, Error> { let key = bincode::serialize(&(layer, node)).unwrap(); Ok(self.tiles.get(key)?.map(|value| bincode::deserialize(&value).unwrap())) } fn update_tile_meta(&self, layer: LayerType, node: VNode, meta: TileMeta) -> Result<(), Error> { let key = bincode::serialize(&(layer, node)).unwrap(); let value = bincode::serialize(&meta).unwrap(); self.tiles.insert(key, value)?; Ok(()) } fn remove_tile_meta(&self, layer: LayerType, node: VNode) -> Result<(), Error> { let key = bincode::serialize(&(layer, node)).unwrap(); self.tiles.remove(key)?; Ok(()) } fn scan_tile_meta<F: FnMut(VNode, TileMeta) -> Result<(), Error>>( &self, layer: LayerType, mut f: F, ) -> Result<(), Error> { let prefix = bincode::serialize(&layer).unwrap(); for i in self.tiles.scan_prefix(&prefix) { let (k, v) = i?; let meta = bincode::deserialize::<TileMeta>(&v)?; let node = bincode::deserialize::<(LayerType, VNode)>(&k)?.1; f(node, meta)?; } Ok(()) } fn lookup_texture(&self, name: &str) -> Result<Option<TextureDescriptor>, Error> { Ok(self.textures.get(name)?.map(|value| serde_json::from_slice(&value).unwrap())) } fn update_texture(&self, name: &str, desc: TextureDescriptor) -> Result<(), Error>
{ let value = serde_json::to_vec(&desc).unwrap(); self.textures.insert(name, value)?; Ok(()) }
identifier_body
mapfile.rs
(&directory).expect(&format!( "Failed to open/create sled database. Deleting the '{}' directory may fix this", directory.display() )); db.insert("version", "1").unwrap(); Self { layers, tiles: db.open_tree("tiles").unwrap(), textures: db.open_tree("textures").unwrap(), _db: db, } } pub(crate) fn tile_state(&self, layer: LayerType, node: VNode) -> Result<TileState, Error> { Ok(match self.lookup_tile_meta(layer, node)? { Some(meta) => meta.state, None => TileState::GpuOnly, }) } pub(crate) fn read_tile(&self, layer: LayerType, node: VNode) -> Option<Vec<u8>> { let filename = Self::tile_name(layer, node); if !filename.exists() { return None; } match layer { LayerType::Albedo => Some(image::open(filename).ok()?.to_rgba().into_vec()), LayerType::Heightmaps => { let mut data = Vec::new(); snap::read::FrameDecoder::new(BufReader::new(File::open(filename).ok()?)) .read_to_end(&mut data) .ok()?; let mut qdata = vec![0i16; data.len() / 2]; bytemuck::cast_slice_mut(&mut qdata).copy_from_slice(&data); let mut prev = 0; let mut fdata = vec![0f32; qdata.len()]; for (f, q) in fdata.iter_mut().zip(qdata.iter()) { let x = (*q).wrapping_add(prev); *f = x as f32; prev = x; } data.clear(); data.extend_from_slice(bytemuck::cast_slice(&fdata)); Some(data) } LayerType::Normals | LayerType::Displacements | LayerType::Roughness => { fs::read(filename).ok() } } } pub(crate) fn write_tile( &mut self, layer: LayerType, node: VNode, data: &[u8], base: bool, ) -> Result<(), Error> { let filename = Self::tile_name(layer, node); match layer { LayerType::Albedo => image::save_buffer_with_format( &filename, data, self.layers[layer].texture_resolution as u32, self.layers[layer].texture_resolution as u32, image::ColorType::Rgba8, image::ImageFormat::Bmp, )?, LayerType::Heightmaps => { let data: &[f32] = bytemuck::cast_slice(data); let mut qdata = vec![0i16; data.len()]; let mut prev = 0; for (q, d) in qdata.iter_mut().zip(data.iter()) { let x = ((*d as i16) / 4) * 4; *q = x.wrapping_sub(prev); prev = x; } snap::write::FrameEncoder::new(File::create(filename)?) .write_all(bytemuck::cast_slice(&qdata))?; } LayerType::Normals | LayerType::Displacements | LayerType::Roughness => { fs::write(filename, data)?; } } self.update_tile_meta( layer, node, TileMeta { crc32: 0, state: if base { TileState::Base } else { TileState::Generated } }, ) } pub(crate) fn read_texture( &self, device: &wgpu::Device, encoder: &mut wgpu::CommandEncoder, name: &str, ) -> Result<wgpu::Texture, Error> { let desc = self.lookup_texture(name)?.unwrap(); let texture = device.create_texture(&wgpu::TextureDescriptor { size: wgpu::Extent3d { width: desc.width, height: desc.height, depth: desc.depth }, format: desc.format.to_wgpu(), mip_level_count: 1, sample_count: 1, dimension: if desc.depth == 1 { wgpu::TextureDimension::D2 } else { wgpu::TextureDimension::D3 }, usage: wgpu::TextureUsage::COPY_SRC | wgpu::TextureUsage::COPY_DST | wgpu::TextureUsage::SAMPLED | wgpu::TextureUsage::STORAGE, label: None, }); let (width, height) = (desc.width as usize, (desc.height * desc.depth) as usize); assert_eq!(width % desc.format.block_size() as usize, 0); assert_eq!(height % desc.format.block_size() as usize, 0); let (width, height) = (width / desc.format.block_size() as usize, height / desc.format.block_size() as usize); let row_bytes = width * desc.format.bytes_per_block(); let row_pitch = (row_bytes + 255) & !255; let data = if desc.format == TextureFormat::RGBA8 { image::open(TERRA_DIRECTORY.join(format!("{}.bmp", name)))?.to_rgba().into_vec() } else { fs::read(TERRA_DIRECTORY.join(format!("{}.raw", name)))? }; let buffer = device.create_buffer(&wgpu::BufferDescriptor { size: (row_pitch * height) as u64, usage: wgpu::BufferUsage::MAP_WRITE | wgpu::BufferUsage::COPY_SRC, label: None, mapped_at_creation: true, }); let mut buffer_view = buffer.slice(..).get_mapped_range_mut(); for row in 0..height { buffer_view[row * row_pitch..][..row_bytes] .copy_from_slice(&data[row * row_bytes..][..row_bytes]); } drop(buffer_view); buffer.unmap(); encoder.copy_buffer_to_texture( wgpu::BufferCopyView { buffer: &buffer, layout: wgpu::TextureDataLayout { offset: 0, bytes_per_row: row_pitch as u32, rows_per_image: height as u32 / desc.depth, }, }, wgpu::TextureCopyView { texture: &texture, mip_level: 0, origin: wgpu::Origin3d { x: 0, y: 0, z: 0 }, }, wgpu::Extent3d { width: width as u32, height: height as u32 / desc.depth, depth: desc.depth, }, ); Ok(texture) } pub(crate) fn write_texture( &self, name: &str, desc: TextureDescriptor, data: &[u8], ) -> Result<(), Error> { self.update_texture(name, desc)?; if desc.format == TextureFormat::RGBA8 { let filename = TERRA_DIRECTORY.join(format!("{}.bmp", name)); Ok(image::save_buffer_with_format( &filename, data, desc.width, desc.height * desc.depth, image::ColorType::Rgba8, image::ImageFormat::Bmp, )?) } else { let filename = TERRA_DIRECTORY.join(format!("{}.raw", name)); Ok(fs::write(&filename, data)?) } } pub(crate) fn reload_texture(&self, name: &str) -> bool { let desc = self.lookup_texture(name); if let Ok(Some(desc)) = desc { if desc.format == TextureFormat::RGBA8 { TERRA_DIRECTORY.join(format!("{}.bmp", name)).exists() } else { TERRA_DIRECTORY.join(format!("{}.raw", name)).exists() } } else { false } } pub(crate) fn layers(&self) -> &VecMap<LayerParams> { &self.layers } pub(crate) fn tile_name(layer: LayerType, node: VNode) -> PathBuf { let face = match node.face() { 0 => "0E", 1 => "180E", 2 => "90E", 3 => "90W", 4 => "N", 5 => "S", _ => unreachable!(), }; let (layer, ext) = match layer { LayerType::Displacements => ("displacements", "raw"), LayerType::Albedo => ("albedo", "bmp"), LayerType::Roughness => ("roughness", "raw"), LayerType::Normals => ("normals", "raw"), LayerType::Heightmaps => ("heightmaps", "raw.sz"), }; TERRA_DIRECTORY.join(&format!( "tiles/{}_{}_{}_{}x{}.{}", layer, node.level(), face, node.x(), node.y(), ext )) } pub(crate) fn
( &self, layer: LayerType, node: VNode, base: bool, ) -> Result<TileState, Error> { let filename = Self::tile_name(layer, node); let meta = self.lookup_tile_meta(layer, node); let exists = filename.exists(); let target_state = if base && exists { Tile
reload_tile_state
identifier_name
mapfile.rs
= vec![0i16; data.len()]; let mut prev = 0; for (q, d) in qdata.iter_mut().zip(data.iter()) { let x = ((*d as i16) / 4) * 4; *q = x.wrapping_sub(prev); prev = x; } snap::write::FrameEncoder::new(File::create(filename)?) .write_all(bytemuck::cast_slice(&qdata))?; } LayerType::Normals | LayerType::Displacements | LayerType::Roughness => { fs::write(filename, data)?; } } self.update_tile_meta( layer, node, TileMeta { crc32: 0, state: if base { TileState::Base } else { TileState::Generated } }, ) } pub(crate) fn read_texture( &self, device: &wgpu::Device, encoder: &mut wgpu::CommandEncoder, name: &str, ) -> Result<wgpu::Texture, Error> { let desc = self.lookup_texture(name)?.unwrap(); let texture = device.create_texture(&wgpu::TextureDescriptor { size: wgpu::Extent3d { width: desc.width, height: desc.height, depth: desc.depth }, format: desc.format.to_wgpu(), mip_level_count: 1, sample_count: 1, dimension: if desc.depth == 1 { wgpu::TextureDimension::D2 } else { wgpu::TextureDimension::D3 }, usage: wgpu::TextureUsage::COPY_SRC | wgpu::TextureUsage::COPY_DST | wgpu::TextureUsage::SAMPLED | wgpu::TextureUsage::STORAGE, label: None, }); let (width, height) = (desc.width as usize, (desc.height * desc.depth) as usize); assert_eq!(width % desc.format.block_size() as usize, 0); assert_eq!(height % desc.format.block_size() as usize, 0); let (width, height) = (width / desc.format.block_size() as usize, height / desc.format.block_size() as usize); let row_bytes = width * desc.format.bytes_per_block(); let row_pitch = (row_bytes + 255) & !255; let data = if desc.format == TextureFormat::RGBA8 { image::open(TERRA_DIRECTORY.join(format!("{}.bmp", name)))?.to_rgba().into_vec() } else { fs::read(TERRA_DIRECTORY.join(format!("{}.raw", name)))? }; let buffer = device.create_buffer(&wgpu::BufferDescriptor { size: (row_pitch * height) as u64, usage: wgpu::BufferUsage::MAP_WRITE | wgpu::BufferUsage::COPY_SRC, label: None, mapped_at_creation: true, }); let mut buffer_view = buffer.slice(..).get_mapped_range_mut(); for row in 0..height { buffer_view[row * row_pitch..][..row_bytes] .copy_from_slice(&data[row * row_bytes..][..row_bytes]); } drop(buffer_view); buffer.unmap(); encoder.copy_buffer_to_texture( wgpu::BufferCopyView { buffer: &buffer, layout: wgpu::TextureDataLayout { offset: 0, bytes_per_row: row_pitch as u32, rows_per_image: height as u32 / desc.depth, }, }, wgpu::TextureCopyView { texture: &texture, mip_level: 0, origin: wgpu::Origin3d { x: 0, y: 0, z: 0 }, }, wgpu::Extent3d { width: width as u32, height: height as u32 / desc.depth, depth: desc.depth, }, ); Ok(texture) } pub(crate) fn write_texture( &self, name: &str, desc: TextureDescriptor, data: &[u8], ) -> Result<(), Error> { self.update_texture(name, desc)?; if desc.format == TextureFormat::RGBA8 { let filename = TERRA_DIRECTORY.join(format!("{}.bmp", name)); Ok(image::save_buffer_with_format( &filename, data, desc.width, desc.height * desc.depth, image::ColorType::Rgba8, image::ImageFormat::Bmp, )?) } else { let filename = TERRA_DIRECTORY.join(format!("{}.raw", name)); Ok(fs::write(&filename, data)?) } } pub(crate) fn reload_texture(&self, name: &str) -> bool { let desc = self.lookup_texture(name); if let Ok(Some(desc)) = desc { if desc.format == TextureFormat::RGBA8 { TERRA_DIRECTORY.join(format!("{}.bmp", name)).exists() } else { TERRA_DIRECTORY.join(format!("{}.raw", name)).exists() } } else { false } } pub(crate) fn layers(&self) -> &VecMap<LayerParams> { &self.layers } pub(crate) fn tile_name(layer: LayerType, node: VNode) -> PathBuf { let face = match node.face() { 0 => "0E", 1 => "180E", 2 => "90E", 3 => "90W", 4 => "N", 5 => "S", _ => unreachable!(), }; let (layer, ext) = match layer { LayerType::Displacements => ("displacements", "raw"), LayerType::Albedo => ("albedo", "bmp"), LayerType::Roughness => ("roughness", "raw"), LayerType::Normals => ("normals", "raw"), LayerType::Heightmaps => ("heightmaps", "raw.sz"), }; TERRA_DIRECTORY.join(&format!( "tiles/{}_{}_{}_{}x{}.{}", layer, node.level(), face, node.x(), node.y(), ext )) } pub(crate) fn reload_tile_state( &self, layer: LayerType, node: VNode, base: bool, ) -> Result<TileState, Error> { let filename = Self::tile_name(layer, node); let meta = self.lookup_tile_meta(layer, node); let exists = filename.exists(); let target_state = if base && exists { TileState::Base } else if base { TileState::MissingBase } else if exists { TileState::Generated } else { TileState::Missing }; if let Ok(Some(TileMeta { state, .. })) = meta { if state == target_state { return Ok(state); } } let new_meta = TileMeta { state: target_state, crc32: 0 }; self.update_tile_meta(layer, node, new_meta)?; Ok(target_state) } // pub(crate) fn set_missing( // &self, // layer: LayerType, // node: VNode, // base: bool, // ) -> Result<(), Error> { // let state = if base { TileState::MissingBase } else { TileState::Missing }; // self.update_tile_meta(layer, node, TileMeta { crc32: 0, state }) // } pub(crate) fn clear_generated(&mut self, layer: LayerType) -> Result<(), Error> { self.scan_tile_meta(layer, |node, meta| { if let TileState::Generated = meta.state { self.remove_tile_meta(layer, node)?; } Ok(()) }) } pub(crate) fn get_missing_base(&self, layer: LayerType) -> Result<Vec<VNode>, Error> { let mut missing = Vec::new(); self.scan_tile_meta(layer, |node, meta| { if let TileState::MissingBase = meta.state { missing.push(node); } Ok(()) })?; Ok(missing) } // // These functions use the database. // fn lookup_tile_meta(&self, layer: LayerType, node: VNode) -> Result<Option<TileMeta>, Error> { let key = bincode::serialize(&(layer, node)).unwrap(); Ok(self.tiles.get(key)?.map(|value| bincode::deserialize(&value).unwrap())) } fn update_tile_meta(&self, layer: LayerType, node: VNode, meta: TileMeta) -> Result<(), Error> { let key = bincode::serialize(&(layer, node)).unwrap(); let value = bincode::serialize(&meta).unwrap(); self.tiles.insert(key, value)?; Ok(()) } fn remove_tile_meta(&self, layer: LayerType, node: VNode) -> Result<(), Error> { let key = bincode::serialize(&(layer, node)).unwrap(); self.tiles.remove(key)?; Ok(()) } fn scan_tile_meta<F: FnMut(VNode, TileMeta) -> Result<(), Error>>( &self, layer: LayerType,
mut f: F, ) -> Result<(), Error> { let prefix = bincode::serialize(&layer).unwrap();
random_line_split
run.py
_index, source_pop in enumerate(par.label): n_target = num_neurons[target_index] n_source = num_neurons[source_index] K[target_index][source_index] = np.log(1. - par.conn_probs[target_index][source_index]) / np.log( 1. - 1. / (n_target * n_source))/n_target return K def get_neuron_number(self): '''stores the neuron numbers in list ordered such as label''' num_neurons = [] layers = ['L23','L4','L5','L6'] keys = ['E', 'I'] for layer in layers: for key in keys: num_neurons.append(par.num_neurons[layer][key]) return num_neurons def build(self): # set populations self.populations = {} # calculate indegrees from connection probability self.indegrees = self.get_indegrees() for layer, exIn in par.num_neurons.items(): # [:1] to remove the first "L" self.populations[layer[1:] + "e"] = pynn.Population( int(exIn["E"] * self.scale), self.model) self.populations[layer[1:] + "i"] = pynn.Population( int(exIn["I"] * self.scale), self.model) # Create projections self.projections = [] self.projectionLabels = [] for targetIndex, targetPop in enumerate(par.label): for sourceIndex, sourcePop in enumerate(par.label): if sourcePop.endswith("e"): target = "excitatory" else: target = "inhibitory" sourceSize = self.populations[sourcePop].size targetSize = self.populations[targetPop].size # In-degree scaling as described in Albada et al. (2015) "Scalability of Asynchronous Networks # Is Limited by One-to-One Mapping between Effective Connectivity and Correlations" # Number of inputs per target neuron (in-degree) for full scale model is scaled with k_scale # To receive total connection number it is multiplied with downscaled target population size (scale) # Connection probability is not preserved if scale == k_scale (multiple connections neglected) n_connection = int(round(self.indegrees[targetIndex][sourceIndex] * self.k_scale * targetSize)) self.totalConnections += n_connection if(n_connection == 0): continue # connection matrix [(neuron_pop1,neuron_pop2,weight,delay),(...)] matrix = np.zeros((4, n_connection),dtype= float) np.random.seed(self.seed) matrix[0] = np.random.randint(0, sourceSize, n_connection) matrix[1] = np.random.randint(0, targetSize, n_connection) # The delay and weight is not important for mapping # PyNN requires it to be set to some value matrix[2] = np.repeat(1, n_connection) # arbitrary weight matrix[3] = np.repeat(0, n_connection) # arbitrary delay matrix = matrix.T matrix = [[int(a),int(b),c,d] for a,b,c,d in matrix] connector = pynn.FromListConnector(matrix) self.projections.append(pynn.Projection( self.populations[sourcePop], self.populations[targetPop], connector, target=target, label=sourcePop + "-" + targetPop)) self.projectionLabels.append(sourcePop + "-" + targetPop) print("total connections:", self.totalConnections) # external input: self.externalInputPops = {} # External spikes or external current external_source = par.external_source # will not work for large networks, for now it is not used due to par.external_source if (external_source == "spikeInput"): print("using external input connections") for layer, amount in par.K_ext.items(): # rate is given in model with 8Hz # will not work for large networks, for now it is not used due to par.external_source rate_to_ex = par.bg_rate * amount["E"] * self.k_scale rate_to_in = par.bg_rate * amount["I"] * self.k_scale self.externalInputPops[layer[1:] + "e"] = pynn.Population( self.populations[layer[1:] + "e"].size, pynn.SpikeSourcePoisson, {'rate': rate_to_ex}) self.externalInputPops[layer[1:] + "i"] = pynn.Population( self.populations[layer[1:] + "i"].size, pynn.SpikeSourcePoisson, {'rate': rate_to_in}) # create connections for sourceKey, sourcePop in self.externalInputPops.items(): # set connector for each pop size since RandomDistribution object not supported by pyhmf # arbitrary weight externalConnector = pynn.OneToOneConnector( weights = 1) # create connection self.projections.append(pynn.Projection( sourcePop, self.populations[sourceKey], externalConnector, target="excitatory")) self.projectionLabels.append("ext.-" + targetPop) def getLoss(self, marocco): perPopulation = {} for i in range(len(self.projections)): synLoss, totalSyn = self.projectionwise_synapse_loss( self.projections[i], marocco) perPopulation[self.projectionLabels[i]] = { "synLoss": synLoss, "TotalSyns": totalSyn} return perPopulation def run(self):
def projectionwise_synapse_loss(self, proj, marocco): """ computes the synapse loss of a projection params: proj - a pyhmf.Projection marocco - the PyMarocco object after the mapping has run. returns: (nr of lost synapses, total synapses in projection) """ orig_weights = proj.getWeights(format='array') mapped_weights = marocco.stats.getWeights(proj) syns = np.where(~np.isnan(orig_weights)) realized_syns = np.where(~np.isnan(mapped_weights)) orig = len(syns[0]) realized = len(realized_syns[0]) if orig > 0: print ("Projection-Wise Synapse Loss", proj, (orig - realized) * 100. / orig) return orig - realized, orig def str2bool(v): if isinstance(v, bool): return v if v.lower() in ('yes', 'true', 't', 'y', '1'): return True elif v.lower() in ('no', 'false', 'f', 'n', '0'): return False else: raise argparse.ArgumentTypeError('Boolean value expected.') def main(): parser = argparse.ArgumentParser() # scale factor of the whole network compared to the original one parser.add_argument('--scale', default=0.01, type=float) # size of one neueron in hw neurons parser.add_argument('--n_size', default=4, type=int) parser.add_argument('--k_scale', type=float) # scale of connections # wafer defects that should be considered in the mapping parser.add_argument('--wafer', '-w', type=int, default=24) # specific path where the defect parts of the wafer are saved # if nothing specified, current defects of the given wafer are used parser.add_argument('--defects_path', type=str) parser.add_argument('--ignore_blacklisting', type=str2bool, nargs='?', default = False, const=True) parser.add_argument('--name', type=str, default='cortical_column_network') # name parser.add_argument('--placer', type=str, default='byNeuron') parser.add_argument('--seed', default=0, type=int) args = parser.parse_args() # k_scale is set to "scale" by deflaut if not args.k_scale: args.k_scale = args.scale taskname = "scale{}_k-scale{}_nsize{}_wafer{}_ignoreBlacklsiting{}".format( args.scale, args.k_scale, args.n_size, args.wafer, args.ignore_blacklisting) marocco = PyMarocco() marocco.neuron_placement.default_neuron_size(args.n_size) if(args.ignore_blacklisting): marocco.defects.backend = Defects.Backend.Without else: marocco.defects.backend = Defects.Backend.XML marocco.skip_mapping = False marocco.backend = PyMarocco.Without marocco.continue_despite_synapse_loss = True marocco.default_wafer = C.Wafer(args.wafer) # give wafer args marocco.calib_backend = PyMarocco.CalibBackend.Default marocco.calib_path = "/wang/data/calibration/brainscales/default" if args.defects_path: marocco.defects.path = args.defects_path else: marocco.defects.path = "/wang/data/commissioning/BSS-1/rackplace/" + str( args.wafer) + "/derived_plus_calib_blacklisting/current" # c 4189 no specification #taskname += "_c4189_" # strategy mar
pynn.run(1) pynn.end()
identifier_body
run.py
_index, source_pop in enumerate(par.label): n_target = num_neurons[target_index] n_source = num_neurons[source_index] K[target_index][source_index] = np.log(1. - par.conn_probs[target_index][source_index]) / np.log( 1. - 1. / (n_target * n_source))/n_target return K def get_neuron_number(self): '''stores the neuron numbers in list ordered such as label''' num_neurons = [] layers = ['L23','L4','L5','L6'] keys = ['E', 'I'] for layer in layers: for key in keys: num_neurons.append(par.num_neurons[layer][key]) return num_neurons def build(self): # set populations self.populations = {} # calculate indegrees from connection probability self.indegrees = self.get_indegrees() for layer, exIn in par.num_neurons.items(): # [:1] to remove the first "L" self.populations[layer[1:] + "e"] = pynn.Population( int(exIn["E"] * self.scale), self.model) self.populations[layer[1:] + "i"] = pynn.Population( int(exIn["I"] * self.scale), self.model) # Create projections self.projections = [] self.projectionLabels = [] for targetIndex, targetPop in enumerate(par.label): for sourceIndex, sourcePop in enumerate(par.label): if sourcePop.endswith("e"): target = "excitatory" else:
sourceSize = self.populations[sourcePop].size targetSize = self.populations[targetPop].size # In-degree scaling as described in Albada et al. (2015) "Scalability of Asynchronous Networks # Is Limited by One-to-One Mapping between Effective Connectivity and Correlations" # Number of inputs per target neuron (in-degree) for full scale model is scaled with k_scale # To receive total connection number it is multiplied with downscaled target population size (scale) # Connection probability is not preserved if scale == k_scale (multiple connections neglected) n_connection = int(round(self.indegrees[targetIndex][sourceIndex] * self.k_scale * targetSize)) self.totalConnections += n_connection if(n_connection == 0): continue # connection matrix [(neuron_pop1,neuron_pop2,weight,delay),(...)] matrix = np.zeros((4, n_connection),dtype= float) np.random.seed(self.seed) matrix[0] = np.random.randint(0, sourceSize, n_connection) matrix[1] = np.random.randint(0, targetSize, n_connection) # The delay and weight is not important for mapping # PyNN requires it to be set to some value matrix[2] = np.repeat(1, n_connection) # arbitrary weight matrix[3] = np.repeat(0, n_connection) # arbitrary delay matrix = matrix.T matrix = [[int(a),int(b),c,d] for a,b,c,d in matrix] connector = pynn.FromListConnector(matrix) self.projections.append(pynn.Projection( self.populations[sourcePop], self.populations[targetPop], connector, target=target, label=sourcePop + "-" + targetPop)) self.projectionLabels.append(sourcePop + "-" + targetPop) print("total connections:", self.totalConnections) # external input: self.externalInputPops = {} # External spikes or external current external_source = par.external_source # will not work for large networks, for now it is not used due to par.external_source if (external_source == "spikeInput"): print("using external input connections") for layer, amount in par.K_ext.items(): # rate is given in model with 8Hz # will not work for large networks, for now it is not used due to par.external_source rate_to_ex = par.bg_rate * amount["E"] * self.k_scale rate_to_in = par.bg_rate * amount["I"] * self.k_scale self.externalInputPops[layer[1:] + "e"] = pynn.Population( self.populations[layer[1:] + "e"].size, pynn.SpikeSourcePoisson, {'rate': rate_to_ex}) self.externalInputPops[layer[1:] + "i"] = pynn.Population( self.populations[layer[1:] + "i"].size, pynn.SpikeSourcePoisson, {'rate': rate_to_in}) # create connections for sourceKey, sourcePop in self.externalInputPops.items(): # set connector for each pop size since RandomDistribution object not supported by pyhmf # arbitrary weight externalConnector = pynn.OneToOneConnector( weights = 1) # create connection self.projections.append(pynn.Projection( sourcePop, self.populations[sourceKey], externalConnector, target="excitatory")) self.projectionLabels.append("ext.-" + targetPop) def getLoss(self, marocco): perPopulation = {} for i in range(len(self.projections)): synLoss, totalSyn = self.projectionwise_synapse_loss( self.projections[i], marocco) perPopulation[self.projectionLabels[i]] = { "synLoss": synLoss, "TotalSyns": totalSyn} return perPopulation def run(self): pynn.run(1) pynn.end() def projectionwise_synapse_loss(self, proj, marocco): """ computes the synapse loss of a projection params: proj - a pyhmf.Projection marocco - the PyMarocco object after the mapping has run. returns: (nr of lost synapses, total synapses in projection) """ orig_weights = proj.getWeights(format='array') mapped_weights = marocco.stats.getWeights(proj) syns = np.where(~np.isnan(orig_weights)) realized_syns = np.where(~np.isnan(mapped_weights)) orig = len(syns[0]) realized = len(realized_syns[0]) if orig > 0: print ("Projection-Wise Synapse Loss", proj, (orig - realized) * 100. / orig) return orig - realized, orig def str2bool(v): if isinstance(v, bool): return v if v.lower() in ('yes', 'true', 't', 'y', '1'): return True elif v.lower() in ('no', 'false', 'f', 'n', '0'): return False else: raise argparse.ArgumentTypeError('Boolean value expected.') def main(): parser = argparse.ArgumentParser() # scale factor of the whole network compared to the original one parser.add_argument('--scale', default=0.01, type=float) # size of one neueron in hw neurons parser.add_argument('--n_size', default=4, type=int) parser.add_argument('--k_scale', type=float) # scale of connections # wafer defects that should be considered in the mapping parser.add_argument('--wafer', '-w', type=int, default=24) # specific path where the defect parts of the wafer are saved # if nothing specified, current defects of the given wafer are used parser.add_argument('--defects_path', type=str) parser.add_argument('--ignore_blacklisting', type=str2bool, nargs='?', default = False, const=True) parser.add_argument('--name', type=str, default='cortical_column_network') # name parser.add_argument('--placer', type=str, default='byNeuron') parser.add_argument('--seed', default=0, type=int) args = parser.parse_args() # k_scale is set to "scale" by deflaut if not args.k_scale: args.k_scale = args.scale taskname = "scale{}_k-scale{}_nsize{}_wafer{}_ignoreBlacklsiting{}".format( args.scale, args.k_scale, args.n_size, args.wafer, args.ignore_blacklisting) marocco = PyMarocco() marocco.neuron_placement.default_neuron_size(args.n_size) if(args.ignore_blacklisting): marocco.defects.backend = Defects.Backend.Without else: marocco.defects.backend = Defects.Backend.XML marocco.skip_mapping = False marocco.backend = PyMarocco.Without marocco.continue_despite_synapse_loss = True marocco.default_wafer = C.Wafer(args.wafer) # give wafer args marocco.calib_backend = PyMarocco.CalibBackend.Default marocco.calib_path = "/wang/data/calibration/brainscales/default" if args.defects_path: marocco.defects.path = args.defects_path else: marocco.defects.path = "/wang/data/commissioning/BSS-1/rackplace/" + str( args.wafer) + "/derived_plus_calib_blacklisting/current" # c 4189 no specification #taskname += "_c4189_" # strategy mar
target = "inhibitory"
conditional_block
run.py
source_index, source_pop in enumerate(par.label): n_target = num_neurons[target_index] n_source = num_neurons[source_index] K[target_index][source_index] = np.log(1. - par.conn_probs[target_index][source_index]) / np.log( 1. - 1. / (n_target * n_source))/n_target return K def get_neuron_number(self): '''stores the neuron numbers in list ordered such as label''' num_neurons = [] layers = ['L23','L4','L5','L6'] keys = ['E', 'I'] for layer in layers: for key in keys: num_neurons.append(par.num_neurons[layer][key]) return num_neurons def build(self): # set populations self.populations = {} # calculate indegrees from connection probability self.indegrees = self.get_indegrees() for layer, exIn in par.num_neurons.items(): # [:1] to remove the first "L" self.populations[layer[1:] + "e"] = pynn.Population( int(exIn["E"] * self.scale), self.model) self.populations[layer[1:] + "i"] = pynn.Population( int(exIn["I"] * self.scale), self.model) # Create projections self.projections = [] self.projectionLabels = [] for targetIndex, targetPop in enumerate(par.label): for sourceIndex, sourcePop in enumerate(par.label): if sourcePop.endswith("e"): target = "excitatory" else: target = "inhibitory" sourceSize = self.populations[sourcePop].size targetSize = self.populations[targetPop].size # In-degree scaling as described in Albada et al. (2015) "Scalability of Asynchronous Networks # Is Limited by One-to-One Mapping between Effective Connectivity and Correlations" # Number of inputs per target neuron (in-degree) for full scale model is scaled with k_scale # To receive total connection number it is multiplied with downscaled target population size (scale) # Connection probability is not preserved if scale == k_scale (multiple connections neglected) n_connection = int(round(self.indegrees[targetIndex][sourceIndex] * self.k_scale * targetSize)) self.totalConnections += n_connection if(n_connection == 0): continue # connection matrix [(neuron_pop1,neuron_pop2,weight,delay),(...)] matrix = np.zeros((4, n_connection),dtype= float) np.random.seed(self.seed) matrix[0] = np.random.randint(0, sourceSize, n_connection) matrix[1] = np.random.randint(0, targetSize, n_connection) # The delay and weight is not important for mapping # PyNN requires it to be set to some value matrix[2] = np.repeat(1, n_connection) # arbitrary weight matrix[3] = np.repeat(0, n_connection) # arbitrary delay matrix = matrix.T matrix = [[int(a),int(b),c,d] for a,b,c,d in matrix] connector = pynn.FromListConnector(matrix) self.projections.append(pynn.Projection( self.populations[sourcePop], self.populations[targetPop], connector, target=target, label=sourcePop + "-" + targetPop)) self.projectionLabels.append(sourcePop + "-" + targetPop) print("total connections:", self.totalConnections) # external input: self.externalInputPops = {} # External spikes or external current external_source = par.external_source # will not work for large networks, for now it is not used due to par.external_source if (external_source == "spikeInput"): print("using external input connections")
self.externalInputPops[layer[1:] + "e"] = pynn.Population( self.populations[layer[1:] + "e"].size, pynn.SpikeSourcePoisson, {'rate': rate_to_ex}) self.externalInputPops[layer[1:] + "i"] = pynn.Population( self.populations[layer[1:] + "i"].size, pynn.SpikeSourcePoisson, {'rate': rate_to_in}) # create connections for sourceKey, sourcePop in self.externalInputPops.items(): # set connector for each pop size since RandomDistribution object not supported by pyhmf # arbitrary weight externalConnector = pynn.OneToOneConnector( weights = 1) # create connection self.projections.append(pynn.Projection( sourcePop, self.populations[sourceKey], externalConnector, target="excitatory")) self.projectionLabels.append("ext.-" + targetPop) def getLoss(self, marocco): perPopulation = {} for i in range(len(self.projections)): synLoss, totalSyn = self.projectionwise_synapse_loss( self.projections[i], marocco) perPopulation[self.projectionLabels[i]] = { "synLoss": synLoss, "TotalSyns": totalSyn} return perPopulation def run(self): pynn.run(1) pynn.end() def projectionwise_synapse_loss(self, proj, marocco): """ computes the synapse loss of a projection params: proj - a pyhmf.Projection marocco - the PyMarocco object after the mapping has run. returns: (nr of lost synapses, total synapses in projection) """ orig_weights = proj.getWeights(format='array') mapped_weights = marocco.stats.getWeights(proj) syns = np.where(~np.isnan(orig_weights)) realized_syns = np.where(~np.isnan(mapped_weights)) orig = len(syns[0]) realized = len(realized_syns[0]) if orig > 0: print ("Projection-Wise Synapse Loss", proj, (orig - realized) * 100. / orig) return orig - realized, orig def str2bool(v): if isinstance(v, bool): return v if v.lower() in ('yes', 'true', 't', 'y', '1'): return True elif v.lower() in ('no', 'false', 'f', 'n', '0'): return False else: raise argparse.ArgumentTypeError('Boolean value expected.') def main(): parser = argparse.ArgumentParser() # scale factor of the whole network compared to the original one parser.add_argument('--scale', default=0.01, type=float) # size of one neueron in hw neurons parser.add_argument('--n_size', default=4, type=int) parser.add_argument('--k_scale', type=float) # scale of connections # wafer defects that should be considered in the mapping parser.add_argument('--wafer', '-w', type=int, default=24) # specific path where the defect parts of the wafer are saved # if nothing specified, current defects of the given wafer are used parser.add_argument('--defects_path', type=str) parser.add_argument('--ignore_blacklisting', type=str2bool, nargs='?', default = False, const=True) parser.add_argument('--name', type=str, default='cortical_column_network') # name parser.add_argument('--placer', type=str, default='byNeuron') parser.add_argument('--seed', default=0, type=int) args = parser.parse_args() # k_scale is set to "scale" by deflaut if not args.k_scale: args.k_scale = args.scale taskname = "scale{}_k-scale{}_nsize{}_wafer{}_ignoreBlacklsiting{}".format( args.scale, args.k_scale, args.n_size, args.wafer, args.ignore_blacklisting) marocco = PyMarocco() marocco.neuron_placement.default_neuron_size(args.n_size) if(args.ignore_blacklisting): marocco.defects.backend = Defects.Backend.Without else: marocco.defects.backend = Defects.Backend.XML marocco.skip_mapping = False marocco.backend = PyMarocco.Without marocco.continue_despite_synapse_loss = True marocco.default_wafer = C.Wafer(args.wafer) # give wafer args marocco.calib_backend = PyMarocco.CalibBackend.Default marocco.calib_path = "/wang/data/calibration/brainscales/default" if args.defects_path: marocco.defects.path = args.defects_path else: marocco.defects.path = "/wang/data/commissioning/BSS-1/rackplace/" + str( args.wafer) + "/derived_plus_calib_blacklisting/current" # c 4189 no specification #taskname += "_c4189_" # strategy marocco
for layer, amount in par.K_ext.items(): # rate is given in model with 8Hz # will not work for large networks, for now it is not used due to par.external_source rate_to_ex = par.bg_rate * amount["E"] * self.k_scale rate_to_in = par.bg_rate * amount["I"] * self.k_scale
random_line_split
run.py
_index, source_pop in enumerate(par.label): n_target = num_neurons[target_index] n_source = num_neurons[source_index] K[target_index][source_index] = np.log(1. - par.conn_probs[target_index][source_index]) / np.log( 1. - 1. / (n_target * n_source))/n_target return K def get_neuron_number(self): '''stores the neuron numbers in list ordered such as label''' num_neurons = [] layers = ['L23','L4','L5','L6'] keys = ['E', 'I'] for layer in layers: for key in keys: num_neurons.append(par.num_neurons[layer][key]) return num_neurons def build(self): # set populations self.populations = {} # calculate indegrees from connection probability self.indegrees = self.get_indegrees() for layer, exIn in par.num_neurons.items(): # [:1] to remove the first "L" self.populations[layer[1:] + "e"] = pynn.Population( int(exIn["E"] * self.scale), self.model) self.populations[layer[1:] + "i"] = pynn.Population( int(exIn["I"] * self.scale), self.model) # Create projections self.projections = [] self.projectionLabels = [] for targetIndex, targetPop in enumerate(par.label): for sourceIndex, sourcePop in enumerate(par.label): if sourcePop.endswith("e"): target = "excitatory" else: target = "inhibitory" sourceSize = self.populations[sourcePop].size targetSize = self.populations[targetPop].size # In-degree scaling as described in Albada et al. (2015) "Scalability of Asynchronous Networks # Is Limited by One-to-One Mapping between Effective Connectivity and Correlations" # Number of inputs per target neuron (in-degree) for full scale model is scaled with k_scale # To receive total connection number it is multiplied with downscaled target population size (scale) # Connection probability is not preserved if scale == k_scale (multiple connections neglected) n_connection = int(round(self.indegrees[targetIndex][sourceIndex] * self.k_scale * targetSize)) self.totalConnections += n_connection if(n_connection == 0): continue # connection matrix [(neuron_pop1,neuron_pop2,weight,delay),(...)] matrix = np.zeros((4, n_connection),dtype= float) np.random.seed(self.seed) matrix[0] = np.random.randint(0, sourceSize, n_connection) matrix[1] = np.random.randint(0, targetSize, n_connection) # The delay and weight is not important for mapping # PyNN requires it to be set to some value matrix[2] = np.repeat(1, n_connection) # arbitrary weight matrix[3] = np.repeat(0, n_connection) # arbitrary delay matrix = matrix.T matrix = [[int(a),int(b),c,d] for a,b,c,d in matrix] connector = pynn.FromListConnector(matrix) self.projections.append(pynn.Projection( self.populations[sourcePop], self.populations[targetPop], connector, target=target, label=sourcePop + "-" + targetPop)) self.projectionLabels.append(sourcePop + "-" + targetPop) print("total connections:", self.totalConnections) # external input: self.externalInputPops = {} # External spikes or external current external_source = par.external_source # will not work for large networks, for now it is not used due to par.external_source if (external_source == "spikeInput"): print("using external input connections") for layer, amount in par.K_ext.items(): # rate is given in model with 8Hz # will not work for large networks, for now it is not used due to par.external_source rate_to_ex = par.bg_rate * amount["E"] * self.k_scale rate_to_in = par.bg_rate * amount["I"] * self.k_scale self.externalInputPops[layer[1:] + "e"] = pynn.Population( self.populations[layer[1:] + "e"].size, pynn.SpikeSourcePoisson, {'rate': rate_to_ex}) self.externalInputPops[layer[1:] + "i"] = pynn.Population( self.populations[layer[1:] + "i"].size, pynn.SpikeSourcePoisson, {'rate': rate_to_in}) # create connections for sourceKey, sourcePop in self.externalInputPops.items(): # set connector for each pop size since RandomDistribution object not supported by pyhmf # arbitrary weight externalConnector = pynn.OneToOneConnector( weights = 1) # create connection self.projections.append(pynn.Projection( sourcePop, self.populations[sourceKey], externalConnector, target="excitatory")) self.projectionLabels.append("ext.-" + targetPop) def getLoss(self, marocco): perPopulation = {} for i in range(len(self.projections)): synLoss, totalSyn = self.projectionwise_synapse_loss( self.projections[i], marocco) perPopulation[self.projectionLabels[i]] = { "synLoss": synLoss, "TotalSyns": totalSyn} return perPopulation def run(self): pynn.run(1) pynn.end() def projectionwise_synapse_loss(self, proj, marocco): """ computes the synapse loss of a projection params: proj - a pyhmf.Projection marocco - the PyMarocco object after the mapping has run. returns: (nr of lost synapses, total synapses in projection) """ orig_weights = proj.getWeights(format='array') mapped_weights = marocco.stats.getWeights(proj) syns = np.where(~np.isnan(orig_weights)) realized_syns = np.where(~np.isnan(mapped_weights)) orig = len(syns[0]) realized = len(realized_syns[0]) if orig > 0: print ("Projection-Wise Synapse Loss", proj, (orig - realized) * 100. / orig) return orig - realized, orig def str2bool(v): if isinstance(v, bool): return v if v.lower() in ('yes', 'true', 't', 'y', '1'): return True elif v.lower() in ('no', 'false', 'f', 'n', '0'): return False else: raise argparse.ArgumentTypeError('Boolean value expected.') def
(): parser = argparse.ArgumentParser() # scale factor of the whole network compared to the original one parser.add_argument('--scale', default=0.01, type=float) # size of one neueron in hw neurons parser.add_argument('--n_size', default=4, type=int) parser.add_argument('--k_scale', type=float) # scale of connections # wafer defects that should be considered in the mapping parser.add_argument('--wafer', '-w', type=int, default=24) # specific path where the defect parts of the wafer are saved # if nothing specified, current defects of the given wafer are used parser.add_argument('--defects_path', type=str) parser.add_argument('--ignore_blacklisting', type=str2bool, nargs='?', default = False, const=True) parser.add_argument('--name', type=str, default='cortical_column_network') # name parser.add_argument('--placer', type=str, default='byNeuron') parser.add_argument('--seed', default=0, type=int) args = parser.parse_args() # k_scale is set to "scale" by deflaut if not args.k_scale: args.k_scale = args.scale taskname = "scale{}_k-scale{}_nsize{}_wafer{}_ignoreBlacklsiting{}".format( args.scale, args.k_scale, args.n_size, args.wafer, args.ignore_blacklisting) marocco = PyMarocco() marocco.neuron_placement.default_neuron_size(args.n_size) if(args.ignore_blacklisting): marocco.defects.backend = Defects.Backend.Without else: marocco.defects.backend = Defects.Backend.XML marocco.skip_mapping = False marocco.backend = PyMarocco.Without marocco.continue_despite_synapse_loss = True marocco.default_wafer = C.Wafer(args.wafer) # give wafer args marocco.calib_backend = PyMarocco.CalibBackend.Default marocco.calib_path = "/wang/data/calibration/brainscales/default" if args.defects_path: marocco.defects.path = args.defects_path else: marocco.defects.path = "/wang/data/commissioning/BSS-1/rackplace/" + str( args.wafer) + "/derived_plus_calib_blacklisting/current" # c 4189 no specification #taskname += "_c4189_" # strategy mar
main
identifier_name
a1_300068688.py.py
n' times print(str(yoda)) #prints 'yoda' ######################## # Question 2 ######################## ''' creates a function called is_prime that takes one argument and check to see whether it is prime or not, if it is a prime the function return true if not it returns else''' def is_prime(n): if n >= 2: # prime numbers are numbers that cant be divided unless its by 1 or it self so 'n' cannot be 1 or 2 for i in range(2, n): #creates a for loop that starts at 2 and ends at 'n' if (n%i == 0): #checks to see whether n divides into i intergerally return n<0 #if it does return false becuase n is not a prime number then else: return n>0 #returns true because n is a prime ######################## # Question 3 ######################## ''' Creates a function points that takes 4 arguments which make up to coordinates x1,y1 x2,y2 this function then computes the slope of the the line of these points and the distance between them and outputs a message a message''' def points(x1,y1,x2,y2): distance = (((x2-x1)**2)+((y2-y1)**2))**(1/2) # finds the distance between these points by finding the hypotenuse of the triangle a^2 + b^2 = c^2 if (x2-x1)!= 0 and (y2-y1)!=0: #checks to see if the slope is not 0 and if it is print a special message slope = (y2-y1)/(x2-x1) #computes slope print("The slope is " + str(slope) + " and the distance is " + str(distance)) #prints out slope and distance else: print("the slope is infinity and the distance is " + str(distance))# slope is infinity so it outprints just the distance ######################## # Question 5 ######################## ''' creates a function reverse_int that takes one argument that is a integer and reverses this integer''' def reverse_int(num): reverse1 = num%10 # gets you the number of the begining of the reversed integer reverse2 = (num%100)-reverse1 # gets you the number of the second number in the reversed inetger reverse1 = reverse1*100 # makes the number at the end now the begining reverse3 = num//100 #gives you last number in the reversed integer final = reverse1 + reverse2 + reverse3 # adds all the numbers together to get you the reversed int return final #function returns the reversed int ######################## # Question 6 ######################## ''' creates a function vowelCount that takes one argument which is a string and then prints out the number of vowels and what vowels are in the word''' def vowelCount(string): print('a, e, i, o, and u appear, respectively, ' + str(string.count('a'))+ ', ' + str(string.count('e'))+ ', ' + str(string.count('i')) + ', ' + str(string.count('o')) + ', ' + str(string.count('u'))+ ' times') ######################## # Question 7 ######################## ''' creayes a function with 3 input paramteres that are strings and checks to see if all the same then returns true if they are and false if they are not''' def allthesame(x,y,z): if x==y and x==z and z==y: #if statment that checks whther they are all the sale return x==y else: return 1<0 ''' creates a function alldifferent that has 3 input paramteres that are all strings and checks whether all the strings are different and returns true if they are and false if they are not''' def alldifferent(x,y,z): if x != y and y != z and x != z: #checks to see whether the strings x, y and z are all different return x!=y else: return 1<0 ''' creates a function called sorted that takes 3 input paramteres and checks to see whether they are in sorted order and returns true if they are and false if they are not''' def sorted(x,y,z): if z>y and z>x and y>x: #checks whther they are sorted return z>y else: return 1<0 ######################## # Question 8 ######################## ''' creates a function called leap that takes on input value that is an integer , this function then computes whether that number was a leap year and returns true if it was and false if it isnt''' def leap(year): if year%4==0 and year%100 != 0 or year%400==0: #checks whether the year was a leap year return year>0 else: return year<0 ######################## # Question 9 ######################## ''' creates a function called letter2number tthat takes a string input that is a letter grade and converts them to a number grade''' def letter2number(letter): if letter== 'A' or'A-' or 'A+': #checks to see if it was an A, A+ or A- if letter == 'A': #assigns the corresponding number value to the letter and prints it print(str(4)) elif letter == 'A+': print(str(4+0.3)) elif letter== 'A-': print(str(4-0.3)) if letter == 'B' or 'B-' or 'B+': #checks to see if it was an B, B+ or B- if letter == 'B': #assigns the corresponding number value to the letter and prints it print(str(3.0)) elif letter == 'B+': print(str(3+0.3)) elif letter == 'B-': print(str(3-0.3)) if letter == 'C' or 'C-' or 'C+': #checks to see if it was an C, C+ or C- if letter== 'C': #assigns the corresponding number value to the letter and prints it print(str(2.0)) elif letter== 'C+': print(str(2+0.3)) elif letter== 'C-': print(str(2-0.3)) if letter == 'D' or 'D-' or 'D+': #checks to see if it was an D, D+ or D- if letter == 'D': #assigns the corresponding number value to the letter and prints it print(str(1)) elif letter=='D+': print(str(1+0.3)) elif letter=='D-': print(str(1-0.3)) if letter == 'F': print(str(0)) ######################## # Question 10 ######################## ''' creates a function called is_Palindrome that takes one input that is a string and determines if that string is the exact same when reversed and returns true if it is and false if it is not ''' def is_Palindrome(string): reverse = string[::-1] #reverses the string if reverse == string:
else: return reverse == string ######################## # Question 12 ######################## ''' creates a function called rps that is a game of rock, paper, scissors and takes two input paramteres that are strings that are either 'S' 'P' 'R' and returns a -1 if player 1 wins and a 1 if player two wins and a zero if its a tie''' def rps(player1, player2): if player1 == 'R' and player2 == 'S' or player1 == 'P' and player2 == 'R' or player1 == 'S' and player2 == 'P': return -1 elif player2 == 'R' and player1 == 'S' or player2 == 'P' and player1 == 'R' or player2 == 'S' and player1 == 'P': return 1 else: return 0 ######################## # Question 14 ######################## ''' creates a function called count_even_digits that takes two arguments and returns the amount of even numbers in the first paramater and the second paramtere is the amount of digits in the first one ''' def count_even_digits(int1, int2): even = str(int1) even = int(even.count('0')) + int(even.count('2')) + int(even.count('4')) + int(even.count('6')) + int(even.count('8')) #uses the count operator to see the amount of even numbers print(str(even)) ######################## # Question 13 ######################## ''' creates a function called alogical that takes one parameter n which must be a integer and counts how many times that integer must be divided by 2 to be less than or equal to 1 ''' def alogical(n): counter = 0 while n > 1 or n==1: # while loop is used to count the amount of times needed to divide by 2 n = n/2 counter = counter + 1 print(str(counter))
return reverse == string
conditional_block
a1_300068688.py.py
n' times print(str(yoda)) #prints 'yoda' ######################## # Question 2 ######################## ''' creates a function called is_prime that takes one argument and check to see whether it is prime or not, if it is a prime the function return true if not it returns else''' def is_prime(n): if n >= 2: # prime numbers are numbers that cant be divided unless its by 1 or it self so 'n' cannot be 1 or 2 for i in range(2, n): #creates a for loop that starts at 2 and ends at 'n' if (n%i == 0): #checks to see whether n divides into i intergerally return n<0 #if it does return false becuase n is not a prime number then else: return n>0 #returns true because n is a prime ######################## # Question 3 ######################## ''' Creates a function points that takes 4 arguments which make up to coordinates x1,y1 x2,y2 this function then computes the slope of the the line of these points and the distance between them and outputs a message a message''' def points(x1,y1,x2,y2): distance = (((x2-x1)**2)+((y2-y1)**2))**(1/2) # finds the distance between these points by finding the hypotenuse of the triangle a^2 + b^2 = c^2 if (x2-x1)!= 0 and (y2-y1)!=0: #checks to see if the slope is not 0 and if it is print a special message slope = (y2-y1)/(x2-x1) #computes slope print("The slope is " + str(slope) + " and the distance is " + str(distance)) #prints out slope and distance else: print("the slope is infinity and the distance is " + str(distance))# slope is infinity so it outprints just the distance ######################## # Question 5 ######################## ''' creates a function reverse_int that takes one argument that is a integer and reverses this integer''' def reverse_int(num):
######################## # Question 6 ######################## ''' creates a function vowelCount that takes one argument which is a string and then prints out the number of vowels and what vowels are in the word''' def vowelCount(string): print('a, e, i, o, and u appear, respectively, ' + str(string.count('a'))+ ', ' + str(string.count('e'))+ ', ' + str(string.count('i')) + ', ' + str(string.count('o')) + ', ' + str(string.count('u'))+ ' times') ######################## # Question 7 ######################## ''' creayes a function with 3 input paramteres that are strings and checks to see if all the same then returns true if they are and false if they are not''' def allthesame(x,y,z): if x==y and x==z and z==y: #if statment that checks whther they are all the sale return x==y else: return 1<0 ''' creates a function alldifferent that has 3 input paramteres that are all strings and checks whether all the strings are different and returns true if they are and false if they are not''' def alldifferent(x,y,z): if x != y and y != z and x != z: #checks to see whether the strings x, y and z are all different return x!=y else: return 1<0 ''' creates a function called sorted that takes 3 input paramteres and checks to see whether they are in sorted order and returns true if they are and false if they are not''' def sorted(x,y,z): if z>y and z>x and y>x: #checks whther they are sorted return z>y else: return 1<0 ######################## # Question 8 ######################## ''' creates a function called leap that takes on input value that is an integer , this function then computes whether that number was a leap year and returns true if it was and false if it isnt''' def leap(year): if year%4==0 and year%100 != 0 or year%400==0: #checks whether the year was a leap year return year>0 else: return year<0 ######################## # Question 9 ######################## ''' creates a function called letter2number tthat takes a string input that is a letter grade and converts them to a number grade''' def letter2number(letter): if letter== 'A' or'A-' or 'A+': #checks to see if it was an A, A+ or A- if letter == 'A': #assigns the corresponding number value to the letter and prints it print(str(4)) elif letter == 'A+': print(str(4+0.3)) elif letter== 'A-': print(str(4-0.3)) if letter == 'B' or 'B-' or 'B+': #checks to see if it was an B, B+ or B- if letter == 'B': #assigns the corresponding number value to the letter and prints it print(str(3.0)) elif letter == 'B+': print(str(3+0.3)) elif letter == 'B-': print(str(3-0.3)) if letter == 'C' or 'C-' or 'C+': #checks to see if it was an C, C+ or C- if letter== 'C': #assigns the corresponding number value to the letter and prints it print(str(2.0)) elif letter== 'C+': print(str(2+0.3)) elif letter== 'C-': print(str(2-0.3)) if letter == 'D' or 'D-' or 'D+': #checks to see if it was an D, D+ or D- if letter == 'D': #assigns the corresponding number value to the letter and prints it print(str(1)) elif letter=='D+': print(str(1+0.3)) elif letter=='D-': print(str(1-0.3)) if letter == 'F': print(str(0)) ######################## # Question 10 ######################## ''' creates a function called is_Palindrome that takes one input that is a string and determines if that string is the exact same when reversed and returns true if it is and false if it is not ''' def is_Palindrome(string): reverse = string[::-1] #reverses the string if reverse == string: return reverse == string else: return reverse == string ######################## # Question 12 ######################## ''' creates a function called rps that is a game of rock, paper, scissors and takes two input paramteres that are strings that are either 'S' 'P' 'R' and returns a -1 if player 1 wins and a 1 if player two wins and a zero if its a tie''' def rps(player1, player2): if player1 == 'R' and player2 == 'S' or player1 == 'P' and player2 == 'R' or player1 == 'S' and player2 == 'P': return -1 elif player2 == 'R' and player1 == 'S' or player2 == 'P' and player1 == 'R' or player2 == 'S' and player1 == 'P': return 1 else: return 0 ######################## # Question 14 ######################## ''' creates a function called count_even_digits that takes two arguments and returns the amount of even numbers in the first paramater and the second paramtere is the amount of digits in the first one ''' def count_even_digits(int1, int2): even = str(int1) even = int(even.count('0')) + int(even.count('2')) + int(even.count('4')) + int(even.count('6')) + int(even.count('8')) #uses the count operator to see the amount of even numbers print(str(even)) ######################## # Question 13 ######################## ''' creates a function called alogical that takes one parameter n which must be a integer and counts how many times that integer must be divided by 2 to be less than or equal to 1 ''' def alogical(n): counter = 0 while n > 1 or n==1: # while loop is used to count the amount of times needed to divide by 2 n = n/2 counter = counter + 1 print(str(counter
reverse1 = num%10 # gets you the number of the begining of the reversed integer reverse2 = (num%100)-reverse1 # gets you the number of the second number in the reversed inetger reverse1 = reverse1*100 # makes the number at the end now the begining reverse3 = num//100 #gives you last number in the reversed integer final = reverse1 + reverse2 + reverse3 # adds all the numbers together to get you the reversed int return final #function returns the reversed int
identifier_body
a1_300068688.py.py
n' times print(str(yoda)) #prints 'yoda' ######################## # Question 2 ######################## ''' creates a function called is_prime that takes one argument and check to see whether it is prime or not, if it is a prime the function return true if not it returns else''' def is_prime(n): if n >= 2: # prime numbers are numbers that cant be divided unless its by 1 or it self so 'n' cannot be 1 or 2 for i in range(2, n): #creates a for loop that starts at 2 and ends at 'n' if (n%i == 0): #checks to see whether n divides into i intergerally return n<0 #if it does return false becuase n is not a prime number then else: return n>0 #returns true because n is a prime ######################## # Question 3 ######################## ''' Creates a function points that takes 4 arguments which make up to coordinates x1,y1 x2,y2 this function then computes the slope of the the line of these points and the distance between them and outputs a message a message''' def points(x1,y1,x2,y2): distance = (((x2-x1)**2)+((y2-y1)**2))**(1/2) # finds the distance between these points by finding the hypotenuse of the triangle a^2 + b^2 = c^2 if (x2-x1)!= 0 and (y2-y1)!=0: #checks to see if the slope is not 0 and if it is print a special message slope = (y2-y1)/(x2-x1) #computes slope print("The slope is " + str(slope) + " and the distance is " + str(distance)) #prints out slope and distance else: print("the slope is infinity and the distance is " + str(distance))# slope is infinity so it outprints just the distance ######################## # Question 5 ######################## ''' creates a function reverse_int that takes one argument that is a integer and reverses this integer''' def reverse_int(num): reverse1 = num%10 # gets you the number of the begining of the reversed integer reverse2 = (num%100)-reverse1 # gets you the number of the second number in the reversed inetger reverse1 = reverse1*100 # makes the number at the end now the begining reverse3 = num//100 #gives you last number in the reversed integer final = reverse1 + reverse2 + reverse3 # adds all the numbers together to get you the reversed int return final #function returns the reversed int ######################## # Question 6 ######################## ''' creates a function vowelCount that takes one argument which is a string and then prints out the number of vowels and what vowels are in the word''' def vowelCount(string): print('a, e, i, o, and u appear, respectively, ' + str(string.count('a'))+ ', ' + str(string.count('e'))+ ', ' + str(string.count('i')) + ', ' + str(string.count('o')) + ', ' + str(string.count('u'))+ ' times') ######################## # Question 7 ######################## ''' creayes a function with 3 input paramteres that are strings and checks to see if all the same then returns true if they are and false if they are not''' def allthesame(x,y,z): if x==y and x==z and z==y: #if statment that checks whther they are all the sale return x==y else: return 1<0 ''' creates a function alldifferent that has 3 input paramteres that are all strings and checks whether all the strings are different and returns true if they are and false if they are not''' def alldifferent(x,y,z): if x != y and y != z and x != z: #checks to see whether the strings x, y and z are all different return x!=y else: return 1<0 ''' creates a function called sorted that takes 3 input paramteres and checks to see whether they are in sorted order and returns true if they are and false if they are not''' def sorted(x,y,z): if z>y and z>x and y>x: #checks whther they are sorted return z>y else: return 1<0 ######################## # Question 8 ######################## ''' creates a function called leap that takes on input value that is an integer , this function then computes whether that number was a leap year and returns true if it was and false if it isnt''' def leap(year): if year%4==0 and year%100 != 0 or year%400==0: #checks whether the year was a leap year return year>0 else: return year<0 ######################## # Question 9 ######################## ''' creates a function called letter2number tthat takes a string input that is a letter grade and converts them to a number grade''' def
(letter): if letter== 'A' or'A-' or 'A+': #checks to see if it was an A, A+ or A- if letter == 'A': #assigns the corresponding number value to the letter and prints it print(str(4)) elif letter == 'A+': print(str(4+0.3)) elif letter== 'A-': print(str(4-0.3)) if letter == 'B' or 'B-' or 'B+': #checks to see if it was an B, B+ or B- if letter == 'B': #assigns the corresponding number value to the letter and prints it print(str(3.0)) elif letter == 'B+': print(str(3+0.3)) elif letter == 'B-': print(str(3-0.3)) if letter == 'C' or 'C-' or 'C+': #checks to see if it was an C, C+ or C- if letter== 'C': #assigns the corresponding number value to the letter and prints it print(str(2.0)) elif letter== 'C+': print(str(2+0.3)) elif letter== 'C-': print(str(2-0.3)) if letter == 'D' or 'D-' or 'D+': #checks to see if it was an D, D+ or D- if letter == 'D': #assigns the corresponding number value to the letter and prints it print(str(1)) elif letter=='D+': print(str(1+0.3)) elif letter=='D-': print(str(1-0.3)) if letter == 'F': print(str(0)) ######################## # Question 10 ######################## ''' creates a function called is_Palindrome that takes one input that is a string and determines if that string is the exact same when reversed and returns true if it is and false if it is not ''' def is_Palindrome(string): reverse = string[::-1] #reverses the string if reverse == string: return reverse == string else: return reverse == string ######################## # Question 12 ######################## ''' creates a function called rps that is a game of rock, paper, scissors and takes two input paramteres that are strings that are either 'S' 'P' 'R' and returns a -1 if player 1 wins and a 1 if player two wins and a zero if its a tie''' def rps(player1, player2): if player1 == 'R' and player2 == 'S' or player1 == 'P' and player2 == 'R' or player1 == 'S' and player2 == 'P': return -1 elif player2 == 'R' and player1 == 'S' or player2 == 'P' and player1 == 'R' or player2 == 'S' and player1 == 'P': return 1 else: return 0 ######################## # Question 14 ######################## ''' creates a function called count_even_digits that takes two arguments and returns the amount of even numbers in the first paramater and the second paramtere is the amount of digits in the first one ''' def count_even_digits(int1, int2): even = str(int1) even = int(even.count('0')) + int(even.count('2')) + int(even.count('4')) + int(even.count('6')) + int(even.count('8')) #uses the count operator to see the amount of even numbers print(str(even)) ######################## # Question 13 ######################## ''' creates a function called alogical that takes one parameter n which must be a integer and counts how many times that integer must be divided by 2 to be less than or equal to 1 ''' def alogical(n): counter = 0 while n > 1 or n==1: # while loop is used to count the amount of times needed to divide by 2 n = n/2 counter = counter + 1 print(str(counter
letter2number
identifier_name
a1_300068688.py.py
delim and repeats 'n' times print(str(yoda)) #prints 'yoda' ######################## # Question 2 ######################## ''' creates a function called is_prime that takes one argument and check to see whether it is prime or not, if it is a prime the function return true if not it returns else''' def is_prime(n): if n >= 2: # prime numbers are numbers that cant be divided unless its by 1 or it self so 'n' cannot be 1 or 2 for i in range(2, n): #creates a for loop that starts at 2 and ends at 'n' if (n%i == 0): #checks to see whether n divides into i intergerally return n<0 #if it does return false becuase n is not a prime number then else: return n>0 #returns true because n is a prime ######################## # Question 3 ######################## ''' Creates a function points that takes 4 arguments which make up to coordinates x1,y1 x2,y2 this function then computes the slope of the the line of these points and the distance between them and outputs a message a message''' def points(x1,y1,x2,y2): distance = (((x2-x1)**2)+((y2-y1)**2))**(1/2) # finds the distance between these points by finding the hypotenuse of the triangle a^2 + b^2 = c^2 if (x2-x1)!= 0 and (y2-y1)!=0: #checks to see if the slope is not 0 and if it is print a special message slope = (y2-y1)/(x2-x1) #computes slope print("The slope is " + str(slope) + " and the distance is " + str(distance)) #prints out slope and distance else: print("the slope is infinity and the distance is " + str(distance))# slope is infinity so it outprints just the distance ######################## # Question 5 ######################## ''' creates a function reverse_int that takes one argument that is a integer and reverses this integer''' def reverse_int(num): reverse1 = num%10 # gets you the number of the begining of the reversed integer reverse2 = (num%100)-reverse1 # gets you the number of the second number in the reversed inetger reverse1 = reverse1*100 # makes the number at the end now the begining reverse3 = num//100 #gives you last number in the reversed integer final = reverse1 + reverse2 + reverse3 # adds all the numbers together to get you the reversed int return final #function returns the reversed int ######################## # Question 6 ######################## ''' creates a function vowelCount that takes one argument which is a string and then prints out the number of vowels and what vowels are in the word''' def vowelCount(string): print('a, e, i, o, and u appear, respectively, ' + str(string.count('a'))+ ', ' + str(string.count('e'))+ ', ' + str(string.count('i')) + ', ' + str(string.count('o')) + ', ' + str(string.count('u'))+ ' times') ######################## # Question 7 ######################## ''' creayes a function with 3 input paramteres that are strings and checks to see if all the same then returns true if they are and false if they are not''' def allthesame(x,y,z): if x==y and x==z and z==y: #if statment that checks whther they are all the sale return x==y else: return 1<0 ''' creates a function alldifferent that has 3 input paramteres that are all strings and checks whether all the strings are different and returns true if they are and false if they are not''' def alldifferent(x,y,z): if x != y and y != z and x != z: #checks to see whether the strings x, y and z are all different return x!=y else: return 1<0 ''' creates a function called sorted that takes 3 input paramteres and checks to see whether they are in sorted order and returns true if they are and false if they are not''' def sorted(x,y,z): if z>y and z>x and y>x: #checks whther they are sorted return z>y else: return 1<0 ######################## # Question 8 ######################## ''' creates a function called leap that takes on input value that is an integer , this function then computes whether that number was a leap year and returns true if it was and false if it isnt''' def leap(year): if year%4==0 and year%100 != 0 or year%400==0: #checks whether the year was a leap year return year>0 else: return year<0 ######################## # Question 9 ######################## ''' creates a function called letter2number tthat takes a string input that is a letter grade and converts them to a number grade''' def letter2number(letter): if letter== 'A' or'A-' or 'A+': #checks to see if it was an A, A+ or A- if letter == 'A': #assigns the corresponding number value to the letter and prints it print(str(4))
print(str(4+0.3)) elif letter== 'A-': print(str(4-0.3)) if letter == 'B' or 'B-' or 'B+': #checks to see if it was an B, B+ or B- if letter == 'B': #assigns the corresponding number value to the letter and prints it print(str(3.0)) elif letter == 'B+': print(str(3+0.3)) elif letter == 'B-': print(str(3-0.3)) if letter == 'C' or 'C-' or 'C+': #checks to see if it was an C, C+ or C- if letter== 'C': #assigns the corresponding number value to the letter and prints it print(str(2.0)) elif letter== 'C+': print(str(2+0.3)) elif letter== 'C-': print(str(2-0.3)) if letter == 'D' or 'D-' or 'D+': #checks to see if it was an D, D+ or D- if letter == 'D': #assigns the corresponding number value to the letter and prints it print(str(1)) elif letter=='D+': print(str(1+0.3)) elif letter=='D-': print(str(1-0.3)) if letter == 'F': print(str(0)) ######################## # Question 10 ######################## ''' creates a function called is_Palindrome that takes one input that is a string and determines if that string is the exact same when reversed and returns true if it is and false if it is not ''' def is_Palindrome(string): reverse = string[::-1] #reverses the string if reverse == string: return reverse == string else: return reverse == string ######################## # Question 12 ######################## ''' creates a function called rps that is a game of rock, paper, scissors and takes two input paramteres that are strings that are either 'S' 'P' 'R' and returns a -1 if player 1 wins and a 1 if player two wins and a zero if its a tie''' def rps(player1, player2): if player1 == 'R' and player2 == 'S' or player1 == 'P' and player2 == 'R' or player1 == 'S' and player2 == 'P': return -1 elif player2 == 'R' and player1 == 'S' or player2 == 'P' and player1 == 'R' or player2 == 'S' and player1 == 'P': return 1 else: return 0 ######################## # Question 14 ######################## ''' creates a function called count_even_digits that takes two arguments and returns the amount of even numbers in the first paramater and the second paramtere is the amount of digits in the first one ''' def count_even_digits(int1, int2): even = str(int1) even = int(even.count('0')) + int(even.count('2')) + int(even.count('4')) + int(even.count('6')) + int(even.count('8')) #uses the count operator to see the amount of even numbers print(str(even)) ######################## # Question 13 ######################## ''' creates a function called alogical that takes one parameter n which must be a integer and counts how many times that integer must be divided by 2 to be less than or equal to 1 ''' def alogical(n): counter = 0 while n > 1 or n==1: # while loop is used to count the amount of times needed to divide by 2 n = n/2 counter = counter +
elif letter == 'A+':
random_line_split
eventsGetUtils.js
' | 'trashed', modifiedSince?: number, includeDeletions?: boolean, }; export type StoreQuery = { id: string, storeId: string, includeTrashed: boolean, expandChildren: boolean, excludedIds: Array<string>, }; let mall; /** * # Stream Query Flow * 1. coerceStreamParam: * - null `streams` is changed to `[{any: ['*]}] * - transform "stringified" `streams` by parsing JSON object * * 2. transformArrayOfStringsToStreamsQuery: * For backwardCompatibility with older streams parameter ['A', 'B'] * - `streams: ['A', 'B', 'C']` => `streams: [{any: 'A'}, {any: 'B'}, {any: 'C'}]` * * 3. validateStreamsQueriesAndSetStore: * - Check syntax and add storeId * `streams: [{any: 'A'}, {any: ':_audit:B'}]` => `streams: [{any: 'A', storeId: 'local'}, {any: 'B', storeId: 'audit'}]` * * 4. streamQueryCheckPermissionsAndReplaceStars: * For `stream.any`ONLY ! (we don't have to check NOT and ALL query as they only reduce scope) * - check if stream exits and if has "read" access * - If "stream.any" contains "*" it's replaced by all root streams with "read" rights * * 5. streamQueryAddForcedAndForbiddenStreams * - Add to streams query `all` streams declared as "forced" * - Add to streams query `not` streams that must not be exposed permissions => with level = "none" * * 6. streamQueryExpandStreams * - Each "streamId" of the queries is "expanded" (i.e. transformed in an array of streamId that includes the streams and it's chidlren) * - Do not expand streams prefixed with a "#" * * - A callBack `expandStreamInContext`is used to link the expand process and the "store" * This callBack is designed to be optimized on a Per-Store basis The current implementation is generic * - If streamId is prefixed with a "#" just return the streamId without "#" * - It queries the stores with and standard `store.streams.get({id: streamId, exludedIds: [....]})` * and return an array of streams. * * - streamsQueryUtils.expandAndTransformStreamQueries * Is in charge of handling 'any', 'all' and 'not' "expand" process * * - "any" is expanded first excluding streamIds in "not" * => The result is kept in `any` * - "all" is expanded in second excluding streamIds in "not" * `all` is tranformed and each "expansion" is kept in `and: [{any: ,..}]` * example: `{all: ['A', 'B']}` => `{and: [{any: [...expand('A')]}, {any: [...expand('B')]}]}` * - "not" is expanded in third and added to `and` -- !! we exclude streamIds that are in 'any' as some authorization might have been given on child now expanded * example: `{all: ['A'], not['B', 'C']}` => `{and: [{any: [...expand('A')]}, {not: [...expand('B')...expand('C')]}]} * */ function coerceStreamsParam(context: MethodContext, params: GetEventsParams, result: Result, next: ApiCallback) { if (params.streams == null) { return next(); } if (! context.acceptStreamsQueryNonStringified) { if (isStringifiedJSON(params.streams)) { try { params.streams = parseStreamsParams(params.streams); } catch (e) { return next(e); } } else if (isStringOrArrayOfStrings(params.streams)) { // good, do nothing } else { return next(errors.invalidRequestStructure('Invalid "streams" parameter. It should be an array of streamIds or JSON logical query.')) } } else { if (isStringifiedJSON(params.streams)) { try { params.streams = parseStreamsParams(params.streams); } catch (e) { return next(e); } } else { // good, do nothing } } // Transform object or string to Array if (!Array.isArray(params.streams)) { params.streams = [params.streams]; } next(); function parseStreamsParams(input: string): ?StreamQuery | ?Array<StreamQuery> { try { return JSON.parse(input); } catch (e) { throw errors.invalidRequestStructure('Invalid "streams" parameter. It should be an array of streamIds or JSON logical query. Error while parsing JSON ' + e, input); } } /** * we detect if it's JSON by looking at first char. * Note: since RFC 7159 JSON can also starts with ", true, false or number - this does not apply in this case. * @param {string} input */ function isStringifiedJSON(input: any): boolean { return (typeof input === 'string') && ['[', '{'].includes(input.substr(0, 1)); } function isStringOrArrayOfStrings(input: any): boolean {
async function applyDefaultsForRetrieval(context: MethodContext, params: GetEventsParams, result: Result, next: ApiCallback) { _.defaults(params, { streams: [{ any: ['*'] }], tags: null, types: null, fromTime: null, toTime: null, sortAscending: false, skip: null, limit: null, state: 'default', modifiedSince: null, includeDeletions: false }); if (params.fromTime == null && params.toTime != null) { params.fromTime = timestamp.add(params.toTime, -24 * 60 * 60); } if (params.fromTime != null && params.toTime == null) { params.toTime = timestamp.now(); } if (params.fromTime == null && params.toTime == null && params.limit == null) { // limit to 20 items by default params.limit = 20; } next(); } function transformArrayOfStringsToStreamsQuery(context: MethodContext, params: GetEventsParams, result: Result, next: ApiCallback) { try { params.arrayOfStreamQueries = streamsQueryUtils.transformArrayOfStringsToStreamsQuery(params.streams); } catch (e) { return next(errors.invalidRequestStructure(e, params.streams)); } next(); } function validateStreamsQueriesAndSetStore(context: MethodContext, params: GetEventsParams, result: Result, next: ApiCallback) { try { streamsQueryUtils.validateStreamsQueriesAndSetStore(params.arrayOfStreamQueries); params.arrayOfStreamQueriesWithStoreId = params.arrayOfStreamQueries; } catch (e) { return next(errors.invalidRequestStructure('Initial filtering: ' + e, params.streams)); } next(); } // the two tasks are joined as '*' replaced have their permissions checked async function streamQueryCheckPermissionsAndReplaceStars(context: MethodContext, params: GetEventsParams, result: Result, next: ApiCallback) { context.tracing.startSpan('streamQueries'); const unAuthorizedStreamIds: Array<string> = []; const unAccessibleStreamIds: Array<string> = []; async function streamExistsAndCanGetEventsOnStream(streamId: string, storeId: string, unAuthorizedStreamIds: Array<string>, unAccessibleStreamIds: Array<string>): Promise<void> { // remove eventual '#' in streamQuery const cleanStreamId: string = streamId.startsWith('#') ? streamId.substr(1) : streamId; const stream: Stream = await context.streamForStreamId(cleanStreamId, storeId); if (stream == null) { unAccessibleStreamIds.push(cleanStreamId); return; } if (! await context.access.canGetEventsOnStream(cleanStreamId, storeId)) { unAuthorizedStreamIds.push(cleanStreamId); } } for (const streamQuery: StreamQueryWithStoreId of params.arrayOfStreamQueriesWithStoreId) { // ------------ "*" case if (streamQuery.any && streamQuery.any.includes('*')) { if (await context.access.canGetEventsOnStream('*', streamQuery.storeId)) continue; // We can keep star // replace any by allowed streams for reading const canReadStreamIds: Array<string> = []; for (const streamPermission of context.access.getStoresPermissions(streamQuery.storeId)) { if (await context.access.canGetEventsOnStream(streamPermission.streamId, streamQuery.storeId)) { canReadStreamIds.push(streamPermission.streamId); }
if (typeof input === 'string') return true; if (! Array.isArray(input)) return false; for (const item of input) { if (typeof item !== 'string') return false; } return true; } }
identifier_body
eventsGetUtils.js
' | 'trashed', modifiedSince?: number, includeDeletions?: boolean, }; export type StoreQuery = { id: string, storeId: string, includeTrashed: boolean, expandChildren: boolean, excludedIds: Array<string>, }; let mall; /** * # Stream Query Flow * 1. coerceStreamParam: * - null `streams` is changed to `[{any: ['*]}] * - transform "stringified" `streams` by parsing JSON object * * 2. transformArrayOfStringsToStreamsQuery: * For backwardCompatibility with older streams parameter ['A', 'B'] * - `streams: ['A', 'B', 'C']` => `streams: [{any: 'A'}, {any: 'B'}, {any: 'C'}]` * * 3. validateStreamsQueriesAndSetStore: * - Check syntax and add storeId * `streams: [{any: 'A'}, {any: ':_audit:B'}]` => `streams: [{any: 'A', storeId: 'local'}, {any: 'B', storeId: 'audit'}]` * * 4. streamQueryCheckPermissionsAndReplaceStars: * For `stream.any`ONLY ! (we don't have to check NOT and ALL query as they only reduce scope) * - check if stream exits and if has "read" access * - If "stream.any" contains "*" it's replaced by all root streams with "read" rights * * 5. streamQueryAddForcedAndForbiddenStreams * - Add to streams query `all` streams declared as "forced" * - Add to streams query `not` streams that must not be exposed permissions => with level = "none" * * 6. streamQueryExpandStreams * - Each "streamId" of the queries is "expanded" (i.e. transformed in an array of streamId that includes the streams and it's chidlren) * - Do not expand streams prefixed with a "#" * * - A callBack `expandStreamInContext`is used to link the expand process and the "store" * This callBack is designed to be optimized on a Per-Store basis The current implementation is generic * - If streamId is prefixed with a "#" just return the streamId without "#" * - It queries the stores with and standard `store.streams.get({id: streamId, exludedIds: [....]})` * and return an array of streams. * * - streamsQueryUtils.expandAndTransformStreamQueries * Is in charge of handling 'any', 'all' and 'not' "expand" process * * - "any" is expanded first excluding streamIds in "not" * => The result is kept in `any` * - "all" is expanded in second excluding streamIds in "not" * `all` is tranformed and each "expansion" is kept in `and: [{any: ,..}]` * example: `{all: ['A', 'B']}` => `{and: [{any: [...expand('A')]}, {any: [...expand('B')]}]}` * - "not" is expanded in third and added to `and` -- !! we exclude streamIds that are in 'any' as some authorization might have been given on child now expanded * example: `{all: ['A'], not['B', 'C']}` => `{and: [{any: [...expand('A')]}, {not: [...expand('B')...expand('C')]}]} * */ function coerceStreamsParam(context: MethodContext, params: GetEventsParams, result: Result, next: ApiCallback) { if (params.streams == null) {
if (! context.acceptStreamsQueryNonStringified) { if (isStringifiedJSON(params.streams)) { try { params.streams = parseStreamsParams(params.streams); } catch (e) { return next(e); } } else if (isStringOrArrayOfStrings(params.streams)) { // good, do nothing } else { return next(errors.invalidRequestStructure('Invalid "streams" parameter. It should be an array of streamIds or JSON logical query.')) } } else { if (isStringifiedJSON(params.streams)) { try { params.streams = parseStreamsParams(params.streams); } catch (e) { return next(e); } } else { // good, do nothing } } // Transform object or string to Array if (!Array.isArray(params.streams)) { params.streams = [params.streams]; } next(); function parseStreamsParams(input: string): ?StreamQuery | ?Array<StreamQuery> { try { return JSON.parse(input); } catch (e) { throw errors.invalidRequestStructure('Invalid "streams" parameter. It should be an array of streamIds or JSON logical query. Error while parsing JSON ' + e, input); } } /** * we detect if it's JSON by looking at first char. * Note: since RFC 7159 JSON can also starts with ", true, false or number - this does not apply in this case. * @param {string} input */ function isStringifiedJSON(input: any): boolean { return (typeof input === 'string') && ['[', '{'].includes(input.substr(0, 1)); } function isStringOrArrayOfStrings(input: any): boolean { if (typeof input === 'string') return true; if (! Array.isArray(input)) return false; for (const item of input) { if (typeof item !== 'string') return false; } return true; } } async function applyDefaultsForRetrieval(context: MethodContext, params: GetEventsParams, result: Result, next: ApiCallback) { _.defaults(params, { streams: [{ any: ['*'] }], tags: null, types: null, fromTime: null, toTime: null, sortAscending: false, skip: null, limit: null, state: 'default', modifiedSince: null, includeDeletions: false }); if (params.fromTime == null && params.toTime != null) { params.fromTime = timestamp.add(params.toTime, -24 * 60 * 60); } if (params.fromTime != null && params.toTime == null) { params.toTime = timestamp.now(); } if (params.fromTime == null && params.toTime == null && params.limit == null) { // limit to 20 items by default params.limit = 20; } next(); } function transformArrayOfStringsToStreamsQuery(context: MethodContext, params: GetEventsParams, result: Result, next: ApiCallback) { try { params.arrayOfStreamQueries = streamsQueryUtils.transformArrayOfStringsToStreamsQuery(params.streams); } catch (e) { return next(errors.invalidRequestStructure(e, params.streams)); } next(); } function validateStreamsQueriesAndSetStore(context: MethodContext, params: GetEventsParams, result: Result, next: ApiCallback) { try { streamsQueryUtils.validateStreamsQueriesAndSetStore(params.arrayOfStreamQueries); params.arrayOfStreamQueriesWithStoreId = params.arrayOfStreamQueries; } catch (e) { return next(errors.invalidRequestStructure('Initial filtering: ' + e, params.streams)); } next(); } // the two tasks are joined as '*' replaced have their permissions checked async function streamQueryCheckPermissionsAndReplaceStars(context: MethodContext, params: GetEventsParams, result: Result, next: ApiCallback) { context.tracing.startSpan('streamQueries'); const unAuthorizedStreamIds: Array<string> = []; const unAccessibleStreamIds: Array<string> = []; async function streamExistsAndCanGetEventsOnStream(streamId: string, storeId: string, unAuthorizedStreamIds: Array<string>, unAccessibleStreamIds: Array<string>): Promise<void> { // remove eventual '#' in streamQuery const cleanStreamId: string = streamId.startsWith('#') ? streamId.substr(1) : streamId; const stream: Stream = await context.streamForStreamId(cleanStreamId, storeId); if (stream == null) { unAccessibleStreamIds.push(cleanStreamId); return; } if (! await context.access.canGetEventsOnStream(cleanStreamId, storeId)) { unAuthorizedStreamIds.push(cleanStreamId); } } for (const streamQuery: StreamQueryWithStoreId of params.arrayOfStreamQueriesWithStoreId) { // ------------ "*" case if (streamQuery.any && streamQuery.any.includes('*')) { if (await context.access.canGetEventsOnStream('*', streamQuery.storeId)) continue; // We can keep star // replace any by allowed streams for reading const canReadStreamIds: Array<string> = []; for (const streamPermission of context.access.getStoresPermissions(streamQuery.storeId)) { if (await context.access.canGetEventsOnStream(streamPermission.streamId, streamQuery.storeId)) { canReadStreamIds.push(streamPermission.streamId); } }
return next(); }
conditional_block
eventsGetUtils.js
{ params.streams = parseStreamsParams(params.streams); } catch (e) { return next(e); } } else if (isStringOrArrayOfStrings(params.streams)) { // good, do nothing } else { return next(errors.invalidRequestStructure('Invalid "streams" parameter. It should be an array of streamIds or JSON logical query.')) } } else { if (isStringifiedJSON(params.streams)) { try { params.streams = parseStreamsParams(params.streams); } catch (e) { return next(e); } } else { // good, do nothing } } // Transform object or string to Array if (!Array.isArray(params.streams)) { params.streams = [params.streams]; } next(); function parseStreamsParams(input: string): ?StreamQuery | ?Array<StreamQuery> { try { return JSON.parse(input); } catch (e) { throw errors.invalidRequestStructure('Invalid "streams" parameter. It should be an array of streamIds or JSON logical query. Error while parsing JSON ' + e, input); } } /** * we detect if it's JSON by looking at first char. * Note: since RFC 7159 JSON can also starts with ", true, false or number - this does not apply in this case. * @param {string} input */ function isStringifiedJSON(input: any): boolean { return (typeof input === 'string') && ['[', '{'].includes(input.substr(0, 1)); } function isStringOrArrayOfStrings(input: any): boolean { if (typeof input === 'string') return true; if (! Array.isArray(input)) return false; for (const item of input) { if (typeof item !== 'string') return false; } return true; } } async function applyDefaultsForRetrieval(context: MethodContext, params: GetEventsParams, result: Result, next: ApiCallback) { _.defaults(params, { streams: [{ any: ['*'] }], tags: null, types: null, fromTime: null, toTime: null, sortAscending: false, skip: null, limit: null, state: 'default', modifiedSince: null, includeDeletions: false }); if (params.fromTime == null && params.toTime != null) { params.fromTime = timestamp.add(params.toTime, -24 * 60 * 60); } if (params.fromTime != null && params.toTime == null) { params.toTime = timestamp.now(); } if (params.fromTime == null && params.toTime == null && params.limit == null) { // limit to 20 items by default params.limit = 20; } next(); } function transformArrayOfStringsToStreamsQuery(context: MethodContext, params: GetEventsParams, result: Result, next: ApiCallback) { try { params.arrayOfStreamQueries = streamsQueryUtils.transformArrayOfStringsToStreamsQuery(params.streams); } catch (e) { return next(errors.invalidRequestStructure(e, params.streams)); } next(); } function validateStreamsQueriesAndSetStore(context: MethodContext, params: GetEventsParams, result: Result, next: ApiCallback) { try { streamsQueryUtils.validateStreamsQueriesAndSetStore(params.arrayOfStreamQueries); params.arrayOfStreamQueriesWithStoreId = params.arrayOfStreamQueries; } catch (e) { return next(errors.invalidRequestStructure('Initial filtering: ' + e, params.streams)); } next(); } // the two tasks are joined as '*' replaced have their permissions checked async function streamQueryCheckPermissionsAndReplaceStars(context: MethodContext, params: GetEventsParams, result: Result, next: ApiCallback) { context.tracing.startSpan('streamQueries'); const unAuthorizedStreamIds: Array<string> = []; const unAccessibleStreamIds: Array<string> = []; async function streamExistsAndCanGetEventsOnStream(streamId: string, storeId: string, unAuthorizedStreamIds: Array<string>, unAccessibleStreamIds: Array<string>): Promise<void> { // remove eventual '#' in streamQuery const cleanStreamId: string = streamId.startsWith('#') ? streamId.substr(1) : streamId; const stream: Stream = await context.streamForStreamId(cleanStreamId, storeId); if (stream == null) { unAccessibleStreamIds.push(cleanStreamId); return; } if (! await context.access.canGetEventsOnStream(cleanStreamId, storeId)) { unAuthorizedStreamIds.push(cleanStreamId); } } for (const streamQuery: StreamQueryWithStoreId of params.arrayOfStreamQueriesWithStoreId) { // ------------ "*" case if (streamQuery.any && streamQuery.any.includes('*')) { if (await context.access.canGetEventsOnStream('*', streamQuery.storeId)) continue; // We can keep star // replace any by allowed streams for reading const canReadStreamIds: Array<string> = []; for (const streamPermission of context.access.getStoresPermissions(streamQuery.storeId)) { if (await context.access.canGetEventsOnStream(streamPermission.streamId, streamQuery.storeId)) { canReadStreamIds.push(streamPermission.streamId); } } streamQuery.any = canReadStreamIds; } else { // ------------ All other cases /** * ! we don't have to check for permissions on 'all' or 'not' as long there is at least one 'any' authorized. */ if (streamQuery?.any?.length === 0) { return next(errors.invalidRequestStructure('streamQueries must have a valid {any: [...]} component')); } for (const streamId: string of streamQuery.any) { await streamExistsAndCanGetEventsOnStream(streamId, streamQuery.storeId, unAuthorizedStreamIds, unAccessibleStreamIds); }; } } if (unAuthorizedStreamIds.length > 0) { context.tracing.finishSpan('streamQueries'); return next(errors.forbidden('stream [' + unAuthorizedStreamIds[0] + '] has not sufficent permission to get events')); } if (unAccessibleStreamIds.length > 0) { context.tracing.finishSpan('streamQueries'); return next(errors.unknownReferencedResource( 'stream' + (unAccessibleStreamIds.length > 1 ? 's' : ''), 'streams', unAccessibleStreamIds)); } next(); } /** * Add "forced" and "none" events from permissions */ function streamQueryAddForcedAndForbiddenStreams(context: MethodContext, params: GetEventsParams, result: Result, next: ApiCallback) { for (const streamQuery: StreamQueryWithStoreId of params.arrayOfStreamQueriesWithStoreId) { // ------------ ALL --------------- // // add forced Streams if exists const forcedStreamIds: Array<string> = context.access.getForcedStreamsGetEventsStreamIds(streamQuery.storeId); if (forcedStreamIds?.length > 0) { if (streamQuery.all == null) streamQuery.all = []; // TODO check for duplicates streamQuery.all.push(...forcedStreamIds); } // ------------- NOT ------------- // const forbiddenStreamIds: Array<string> = context.access.getForbiddenGetEventsStreamIds(streamQuery.storeId); if (forbiddenStreamIds?.length > 0) { if (streamQuery.not == null) streamQuery.not = []; // TODO check for duplicates streamQuery.not.push(...forbiddenStreamIds); } } next(); } async function streamQueryExpandStreams(context: MethodContext, params: GetEventsParams, result: Result, next: ApiCallback) { async function expandStreamInContext(streamId: string, storeId: string, excludedIds) { // remove eventual '#' in streamQuery if (streamId.startsWith('#')) { return [streamId.substr(1)]; // do not expand Stream } const query: StoreQuery = { id: streamId, storeId: storeId, includeTrashed: params.state === 'all' || params.state === 'trashed', expandChildren: true, excludedIds: excludedIds }; const tree: Array<Stream> = await mall.streams.get(context.user.id, query); // collect streamIds const resultWithPrefix: Array<string> = treeUtils.collectPluck(tree, 'id'); // remove storePrefix const result: Array<string> = resultWithPrefix.map((fullStreamId: string) => StreamsUtils.storeIdAndStreamIdForStreamId(fullStreamId)[1]); return result; } try { params.arrayOfStreamQueriesWithStoreId = await streamsQueryUtils.expandAndTransformStreamQueries(params.arrayOfStreamQueriesWithStoreId, expandStreamInContext); } catch (e) { console.log(e); context.tracing.finishSpan('streamQueries'); return next(e); } // delete streamQueries with no inclusions params.arrayOfStreamQueriesWithStoreId = params.arrayOfStreamQueriesWithStoreId.filter(streamQuery => streamQuery.any != null || streamQuery.and != null); context.tracing.finishSpan('streamQueries'); next(); }
/** * - Create a copy of the params per query
random_line_split
eventsGetUtils.js
' | 'trashed', modifiedSince?: number, includeDeletions?: boolean, }; export type StoreQuery = { id: string, storeId: string, includeTrashed: boolean, expandChildren: boolean, excludedIds: Array<string>, }; let mall; /** * # Stream Query Flow * 1. coerceStreamParam: * - null `streams` is changed to `[{any: ['*]}] * - transform "stringified" `streams` by parsing JSON object * * 2. transformArrayOfStringsToStreamsQuery: * For backwardCompatibility with older streams parameter ['A', 'B'] * - `streams: ['A', 'B', 'C']` => `streams: [{any: 'A'}, {any: 'B'}, {any: 'C'}]` * * 3. validateStreamsQueriesAndSetStore: * - Check syntax and add storeId * `streams: [{any: 'A'}, {any: ':_audit:B'}]` => `streams: [{any: 'A', storeId: 'local'}, {any: 'B', storeId: 'audit'}]` * * 4. streamQueryCheckPermissionsAndReplaceStars: * For `stream.any`ONLY ! (we don't have to check NOT and ALL query as they only reduce scope) * - check if stream exits and if has "read" access * - If "stream.any" contains "*" it's replaced by all root streams with "read" rights * * 5. streamQueryAddForcedAndForbiddenStreams * - Add to streams query `all` streams declared as "forced" * - Add to streams query `not` streams that must not be exposed permissions => with level = "none" * * 6. streamQueryExpandStreams * - Each "streamId" of the queries is "expanded" (i.e. transformed in an array of streamId that includes the streams and it's chidlren) * - Do not expand streams prefixed with a "#" * * - A callBack `expandStreamInContext`is used to link the expand process and the "store" * This callBack is designed to be optimized on a Per-Store basis The current implementation is generic * - If streamId is prefixed with a "#" just return the streamId without "#" * - It queries the stores with and standard `store.streams.get({id: streamId, exludedIds: [....]})` * and return an array of streams. * * - streamsQueryUtils.expandAndTransformStreamQueries * Is in charge of handling 'any', 'all' and 'not' "expand" process * * - "any" is expanded first excluding streamIds in "not" * => The result is kept in `any` * - "all" is expanded in second excluding streamIds in "not" * `all` is tranformed and each "expansion" is kept in `and: [{any: ,..}]` * example: `{all: ['A', 'B']}` => `{and: [{any: [...expand('A')]}, {any: [...expand('B')]}]}` * - "not" is expanded in third and added to `and` -- !! we exclude streamIds that are in 'any' as some authorization might have been given on child now expanded * example: `{all: ['A'], not['B', 'C']}` => `{and: [{any: [...expand('A')]}, {not: [...expand('B')...expand('C')]}]} * */ function coerceStreamsParam(context: MethodContext, params: GetEventsParams, result: Result, next: ApiCallback) { if (params.streams == null) { return next(); } if (! context.acceptStreamsQueryNonStringified) { if (isStringifiedJSON(params.streams)) { try { params.streams = parseStreamsParams(params.streams); } catch (e) { return next(e); } } else if (isStringOrArrayOfStrings(params.streams)) { // good, do nothing } else { return next(errors.invalidRequestStructure('Invalid "streams" parameter. It should be an array of streamIds or JSON logical query.')) } } else { if (isStringifiedJSON(params.streams)) { try { params.streams = parseStreamsParams(params.streams); } catch (e) { return next(e); } } else { // good, do nothing } } // Transform object or string to Array if (!Array.isArray(params.streams)) { params.streams = [params.streams]; } next(); function parseStreamsParams(input: string): ?StreamQuery | ?Array<StreamQuery> { try { return JSON.parse(input); } catch (e) { throw errors.invalidRequestStructure('Invalid "streams" parameter. It should be an array of streamIds or JSON logical query. Error while parsing JSON ' + e, input); } } /** * we detect if it's JSON by looking at first char. * Note: since RFC 7159 JSON can also starts with ", true, false or number - this does not apply in this case. * @param {string} input */ function isStringifiedJSON(input: any): boolean { return (typeof input === 'string') && ['[', '{'].includes(input.substr(0, 1)); } function isStringOrArrayOfStrings(input: any): boolean { if (typeof input === 'string') return true; if (! Array.isArray(input)) return false; for (const item of input) { if (typeof item !== 'string') return false; } return true; } } async function applyDefaultsForRetrieval(context: MethodContext, params: GetEventsParams, result: Result, next: ApiCallback) { _.defaults(params, { streams: [{ any: ['*'] }], tags: null, types: null, fromTime: null, toTime: null, sortAscending: false, skip: null, limit: null, state: 'default', modifiedSince: null, includeDeletions: false }); if (params.fromTime == null && params.toTime != null) { params.fromTime = timestamp.add(params.toTime, -24 * 60 * 60); } if (params.fromTime != null && params.toTime == null) { params.toTime = timestamp.now(); } if (params.fromTime == null && params.toTime == null && params.limit == null) { // limit to 20 items by default params.limit = 20; } next(); } function tra
ntext: MethodContext, params: GetEventsParams, result: Result, next: ApiCallback) { try { params.arrayOfStreamQueries = streamsQueryUtils.transformArrayOfStringsToStreamsQuery(params.streams); } catch (e) { return next(errors.invalidRequestStructure(e, params.streams)); } next(); } function validateStreamsQueriesAndSetStore(context: MethodContext, params: GetEventsParams, result: Result, next: ApiCallback) { try { streamsQueryUtils.validateStreamsQueriesAndSetStore(params.arrayOfStreamQueries); params.arrayOfStreamQueriesWithStoreId = params.arrayOfStreamQueries; } catch (e) { return next(errors.invalidRequestStructure('Initial filtering: ' + e, params.streams)); } next(); } // the two tasks are joined as '*' replaced have their permissions checked async function streamQueryCheckPermissionsAndReplaceStars(context: MethodContext, params: GetEventsParams, result: Result, next: ApiCallback) { context.tracing.startSpan('streamQueries'); const unAuthorizedStreamIds: Array<string> = []; const unAccessibleStreamIds: Array<string> = []; async function streamExistsAndCanGetEventsOnStream(streamId: string, storeId: string, unAuthorizedStreamIds: Array<string>, unAccessibleStreamIds: Array<string>): Promise<void> { // remove eventual '#' in streamQuery const cleanStreamId: string = streamId.startsWith('#') ? streamId.substr(1) : streamId; const stream: Stream = await context.streamForStreamId(cleanStreamId, storeId); if (stream == null) { unAccessibleStreamIds.push(cleanStreamId); return; } if (! await context.access.canGetEventsOnStream(cleanStreamId, storeId)) { unAuthorizedStreamIds.push(cleanStreamId); } } for (const streamQuery: StreamQueryWithStoreId of params.arrayOfStreamQueriesWithStoreId) { // ------------ "*" case if (streamQuery.any && streamQuery.any.includes('*')) { if (await context.access.canGetEventsOnStream('*', streamQuery.storeId)) continue; // We can keep star // replace any by allowed streams for reading const canReadStreamIds: Array<string> = []; for (const streamPermission of context.access.getStoresPermissions(streamQuery.storeId)) { if (await context.access.canGetEventsOnStream(streamPermission.streamId, streamQuery.storeId)) { canReadStreamIds.push(streamPermission.streamId);
nsformArrayOfStringsToStreamsQuery(co
identifier_name
imax_logger.py
settings['run_text'] = new text_input.on_change("value", text_input_handler) textsource = ColumnDataSource(data=dict(time = [],msg = [])) columns = [ TableColumn(field="time", title="Time"), TableColumn(field="msg", title="Msg", width = 600)] data_table = DataTable(source=textsource, columns=columns, width=600) button_save = Button(label="Save Run", button_type = 'warning') def button_save_handler(): global read_data print('button save worked') run_modes= {'bat_type':settings['bat_type'], 'chrg_type':settings['chrg_type'], 'nominal_mah':settings['nominal_mah'], 'DC_or_CD':settings['DC_or_CD'], 'cycles':settings['cycles'], 'cells':settings['cells'], 'run_text':settings['run_text']} excel_out = imax_0.write_excel_file(run_modes, settings['final_read'], read_data, settings['settings_dict']) msg = 'Data saved to: ' + excel_out print(msg) text_update(datetime.datetime.now().strftime("%Y-%m-%d %H:%M"), msg) button_save.on_click(button_save_handler) def
(t, msg): global data_table print('time and msg: ', t, msg) new_data = dict(time=[t], msg=[msg],) textsource.stream(new_data, 20) #adding the value is a scrolloff lines data_table.update() import imax_0 def check_device_status(): global device_dict #used by startstop btn handler to determine if imax start btn was pressed, or imax running. #global device_dict EndPt_out = device_dict['EndPt_out'] EndPt_in = device_dict['EndPt_in'] device = device_dict['device'] #send host->imax packet to trigger imax to do imax->host transfer #make sure device is still connected if device: w_out = device.write(EndPt_out, settings['data_out_packet']) data = device.read(EndPt_in.bEndpointAddress,EndPt_in.wMaxPacketSize) settings['run_status'] = int(str(data[4])) return settings['run_status'] else: print('Check device failed.') return None def get_final_data(): global device_dict device = device_dict['device'] EndPt_out = device_dict['EndPt_out'] EndPt_in = device_dict['EndPt_in'] final_out = [0x0f, 0x03, 0xfe, 0x000, 0xfe, 0xff, 0xff] + [0]*57 w_out = device.write(EndPt_out, final_out) final_data = device.read(EndPt_in.bEndpointAddress,EndPt_in.wMaxPacketSize) if final_data: final_mah = str(final_data[5] * 256 + final_data[6]) #energy final_t = str(final_data[7] * 256 + final_data[8]) #timer sec final_V = str((final_data[9] * 256 + final_data[10]) / 1000.0) #voltage, V final_T = str(final_data[14]) #Temperature, deg C, if F??? #int temp settings['final_read'] = {'final_mah':final_mah, 'final_t':final_t, 'final_V':final_V, 'final_T':final_T} msg = 'Run completed. Final values: ' + final_mah + ' mah; '+ final_V + ' mV; ' + final_t + ' s.' print(msg) text_update(datetime.datetime.now().strftime("%Y-%m-%d %H:%M"), msg) def start_device(): global read_data global device_dict #sets up device if connected, returns imax settings, configrs, and sets data dictionary. #check for device, if not there wait for connection. device_str = imax_0.find_my_device() #returns True if device found if "No" in device_str: text_update(datetime.datetime.now().strftime("%Y-%m-%d %H:%M"), device_str) print(device_str) text_update(datetime.datetime.now().strftime("%Y-%m-%d %H:%M"), 'Imax offline. Cycling for one minute.') nowtime = datetime.datetime.now() futuretime = datetime.datetime.now() + datetime.timedelta(minutes = 1) while "No" in device_str: device_str = imax_0.find_my_device() #returns msg which has "No" in it ,if device not found if "No" in device_str: if datetime.datetime.now()> futuretime: text_update(datetime.datetime.now().strftime("%Y-%m-%d %H:%M"), 'Could not find device; check device and connection.') return False time.sleep(1) print('device found') text_update(datetime.datetime.now().strftime("%Y-%m-%d %H:%M"), device_str) #device was found, engage device, get parameters and dictionaries device_dict, read_data, settings_dict, data_out_packet = imax_0.start_imax() print('Loading settings) settings['settings_dict'] = settings_dict settings['data_out_packet'] = data_out_packet #Determine if device is idling, or already running: run_status = 2 or 3, or is running = 1 settings['run_status'] = check_device_status() nowtime = datetime.datetime.now() futuretime = datetime.datetime.now() + datetime.timedelta(minutes = 1) text_update(datetime.datetime.now().strftime("%Y-%m-%d %H:%M"), 'Waiting for Imax button press for one minute.') while settings['run_status'] > 1: settings['run_status'] = check_device_status() if datetime.datetime.now()> futuretime: text_update(datetime.datetime.now().strftime("%Y-%m-%d %H:%M"), 'Imax button not pressed in 1 min...aborting start.') return False time.sleep(1) print('out of loop run_status is: ', settings['run_status']) return True def add_lines(plot, source, cells_num = 0): #called from button_startstop_handler if bat_type LiPO, note cells must be > 1 color_list = ['orange', 'yellow', 'green', 'blue', 'violet', 'darkmagenta'] if cells_num > 1: for i in range(cells_num): p1.line(x = 'timer', y = 'cell'+ str(i+1), source = source, color = color_list[i], line_width = 2) button_startstop = Button(label = "Start", button_type = 'success') def button_startstop_handler(): global button_startstop #read btn label and isas driver for condition #label = "Start": and device_started = False is initial start up condition if button_startstop.label == "Start": button_startstop.label = "Connecting" settings['device_started'] = start_device() #returns True if device found,connected and started if settings['device_started']: text_update(datetime.datetime.now().strftime("%Y-%m-%d %H:%M"), 'Imax found & running.') button_startstop.label = "Stop" settings['start_cycle'] = curdoc().add_periodic_callback(update, 10000) print('device found') else: if not settings['device_started']: button_startstop.label = "Start" text_update(datetime.datetime.now().strftime("%Y-%m-%d %H:%M"), 'Imax start failed. Check everything') else: #deal with stop conditions; user pressed app stop button, or run_status > 1 (imax buttons pressed.) if button_startstop.label == "Stop": text_update(datetime.datetime.now().strftime("%Y-%m-%d %H:%M"), 'Run stopped.') curdoc().remove_periodic_callback(settings['start_cycle']) get_final_data() button_startstop.label = "Start" if "Li" in settings['bat_type']: add_lines(p1, source, cells_num = int(settings['cells'])) button_startstop.on_click(button_startstop_handler) def read_imax(): global read_data global device_dict # see above call to start_imax() for globals set from imax.py #global data_out_packet global out_data #global run_status device = device_dict['device'] EndPt_out = device_dict['EndPt_out'] EndPt_in = device_dict['EndPt_in'] #send the host->imax packet to trigger imax to fill buffer and do imax->host transfer #make sure device is still connected w_out = device.write(EndPt_out, settings['data_out_packet']) try: data = device.read(EndPt_in.bEndpointAddress,EndPt_in.wMaxPacketSize) #using more general form of Endpoint_IN attributes except Exception as e: print('Something went wrong: no data incoming; error
text_update
identifier_name
imax_logger.py
settings['run_text'] = new text_input.on_change("value", text_input_handler) textsource = ColumnDataSource(data=dict(time = [],msg = [])) columns = [ TableColumn(field="time", title="Time"), TableColumn(field="msg", title="Msg", width = 600)] data_table = DataTable(source=textsource, columns=columns, width=600) button_save = Button(label="Save Run", button_type = 'warning') def button_save_handler(): global read_data print('button save worked') run_modes= {'bat_type':settings['bat_type'], 'chrg_type':settings['chrg_type'], 'nominal_mah':settings['nominal_mah'], 'DC_or_CD':settings['DC_or_CD'], 'cycles':settings['cycles'], 'cells':settings['cells'], 'run_text':settings['run_text']} excel_out = imax_0.write_excel_file(run_modes, settings['final_read'], read_data, settings['settings_dict']) msg = 'Data saved to: ' + excel_out print(msg) text_update(datetime.datetime.now().strftime("%Y-%m-%d %H:%M"), msg) button_save.on_click(button_save_handler) def text_update(t, msg): global data_table print('time and msg: ', t, msg) new_data = dict(time=[t], msg=[msg],) textsource.stream(new_data, 20) #adding the value is a scrolloff lines data_table.update() import imax_0 def check_device_status(): global device_dict #used by startstop btn handler to determine if imax start btn was pressed, or imax running. #global device_dict EndPt_out = device_dict['EndPt_out'] EndPt_in = device_dict['EndPt_in'] device = device_dict['device'] #send host->imax packet to trigger imax to do imax->host transfer #make sure device is still connected if device: w_out = device.write(EndPt_out, settings['data_out_packet']) data = device.read(EndPt_in.bEndpointAddress,EndPt_in.wMaxPacketSize) settings['run_status'] = int(str(data[4])) return settings['run_status'] else: print('Check device failed.') return None def get_final_data(): global device_dict device = device_dict['device'] EndPt_out = device_dict['EndPt_out'] EndPt_in = device_dict['EndPt_in'] final_out = [0x0f, 0x03, 0xfe, 0x000, 0xfe, 0xff, 0xff] + [0]*57 w_out = device.write(EndPt_out, final_out) final_data = device.read(EndPt_in.bEndpointAddress,EndPt_in.wMaxPacketSize) if final_data: final_mah = str(final_data[5] * 256 + final_data[6]) #energy final_t = str(final_data[7] * 256 + final_data[8]) #timer sec final_V = str((final_data[9] * 256 + final_data[10]) / 1000.0) #voltage, V final_T = str(final_data[14]) #Temperature, deg C, if F??? #int temp settings['final_read'] = {'final_mah':final_mah, 'final_t':final_t, 'final_V':final_V, 'final_T':final_T} msg = 'Run completed. Final values: ' + final_mah + ' mah; '+ final_V + ' mV; ' + final_t + ' s.' print(msg) text_update(datetime.datetime.now().strftime("%Y-%m-%d %H:%M"), msg) def start_device(): global read_data global device_dict #sets up device if connected, returns imax settings, configrs, and sets data dictionary. #check for device, if not there wait for connection. device_str = imax_0.find_my_device() #returns True if device found if "No" in device_str: text_update(datetime.datetime.now().strftime("%Y-%m-%d %H:%M"), device_str) print(device_str) text_update(datetime.datetime.now().strftime("%Y-%m-%d %H:%M"), 'Imax offline. Cycling for one minute.') nowtime = datetime.datetime.now() futuretime = datetime.datetime.now() + datetime.timedelta(minutes = 1) while "No" in device_str:
print('device found') text_update(datetime.datetime.now().strftime("%Y-%m-%d %H:%M"), device_str) #device was found, engage device, get parameters and dictionaries device_dict, read_data, settings_dict, data_out_packet = imax_0.start_imax() print('Loading settings) settings['settings_dict'] = settings_dict settings['data_out_packet'] = data_out_packet #Determine if device is idling, or already running: run_status = 2 or 3, or is running = 1 settings['run_status'] = check_device_status() nowtime = datetime.datetime.now() futuretime = datetime.datetime.now() + datetime.timedelta(minutes = 1) text_update(datetime.datetime.now().strftime("%Y-%m-%d %H:%M"), 'Waiting for Imax button press for one minute.') while settings['run_status'] > 1: settings['run_status'] = check_device_status() if datetime.datetime.now()> futuretime: text_update(datetime.datetime.now().strftime("%Y-%m-%d %H:%M"), 'Imax button not pressed in 1 min...aborting start.') return False time.sleep(1) print('out of loop run_status is: ', settings['run_status']) return True def add_lines(plot, source, cells_num = 0): #called from button_startstop_handler if bat_type LiPO, note cells must be > 1 color_list = ['orange', 'yellow', 'green', 'blue', 'violet', 'darkmagenta'] if cells_num > 1: for i in range(cells_num): p1.line(x = 'timer', y = 'cell'+ str(i+1), source = source, color = color_list[i], line_width = 2) button_startstop = Button(label = "Start", button_type = 'success') def button_startstop_handler(): global button_startstop #read btn label and isas driver for condition #label = "Start": and device_started = False is initial start up condition if button_startstop.label == "Start": button_startstop.label = "Connecting" settings['device_started'] = start_device() #returns True if device found,connected and started if settings['device_started']: text_update(datetime.datetime.now().strftime("%Y-%m-%d %H:%M"), 'Imax found & running.') button_startstop.label = "Stop" settings['start_cycle'] = curdoc().add_periodic_callback(update, 10000) print('device found') else: if not settings['device_started']: button_startstop.label = "Start" text_update(datetime.datetime.now().strftime("%Y-%m-%d %H:%M"), 'Imax start failed. Check everything') else: #deal with stop conditions; user pressed app stop button, or run_status > 1 (imax buttons pressed.) if button_startstop.label == "Stop": text_update(datetime.datetime.now().strftime("%Y-%m-%d %H:%M"), 'Run stopped.') curdoc().remove_periodic_callback(settings['start_cycle']) get_final_data() button_startstop.label = "Start" if "Li" in settings['bat_type']: add_lines(p1, source, cells_num = int(settings['cells'])) button_startstop.on_click(button_startstop_handler) def read_imax(): global read_data global device_dict # see above call to start_imax() for globals set from imax.py #global data_out_packet global out_data #global run_status device = device_dict['device'] EndPt_out = device_dict['EndPt_out'] EndPt_in = device_dict['EndPt_in'] #send the host->imax packet to trigger imax to fill buffer and do imax->host transfer #make sure device is still connected w_out = device.write(EndPt_out, settings['data_out_packet']) try: data = device.read(EndPt_in.bEndpointAddress,EndPt_in.wMaxPacketSize) #using more general form of Endpoint_IN attributes except Exception as e: print('Something went wrong: no data incoming; error is
device_str = imax_0.find_my_device() #returns msg which has "No" in it ,if device not found if "No" in device_str: if datetime.datetime.now()> futuretime: text_update(datetime.datetime.now().strftime("%Y-%m-%d %H:%M"), 'Could not find device; check device and connection.') return False time.sleep(1)
conditional_block
imax_logger.py
DC_radio_group.on_click(DC_radio_handler) select_cells = Select(title = "No. of Cells", value ="4", options = [str(i) for i in range(1,13)]) def select_cells_handler(attr, old, new): settings['cells'] = new select_cells.on_change('value', select_cells_handler) #imax discharge/charge # of cycles; imax limit is 5 #no way to read this from imax only shows up in set up packet select_cycles = Select(title = "Cycles", value ="1", options = [str(i) for i in range(1,6)]) def select_cycles_handler(attr, old, new): settings['cycles'] = new select_cycles.on_change('value', select_cycles_handler) text_input = TextInput(value="Enter run information", title="Run Info::") def text_input_handler(attr, old, new): settings['run_text'] = new text_input.on_change("value", text_input_handler) textsource = ColumnDataSource(data=dict(time = [],msg = [])) columns = [ TableColumn(field="time", title="Time"), TableColumn(field="msg", title="Msg", width = 600)] data_table = DataTable(source=textsource, columns=columns, width=600) button_save = Button(label="Save Run", button_type = 'warning') def button_save_handler(): global read_data print('button save worked') run_modes= {'bat_type':settings['bat_type'], 'chrg_type':settings['chrg_type'], 'nominal_mah':settings['nominal_mah'], 'DC_or_CD':settings['DC_or_CD'], 'cycles':settings['cycles'], 'cells':settings['cells'], 'run_text':settings['run_text']} excel_out = imax_0.write_excel_file(run_modes, settings['final_read'], read_data, settings['settings_dict']) msg = 'Data saved to: ' + excel_out print(msg) text_update(datetime.datetime.now().strftime("%Y-%m-%d %H:%M"), msg) button_save.on_click(button_save_handler) def text_update(t, msg): global data_table print('time and msg: ', t, msg) new_data = dict(time=[t], msg=[msg],) textsource.stream(new_data, 20) #adding the value is a scrolloff lines data_table.update() import imax_0 def check_device_status(): global device_dict #used by startstop btn handler to determine if imax start btn was pressed, or imax running. #global device_dict EndPt_out = device_dict['EndPt_out'] EndPt_in = device_dict['EndPt_in'] device = device_dict['device'] #send host->imax packet to trigger imax to do imax->host transfer #make sure device is still connected if device: w_out = device.write(EndPt_out, settings['data_out_packet']) data = device.read(EndPt_in.bEndpointAddress,EndPt_in.wMaxPacketSize) settings['run_status'] = int(str(data[4])) return settings['run_status'] else: print('Check device failed.') return None def get_final_data(): global device_dict device = device_dict['device'] EndPt_out = device_dict['EndPt_out'] EndPt_in = device_dict['EndPt_in'] final_out = [0x0f, 0x03, 0xfe, 0x000, 0xfe, 0xff, 0xff] + [0]*57 w_out = device.write(EndPt_out, final_out) final_data = device.read(EndPt_in.bEndpointAddress,EndPt_in.wMaxPacketSize) if final_data: final_mah = str(final_data[5] * 256 + final_data[6]) #energy final_t = str(final_data[7] * 256 + final_data[8]) #timer sec final_V = str((final_data[9] * 256 + final_data[10]) / 1000.0) #voltage, V final_T = str(final_data[14]) #Temperature, deg C, if F??? #int temp settings['final_read'] = {'final_mah':final_mah, 'final_t':final_t, 'final_V':final_V, 'final_T':final_T} msg = 'Run completed. Final values: ' + final_mah + ' mah; '+ final_V + ' mV; ' + final_t + ' s.' print(msg) text_update(datetime.datetime.now().strftime("%Y-%m-%d %H:%M"), msg) def start_device(): global read_data global device_dict #sets up device if connected, returns imax settings, configrs, and sets data dictionary. #check for device, if not there wait for connection. device_str = imax_0.find_my_device() #returns True if device found if "No" in device_str: text_update(datetime.datetime.now().strftime("%Y-%m-%d %H:%M"), device_str) print(device_str) text_update(datetime.datetime.now().strftime("%Y-%m-%d %H:%M"), 'Imax offline. Cycling for one minute.') nowtime = datetime.datetime.now() futuretime = datetime.datetime.now() + datetime.timedelta(minutes = 1) while "No" in device_str: device_str = imax_0.find_my_device() #returns msg which has "No" in it ,if device not found if "No" in device_str: if datetime.datetime.now()> futuretime: text_update(datetime.datetime.now().strftime("%Y-%m-%d %H:%M"), 'Could not find device; check device and connection.') return False time.sleep(1) print('device found') text_update(datetime.datetime.now().strftime("%Y-%m-%d %H:%M"), device_str) #device was found, engage device, get parameters and dictionaries device_dict, read_data, settings_dict, data_out_packet = imax_0.start_imax() print('Loading settings) settings['settings_dict'] = settings_dict settings['data_out_packet'] = data_out_packet #Determine if device is idling, or already running: run_status = 2 or 3, or is running = 1 settings['run_status'] = check_device_status() nowtime = datetime.datetime.now() futuretime = datetime.datetime.now() + datetime.timedelta(minutes = 1) text_update(datetime.datetime.now().strftime("%Y-%m-%d %H:%M"), 'Waiting for Imax button press for one minute.') while settings['run_status'] > 1: settings['run_status'] = check_device_status() if datetime.datetime.now()> futuretime: text_update(datetime.datetime.now().strftime("%Y-%m-%d %H:%M"), 'Imax button not pressed in 1 min...aborting start.') return False time.sleep(1) print('out of loop run_status is: ', settings['run_status']) return True def add_lines(plot, source, cells_num = 0): #called from button_startstop_handler if bat_type LiPO, note cells must be > 1 color_list = ['orange', 'yellow', 'green', 'blue', 'violet', 'darkmagenta'] if cells_num > 1: for i in range(cells_num): p1.line(x = 'timer', y = 'cell'+ str(i+1), source = source, color = color_list[i], line_width = 2) button_startstop = Button(label = "Start", button_type = 'success') def button_startstop_handler(): global button_startstop #read btn label and isas driver for condition #label = "Start": and device_started = False is initial start up condition if button_startstop.label == "Start": button_startstop.label = "Connecting" settings['device_started'] = start_device() #returns True if device found,connected and started if settings['device_started']: text_update(datetime.datetime.now().strftime("%Y-%m-%d %H:%M"), 'Imax found & running.') button_startstop.label = "Stop" settings['start_cycle'] = curdoc().add_periodic_callback(update, 10000) print('device found') else: if not settings['device_started']: button_startstop.label = "Start" text_update(datetime.datetime.now().strftime("%Y-%m-%d %H:%M"), 'Imax start failed. Check everything') else: #deal with stop conditions; user pressed app stop button, or run_status > 1 (imax buttons pressed.) if button_startstop.label == "Stop": text_update(datetime.datetime.now().strftime("%Y-%m-%d %H:%M"), 'Run stopped.') curdoc().remove_periodic_callback(settings['start_cycle']) get_final_data() button_startstop.label = "Start" if "Li" in settings['bat_type']: add_lines(p1, source, cells_num = int(settings['cells']))
settings['DC'] = new
identifier_body
imax_logger.py
#Create the header for page notice1 = Div(text="""The data input here are for logging identification and saving conditions, which are already manually chosen on the imax. They do not set or reset the imax.""", sizing_mode = "scale_width") select_battype = Select(title = "Battery Type", value ="", options = ['NiMH', 'NiCd', 'LiPO', 'LiFe', 'LiIO', 'LiHV']) def select_battype_handler(attr, old, new): settings['bat_type'] = new select_battype.on_change('value', select_battype_handler) select_chrg_type = Select(title="Charge Type", value="Charge", options=["Charge", "Discharge", "Cycle", "Re-Peak", "AutoCharge", "Balance Charge", "Fast Charge", "Storage"]) def select_chrg_type_handler(attr, old, new): settings['chrg_type'] = new select_chrg_type.on_change('value', select_chrg_type_handler) maxmah_slider = Slider(start=50, end=24000, value=1, step=50, title="Bat. Specified Capacity, mah") def maxmah_handler(attr, old, new): settings['nominal_mah'] = new maxmah_slider.on_change('value', maxmah_handler) DC_radio_group = RadioGroup(labels=["Discharge/Charge", "Charge/Discharge"], active=0) def DC_radio_handler(new): settings['DC'] = new DC_radio_group.on_click(DC_radio_handler) select_cells = Select(title = "No. of Cells", value ="4", options = [str(i) for i in range(1,13)]) def select_cells_handler(attr, old, new): settings['cells'] = new select_cells.on_change('value', select_cells_handler) #imax discharge/charge # of cycles; imax limit is 5 #no way to read this from imax only shows up in set up packet select_cycles = Select(title = "Cycles", value ="1", options = [str(i) for i in range(1,6)]) def select_cycles_handler(attr, old, new): settings['cycles'] = new select_cycles.on_change('value', select_cycles_handler) text_input = TextInput(value="Enter run information", title="Run Info::") def text_input_handler(attr, old, new): settings['run_text'] = new text_input.on_change("value", text_input_handler) textsource = ColumnDataSource(data=dict(time = [],msg = [])) columns = [ TableColumn(field="time", title="Time"), TableColumn(field="msg", title="Msg", width = 600)] data_table = DataTable(source=textsource, columns=columns, width=600) button_save = Button(label="Save Run", button_type = 'warning') def button_save_handler(): global read_data print('button save worked') run_modes= {'bat_type':settings['bat_type'], 'chrg_type':settings['chrg_type'], 'nominal_mah':settings['nominal_mah'], 'DC_or_CD':settings['DC_or_CD'], 'cycles':settings['cycles'], 'cells':settings['cells'], 'run_text':settings['run_text']} excel_out = imax_0.write_excel_file(run_modes, settings['final_read'], read_data, settings['settings_dict']) msg = 'Data saved to: ' + excel_out print(msg) text_update(datetime.datetime.now().strftime("%Y-%m-%d %H:%M"), msg) button_save.on_click(button_save_handler) def text_update(t, msg): global data_table print('time and msg: ', t, msg) new_data = dict(time=[t], msg=[msg],) textsource.stream(new_data, 20) #adding the value is a scrolloff lines data_table.update() import imax_0 def check_device_status(): global device_dict #used by startstop btn handler to determine if imax start btn was pressed, or imax running. #global device_dict EndPt_out = device_dict['EndPt_out'] EndPt_in = device_dict['EndPt_in'] device = device_dict['device'] #send host->imax packet to trigger imax to do imax->host transfer #make sure device is still connected if device: w_out = device.write(EndPt_out, settings['data_out_packet']) data = device.read(EndPt_in.bEndpointAddress,EndPt_in.wMaxPacketSize) settings['run_status'] = int(str(data[4])) return settings['run_status'] else: print('Check device failed.') return None def get_final_data(): global device_dict device = device_dict['device'] EndPt_out = device_dict['EndPt_out'] EndPt_in = device_dict['EndPt_in'] final_out = [0x0f, 0x03, 0xfe, 0x000, 0xfe, 0xff, 0xff] + [0]*57 w_out = device.write(EndPt_out, final_out) final_data = device.read(EndPt_in.bEndpointAddress,EndPt_in.wMaxPacketSize) if final_data: final_mah = str(final_data[5] * 256 + final_data[6]) #energy final_t = str(final_data[7] * 256 + final_data[8]) #timer sec final_V = str((final_data[9] * 256 + final_data[10]) / 1000.0) #voltage, V final_T = str(final_data[14]) #Temperature, deg C, if F??? #int temp settings['final_read'] = {'final_mah':final_mah, 'final_t':final_t, 'final_V':final_V, 'final_T':final_T} msg = 'Run completed. Final values: ' + final_mah + ' mah; '+ final_V + ' mV; ' + final_t + ' s.' print(msg) text_update(datetime.datetime.now().strftime("%Y-%m-%d %H:%M"), msg) def start_device(): global read_data global device_dict #sets up device if connected, returns imax settings, configrs, and sets data dictionary. #check for device, if not there wait for connection. device_str = imax_0.find_my_device() #returns True if device found if "No" in device_str: text_update(datetime.datetime.now().strftime("%Y-%m-%d %H:%M"), device_str) print(device_str) text_update(datetime.datetime.now().strftime("%Y-%m-%d %H:%M"), 'Imax offline. Cycling for one minute.') nowtime = datetime.datetime.now() futuretime = datetime.datetime.now() + datetime.timedelta(minutes = 1) while "No" in device_str: device_str = imax_0.find_my_device() #returns msg which has "No" in it ,if device not found if "No" in device_str: if datetime.datetime.now()> futuretime: text_update(datetime.datetime.now().strftime("%Y-%m-%d %H:%M"), 'Could not find device; check device and connection.') return False time.sleep(1) print('device found') text_update(datetime.datetime.now().strftime("%Y-%m-%d %H:%M"), device_str) #device was found, engage device, get parameters and dictionaries device_dict, read_data, settings_dict, data_out_packet = imax_0.start_imax() print('Loading settings) settings['settings_dict'] = settings_dict settings['data_out_packet'] = data_out_packet #Determine if device is idling, or already running: run_status = 2 or 3, or is running = 1 settings['run_status'] = check_device_status() nowtime = datetime.datetime.now() futuretime = datetime.datetime.now() + datetime.timedelta(minutes = 1) text_update(datetime.datetime.now().strftime("%Y-%m-%d %H:%M"), 'Waiting for Imax button press for one minute.') while settings['run_status'] > 1: settings['run_status'] = check_device_status() if datetime.datetime.now()> futuretime: text_update(datetime.datetime.now().strftime("%Y-%m-%d %H:%M"), 'Imax button not pressed in 1 min...aborting start.') return False time.sleep(1) print('out of loop run_status is: ', settings['run_status']) return True def add_lines(plot, source, cells_num = 0): #called from button_startstop_handler if bat_type LiPO, note cells must be > 1 color_list = ['orange', 'yellow', 'green', 'blue', 'violet', 'darkmagenta'] if cells_num > 1: for i in range(cells_num): p1.line(x = 'timer', y = 'cell'+ str(i+1), source = source, color = color_list[i], line_width = 2) button_startstop = Button(label
'settings_dict':settings_dict, 'device_dict':device_dict }
random_line_split
network.py
:param filter_shape: shape of the convolution filter. :param stride: stride for the convolution :param padding: padding mode, either 'VALID' or 'SAME' :return: shape of the output tensor as a plain list of integers """ filter_shape = tf.TensorShape(filter_shape).as_list() filter_out = filter_shape[-1] filter_patch_shape = np.array(filter_shape[0:2]) input_shape_list = tf.TensorShape(input_shape).as_list() batch = input_shape_list[:-3] input_shape = np.array(input_shape_list[-3:]) stride = np.array(stride) if padding == 'VALID': shift = -filter_patch_shape + 1 elif padding == 'SAME': shift = 0 else: raise ValueError('padding must be either "VALID" or "SAME", but "%s" was given' % padding) output_shape = np.ceil((input_shape[:2] + shift) / stride[1:3]) return batch + output_shape.astype(np.int).tolist() + [filter_out] def conv2d_config(input_shape, output_shape, filter_shape): """ Based on the desired input, output and filter shape, figure out the correct 2D convolution configuration to use including the type (normal or full convolution), stride size, padding type/size :param input_shape: :param output_shape: :param filter_shape: :return: """ input_shape = tf.TensorShape(input_shape).as_list() if len(input_shape) == 4: batch_size = input_shape[0] else: batch_size = None input_shape = np.array(input_shape[-3:]) output_shape = np.array(tf.TensorShape(output_shape).as_list()[-3:]) # Determine what kind of convolution to use if np.all(input_shape[-3:-1] >= output_shape[-3:-1]): conv_type = "NORMAL" elif np.all(input_shape[-3:-1] <= output_shape[-3:-1]): conv_type = 'FULL' # swap input and output shape input_shape, output_shape = output_shape, input_shape else: raise ValueError('Input shape dimensions must be both bigger than or both smaller than output shape dimensions') filter_shape = np.array(tf.TensorShape(filter_shape).as_list()[:2] + [input_shape[-1], output_shape[-1]]) stride = np.ceil((input_shape[:2] - filter_shape[:2] + 1) / output_shape[:2]).astype(np.int) padding = output_shape[:2] * stride - input_shape[:2] + filter_shape[:2] - 1 # Determine what type of padding can be used if np.all(np.ceil(input_shape[:2] / stride) == output_shape[:2]): padding_type = 'SAME' else: padding_type = 'VALID' # get padded input shape input_shape[:2] = input_shape[:2] + padding.astype(np.int) padded_shape = [batch_size] + input_shape.tolist() left_padding = np.ceil(padding / 2).astype(np.int) right_padding = np.floor(padding / 2).astype(np.int) padding = [[0, 0], [left_padding[0], right_padding[0]], [left_padding[1], right_padding[1]], [0, 0]] stride = [1, stride[0], stride[1], 1] return filter_shape.tolist(), stride, padding, padded_shape, conv_type, padding_type def get_convolution_op(input_shape, output_shape, kernel_shape): """ Given the desired shapes of the input, output and filter tensors, returns the shape of the appropriate convolution filter and a correctly configured op function. The returned op function should be called with the input tensor and weight tensor, and returns a result of 2D convolution that matches the desired output_shape :param input_shape: desired input shape into the convolution operation :param output_shape: desired output shape from the convolution operation :param kernel_shape: desired convolution kernel shape. Only the first two diemensions (height and width) will be used. :return: (filter_shape, conv_op) The shape of the appropriate convolution filter/weight to be used (filter_shape) and a function that can be invoked with inputs tensor and correctly sized filter tensor to define the convolution operation. """ filter_shape, strides, padding, padded_shape, conv_type, padding_type = conv2d_config(input_shape, output_shape, kernel_shape) if conv_type == 'NORMAL':
else: def conv_op(inputs, weight, name='generic_convolution'): if padding_type=='SAME': padded_output = [padded_shape[0]] + output_shape[-3:] else: padded_output = padded_shape with tf.name_scope(name): if padded_output[0] is None: batch_size = tf.shape(inputs)[0] padded_output = [batch_size] + padded_output[1:] output = tf.nn.conv2d_transpose(inputs, weight, padded_output, strides, padding_type, name='transpose_convolution') if padding_type=='VALID' and np.sum(padding) > 0: output = tf.slice(output, [0, padding[1][0], padding[2][0], 0], [-1] + output_shape[-3:], name='cropping') return output return filter_shape, conv_op def normalize_weights(w, dims=(0,), bias=1e-5): """ L2 normalize weights of the given tensor along specified dimension(s). Args: w: Tensor to be normalized dims: dimension(s) along which to normalize the Tensor. Defaults to (0,) bias: Bias value added to the computed norm to prevent dividing by 0. Defaults to 1e-5 Returns: Tensor of same type and shape as `w` whose norm is set to approximately 1 along the specificed dimensions """ with tf.name_scope('normalization'): return w / (tf.sqrt(tf.reduce_sum(tf.square(w), dims, keep_dims=True) + bias)) def weight_variable(shape, name='weight', mean=0.0, stddev=None, initializer=None, constrain=None, dtype=tf.float32): """ Creates and returns a variable initialized with random_normal_initializer, suitable for use as a weight. In the current variable scope, creates (if necessary) and returns a named variable with `tf.random_normal_initializer`. Args: shape: Required. Shape of the variable name: Optional. Name of the variable. Defaults to 'weight' mean: Optional. Mean of the `random_normal_initializer`. Defaults to 0.0 stddev: Optional. Standard deviation of the `random_normal_initializer`. Defaults to 1e-3 dtype: Optional. Data type of the variable. Default to `tf.float32`. Returns: Weight variable with specified name and shape with random normal initialization. """ if stddev is None: raise ValueError('stddev not specified!') if initializer is None: initializer = tf.random_normal_initializer(mean=mean, stddev=stddev) weights = tf.get_variable(name, shape=shape, initializer=initializer, dtype=dtype) if constrain is not None: constrain(weights) return weights def bias_variable(shape, name='bias', value=0.0, initializer=None, constrain=None, dtype=tf.float32): """ Creates and returns a variable initialized with random_normal_initializer, suitable for use as a bias. In the current variable scope, creates (if necessary) and returns a named variable with `tf.random_normal_initializer`. Args: shape: Required. Shape of the variable name: Optional. Name of the variable. Defaults to 'bias' value: Optional. Constant value to which the variable is initialized. Defaults to 0.0 dtype: Optional. Data type of the variable. Default to `tf.float32`. Returns: Bias variable with specified name and shape initialized to a constant. """ if initializer is None: initializer = tf.constant_initializer(value=value) biases = tf.get_variable(name, shape=shape, initializer=initializer, dtype=dtype) if constrain is not None: constrain(biases) return biases def factorized_readout(inputs, n_outputs=100, constrain=True): width, height, n_features = inputs.get_shape()[1:].as_list() n_pixels = width * height with tf.variable_scope('readout'): # spatial readout w_spatial = weight_variable([n_pixels, 1, n_outputs], name='weight_spatial') if constrain: constraints.positive_constrain(w_spatial) w_spatial_norm = normalize_weights(w_spatial, dims=(0,)) # feature readout w_feature = weight_variable([1, n_features, n_outputs], name='weight_feature') if constrain: constraints.positive_constrain(w_feature) w_feature_norm = normalize_weights(w_feature, dims=(1,)) # scaling w_scale = bias_variable([n_outputs], name='weight_scale', value=1
def conv_op(inputs, weight, name='generic_convolution'): with tf.name_scope(name): if padding_type=='VALID' and np.sum(padding) > 0: inputs = tf.pad(inputs, padding, name='padding') return tf.nn.conv2d(inputs, weight, strides, padding_type, name='convolution')
conditional_block
network.py
:param filter_shape: shape of the convolution filter. :param stride: stride for the convolution :param padding: padding mode, either 'VALID' or 'SAME' :return: shape of the output tensor as a plain list of integers """ filter_shape = tf.TensorShape(filter_shape).as_list() filter_out = filter_shape[-1] filter_patch_shape = np.array(filter_shape[0:2]) input_shape_list = tf.TensorShape(input_shape).as_list() batch = input_shape_list[:-3] input_shape = np.array(input_shape_list[-3:]) stride = np.array(stride) if padding == 'VALID': shift = -filter_patch_shape + 1 elif padding == 'SAME': shift = 0 else: raise ValueError('padding must be either "VALID" or "SAME", but "%s" was given' % padding) output_shape = np.ceil((input_shape[:2] + shift) / stride[1:3]) return batch + output_shape.astype(np.int).tolist() + [filter_out] def
(input_shape, output_shape, filter_shape): """ Based on the desired input, output and filter shape, figure out the correct 2D convolution configuration to use including the type (normal or full convolution), stride size, padding type/size :param input_shape: :param output_shape: :param filter_shape: :return: """ input_shape = tf.TensorShape(input_shape).as_list() if len(input_shape) == 4: batch_size = input_shape[0] else: batch_size = None input_shape = np.array(input_shape[-3:]) output_shape = np.array(tf.TensorShape(output_shape).as_list()[-3:]) # Determine what kind of convolution to use if np.all(input_shape[-3:-1] >= output_shape[-3:-1]): conv_type = "NORMAL" elif np.all(input_shape[-3:-1] <= output_shape[-3:-1]): conv_type = 'FULL' # swap input and output shape input_shape, output_shape = output_shape, input_shape else: raise ValueError('Input shape dimensions must be both bigger than or both smaller than output shape dimensions') filter_shape = np.array(tf.TensorShape(filter_shape).as_list()[:2] + [input_shape[-1], output_shape[-1]]) stride = np.ceil((input_shape[:2] - filter_shape[:2] + 1) / output_shape[:2]).astype(np.int) padding = output_shape[:2] * stride - input_shape[:2] + filter_shape[:2] - 1 # Determine what type of padding can be used if np.all(np.ceil(input_shape[:2] / stride) == output_shape[:2]): padding_type = 'SAME' else: padding_type = 'VALID' # get padded input shape input_shape[:2] = input_shape[:2] + padding.astype(np.int) padded_shape = [batch_size] + input_shape.tolist() left_padding = np.ceil(padding / 2).astype(np.int) right_padding = np.floor(padding / 2).astype(np.int) padding = [[0, 0], [left_padding[0], right_padding[0]], [left_padding[1], right_padding[1]], [0, 0]] stride = [1, stride[0], stride[1], 1] return filter_shape.tolist(), stride, padding, padded_shape, conv_type, padding_type def get_convolution_op(input_shape, output_shape, kernel_shape): """ Given the desired shapes of the input, output and filter tensors, returns the shape of the appropriate convolution filter and a correctly configured op function. The returned op function should be called with the input tensor and weight tensor, and returns a result of 2D convolution that matches the desired output_shape :param input_shape: desired input shape into the convolution operation :param output_shape: desired output shape from the convolution operation :param kernel_shape: desired convolution kernel shape. Only the first two diemensions (height and width) will be used. :return: (filter_shape, conv_op) The shape of the appropriate convolution filter/weight to be used (filter_shape) and a function that can be invoked with inputs tensor and correctly sized filter tensor to define the convolution operation. """ filter_shape, strides, padding, padded_shape, conv_type, padding_type = conv2d_config(input_shape, output_shape, kernel_shape) if conv_type == 'NORMAL': def conv_op(inputs, weight, name='generic_convolution'): with tf.name_scope(name): if padding_type=='VALID' and np.sum(padding) > 0: inputs = tf.pad(inputs, padding, name='padding') return tf.nn.conv2d(inputs, weight, strides, padding_type, name='convolution') else: def conv_op(inputs, weight, name='generic_convolution'): if padding_type=='SAME': padded_output = [padded_shape[0]] + output_shape[-3:] else: padded_output = padded_shape with tf.name_scope(name): if padded_output[0] is None: batch_size = tf.shape(inputs)[0] padded_output = [batch_size] + padded_output[1:] output = tf.nn.conv2d_transpose(inputs, weight, padded_output, strides, padding_type, name='transpose_convolution') if padding_type=='VALID' and np.sum(padding) > 0: output = tf.slice(output, [0, padding[1][0], padding[2][0], 0], [-1] + output_shape[-3:], name='cropping') return output return filter_shape, conv_op def normalize_weights(w, dims=(0,), bias=1e-5): """ L2 normalize weights of the given tensor along specified dimension(s). Args: w: Tensor to be normalized dims: dimension(s) along which to normalize the Tensor. Defaults to (0,) bias: Bias value added to the computed norm to prevent dividing by 0. Defaults to 1e-5 Returns: Tensor of same type and shape as `w` whose norm is set to approximately 1 along the specificed dimensions """ with tf.name_scope('normalization'): return w / (tf.sqrt(tf.reduce_sum(tf.square(w), dims, keep_dims=True) + bias)) def weight_variable(shape, name='weight', mean=0.0, stddev=None, initializer=None, constrain=None, dtype=tf.float32): """ Creates and returns a variable initialized with random_normal_initializer, suitable for use as a weight. In the current variable scope, creates (if necessary) and returns a named variable with `tf.random_normal_initializer`. Args: shape: Required. Shape of the variable name: Optional. Name of the variable. Defaults to 'weight' mean: Optional. Mean of the `random_normal_initializer`. Defaults to 0.0 stddev: Optional. Standard deviation of the `random_normal_initializer`. Defaults to 1e-3 dtype: Optional. Data type of the variable. Default to `tf.float32`. Returns: Weight variable with specified name and shape with random normal initialization. """ if stddev is None: raise ValueError('stddev not specified!') if initializer is None: initializer = tf.random_normal_initializer(mean=mean, stddev=stddev) weights = tf.get_variable(name, shape=shape, initializer=initializer, dtype=dtype) if constrain is not None: constrain(weights) return weights def bias_variable(shape, name='bias', value=0.0, initializer=None, constrain=None, dtype=tf.float32): """ Creates and returns a variable initialized with random_normal_initializer, suitable for use as a bias. In the current variable scope, creates (if necessary) and returns a named variable with `tf.random_normal_initializer`. Args: shape: Required. Shape of the variable name: Optional. Name of the variable. Defaults to 'bias' value: Optional. Constant value to which the variable is initialized. Defaults to 0.0 dtype: Optional. Data type of the variable. Default to `tf.float32`. Returns: Bias variable with specified name and shape initialized to a constant. """ if initializer is None: initializer = tf.constant_initializer(value=value) biases = tf.get_variable(name, shape=shape, initializer=initializer, dtype=dtype) if constrain is not None: constrain(biases) return biases def factorized_readout(inputs, n_outputs=100, constrain=True): width, height, n_features = inputs.get_shape()[1:].as_list() n_pixels = width * height with tf.variable_scope('readout'): # spatial readout w_spatial = weight_variable([n_pixels, 1, n_outputs], name='weight_spatial') if constrain: constraints.positive_constrain(w_spatial) w_spatial_norm = normalize_weights(w_spatial, dims=(0,)) # feature readout w_feature = weight_variable([1, n_features, n_outputs], name='weight_feature') if constrain: constraints.positive_constrain(w_feature) w_feature_norm = normalize_weights(w_feature, dims=(1,)) # scaling w_scale = bias_variable([n_outputs], name='weight_scale', value=1
conv2d_config
identifier_name
network.py
[1:3]) return batch + output_shape.astype(np.int).tolist() + [filter_out] def conv2d_config(input_shape, output_shape, filter_shape): """ Based on the desired input, output and filter shape, figure out the correct 2D convolution configuration to use including the type (normal or full convolution), stride size, padding type/size :param input_shape: :param output_shape: :param filter_shape: :return: """ input_shape = tf.TensorShape(input_shape).as_list() if len(input_shape) == 4: batch_size = input_shape[0] else: batch_size = None input_shape = np.array(input_shape[-3:]) output_shape = np.array(tf.TensorShape(output_shape).as_list()[-3:]) # Determine what kind of convolution to use if np.all(input_shape[-3:-1] >= output_shape[-3:-1]): conv_type = "NORMAL" elif np.all(input_shape[-3:-1] <= output_shape[-3:-1]): conv_type = 'FULL' # swap input and output shape input_shape, output_shape = output_shape, input_shape else: raise ValueError('Input shape dimensions must be both bigger than or both smaller than output shape dimensions') filter_shape = np.array(tf.TensorShape(filter_shape).as_list()[:2] + [input_shape[-1], output_shape[-1]]) stride = np.ceil((input_shape[:2] - filter_shape[:2] + 1) / output_shape[:2]).astype(np.int) padding = output_shape[:2] * stride - input_shape[:2] + filter_shape[:2] - 1 # Determine what type of padding can be used if np.all(np.ceil(input_shape[:2] / stride) == output_shape[:2]): padding_type = 'SAME' else: padding_type = 'VALID' # get padded input shape input_shape[:2] = input_shape[:2] + padding.astype(np.int) padded_shape = [batch_size] + input_shape.tolist() left_padding = np.ceil(padding / 2).astype(np.int) right_padding = np.floor(padding / 2).astype(np.int) padding = [[0, 0], [left_padding[0], right_padding[0]], [left_padding[1], right_padding[1]], [0, 0]] stride = [1, stride[0], stride[1], 1] return filter_shape.tolist(), stride, padding, padded_shape, conv_type, padding_type def get_convolution_op(input_shape, output_shape, kernel_shape): """ Given the desired shapes of the input, output and filter tensors, returns the shape of the appropriate convolution filter and a correctly configured op function. The returned op function should be called with the input tensor and weight tensor, and returns a result of 2D convolution that matches the desired output_shape :param input_shape: desired input shape into the convolution operation :param output_shape: desired output shape from the convolution operation :param kernel_shape: desired convolution kernel shape. Only the first two diemensions (height and width) will be used. :return: (filter_shape, conv_op) The shape of the appropriate convolution filter/weight to be used (filter_shape) and a function that can be invoked with inputs tensor and correctly sized filter tensor to define the convolution operation. """ filter_shape, strides, padding, padded_shape, conv_type, padding_type = conv2d_config(input_shape, output_shape, kernel_shape) if conv_type == 'NORMAL': def conv_op(inputs, weight, name='generic_convolution'): with tf.name_scope(name): if padding_type=='VALID' and np.sum(padding) > 0: inputs = tf.pad(inputs, padding, name='padding') return tf.nn.conv2d(inputs, weight, strides, padding_type, name='convolution') else: def conv_op(inputs, weight, name='generic_convolution'): if padding_type=='SAME': padded_output = [padded_shape[0]] + output_shape[-3:] else: padded_output = padded_shape with tf.name_scope(name): if padded_output[0] is None: batch_size = tf.shape(inputs)[0] padded_output = [batch_size] + padded_output[1:] output = tf.nn.conv2d_transpose(inputs, weight, padded_output, strides, padding_type, name='transpose_convolution') if padding_type=='VALID' and np.sum(padding) > 0: output = tf.slice(output, [0, padding[1][0], padding[2][0], 0], [-1] + output_shape[-3:], name='cropping') return output return filter_shape, conv_op def normalize_weights(w, dims=(0,), bias=1e-5): """ L2 normalize weights of the given tensor along specified dimension(s). Args: w: Tensor to be normalized dims: dimension(s) along which to normalize the Tensor. Defaults to (0,) bias: Bias value added to the computed norm to prevent dividing by 0. Defaults to 1e-5 Returns: Tensor of same type and shape as `w` whose norm is set to approximately 1 along the specificed dimensions """ with tf.name_scope('normalization'): return w / (tf.sqrt(tf.reduce_sum(tf.square(w), dims, keep_dims=True) + bias)) def weight_variable(shape, name='weight', mean=0.0, stddev=None, initializer=None, constrain=None, dtype=tf.float32): """ Creates and returns a variable initialized with random_normal_initializer, suitable for use as a weight. In the current variable scope, creates (if necessary) and returns a named variable with `tf.random_normal_initializer`. Args: shape: Required. Shape of the variable name: Optional. Name of the variable. Defaults to 'weight' mean: Optional. Mean of the `random_normal_initializer`. Defaults to 0.0 stddev: Optional. Standard deviation of the `random_normal_initializer`. Defaults to 1e-3 dtype: Optional. Data type of the variable. Default to `tf.float32`. Returns: Weight variable with specified name and shape with random normal initialization. """ if stddev is None: raise ValueError('stddev not specified!') if initializer is None: initializer = tf.random_normal_initializer(mean=mean, stddev=stddev) weights = tf.get_variable(name, shape=shape, initializer=initializer, dtype=dtype) if constrain is not None: constrain(weights) return weights def bias_variable(shape, name='bias', value=0.0, initializer=None, constrain=None, dtype=tf.float32): """ Creates and returns a variable initialized with random_normal_initializer, suitable for use as a bias. In the current variable scope, creates (if necessary) and returns a named variable with `tf.random_normal_initializer`. Args: shape: Required. Shape of the variable name: Optional. Name of the variable. Defaults to 'bias' value: Optional. Constant value to which the variable is initialized. Defaults to 0.0 dtype: Optional. Data type of the variable. Default to `tf.float32`. Returns: Bias variable with specified name and shape initialized to a constant. """ if initializer is None: initializer = tf.constant_initializer(value=value) biases = tf.get_variable(name, shape=shape, initializer=initializer, dtype=dtype) if constrain is not None: constrain(biases) return biases def factorized_readout(inputs, n_outputs=100, constrain=True): width, height, n_features = inputs.get_shape()[1:].as_list() n_pixels = width * height with tf.variable_scope('readout'): # spatial readout w_spatial = weight_variable([n_pixels, 1, n_outputs], name='weight_spatial') if constrain: constraints.positive_constrain(w_spatial) w_spatial_norm = normalize_weights(w_spatial, dims=(0,)) # feature readout w_feature = weight_variable([1, n_features, n_outputs], name='weight_feature') if constrain: constraints.positive_constrain(w_feature) w_feature_norm = normalize_weights(w_feature, dims=(1,)) # scaling w_scale = bias_variable([n_outputs], name='weight_scale', value=1.0) if constrain: constraints.positive_constrain(w_scale) # total readout weight w_out = tf.reshape(w_spatial_norm * w_feature_norm * w_scale, [n_pixels * n_features, n_outputs], 'weight_readout') output = tf.matmul(tf.reshape(inputs, [-1, n_pixels * n_features]), w_out) return output, w_spatial_norm, w_feature_norm, w_scale, w_out def batch_norm(inputs, *args, tag=None, add_summary=True, step=0, **kwargs):
if step > 0 and 'updates_collections' not in kwargs: kwargs['updates_collections'] = 'dump' output = layers.batch_norm(inputs, *args, **kwargs) if add_summary: if tag is None: tag = inputs.op.name.split('/')[-1] tag = 'batch_norm/' + tag tf.histogram_summary(tag, inputs) tf.histogram_summary(tag + '_bn', output) return output
identifier_body
network.py
:param filter_shape: shape of the convolution filter. :param stride: stride for the convolution :param padding: padding mode, either 'VALID' or 'SAME' :return: shape of the output tensor as a plain list of integers """ filter_shape = tf.TensorShape(filter_shape).as_list() filter_out = filter_shape[-1] filter_patch_shape = np.array(filter_shape[0:2]) input_shape_list = tf.TensorShape(input_shape).as_list() batch = input_shape_list[:-3] input_shape = np.array(input_shape_list[-3:]) stride = np.array(stride) if padding == 'VALID': shift = -filter_patch_shape + 1 elif padding == 'SAME': shift = 0 else: raise ValueError('padding must be either "VALID" or "SAME", but "%s" was given' % padding) output_shape = np.ceil((input_shape[:2] + shift) / stride[1:3]) return batch + output_shape.astype(np.int).tolist() + [filter_out] def conv2d_config(input_shape, output_shape, filter_shape): """ Based on the desired input, output and filter shape, figure out the correct 2D convolution configuration to use including the type (normal or full convolution), stride size, padding type/size :param input_shape: :param output_shape: :param filter_shape: :return: """ input_shape = tf.TensorShape(input_shape).as_list() if len(input_shape) == 4: batch_size = input_shape[0] else: batch_size = None input_shape = np.array(input_shape[-3:]) output_shape = np.array(tf.TensorShape(output_shape).as_list()[-3:]) # Determine what kind of convolution to use if np.all(input_shape[-3:-1] >= output_shape[-3:-1]): conv_type = "NORMAL" elif np.all(input_shape[-3:-1] <= output_shape[-3:-1]): conv_type = 'FULL' # swap input and output shape input_shape, output_shape = output_shape, input_shape else: raise ValueError('Input shape dimensions must be both bigger than or both smaller than output shape dimensions') filter_shape = np.array(tf.TensorShape(filter_shape).as_list()[:2] + [input_shape[-1], output_shape[-1]]) stride = np.ceil((input_shape[:2] - filter_shape[:2] + 1) / output_shape[:2]).astype(np.int) padding = output_shape[:2] * stride - input_shape[:2] + filter_shape[:2] - 1 # Determine what type of padding can be used if np.all(np.ceil(input_shape[:2] / stride) == output_shape[:2]): padding_type = 'SAME' else: padding_type = 'VALID' # get padded input shape input_shape[:2] = input_shape[:2] + padding.astype(np.int) padded_shape = [batch_size] + input_shape.tolist() left_padding = np.ceil(padding / 2).astype(np.int) right_padding = np.floor(padding / 2).astype(np.int) padding = [[0, 0], [left_padding[0], right_padding[0]], [left_padding[1], right_padding[1]], [0, 0]] stride = [1, stride[0], stride[1], 1] return filter_shape.tolist(), stride, padding, padded_shape, conv_type, padding_type def get_convolution_op(input_shape, output_shape, kernel_shape): """ Given the desired shapes of the input, output and filter tensors, returns the shape of the appropriate convolution filter and a correctly configured op function. The returned op function should be called with the input tensor and weight tensor, and returns a result of 2D convolution that matches the desired output_shape :param input_shape: desired input shape into the convolution operation :param output_shape: desired output shape from the convolution operation :param kernel_shape: desired convolution kernel shape. Only the first two diemensions (height and width) will be used. :return: (filter_shape, conv_op) The shape of the appropriate convolution filter/weight to be used (filter_shape) and a function that can be invoked with inputs tensor and correctly sized filter tensor to define the convolution operation. """ filter_shape, strides, padding, padded_shape, conv_type, padding_type = conv2d_config(input_shape, output_shape, kernel_shape) if conv_type == 'NORMAL': def conv_op(inputs, weight, name='generic_convolution'): with tf.name_scope(name): if padding_type=='VALID' and np.sum(padding) > 0: inputs = tf.pad(inputs, padding, name='padding') return tf.nn.conv2d(inputs, weight, strides, padding_type, name='convolution') else: def conv_op(inputs, weight, name='generic_convolution'): if padding_type=='SAME': padded_output = [padded_shape[0]] + output_shape[-3:] else: padded_output = padded_shape with tf.name_scope(name): if padded_output[0] is None:
output = tf.slice(output, [0, padding[1][0], padding[2][0], 0], [-1] + output_shape[-3:], name='cropping') return output return filter_shape, conv_op def normalize_weights(w, dims=(0,), bias=1e-5): """ L2 normalize weights of the given tensor along specified dimension(s). Args: w: Tensor to be normalized dims: dimension(s) along which to normalize the Tensor. Defaults to (0,) bias: Bias value added to the computed norm to prevent dividing by 0. Defaults to 1e-5 Returns: Tensor of same type and shape as `w` whose norm is set to approximately 1 along the specificed dimensions """ with tf.name_scope('normalization'): return w / (tf.sqrt(tf.reduce_sum(tf.square(w), dims, keep_dims=True) + bias)) def weight_variable(shape, name='weight', mean=0.0, stddev=None, initializer=None, constrain=None, dtype=tf.float32): """ Creates and returns a variable initialized with random_normal_initializer, suitable for use as a weight. In the current variable scope, creates (if necessary) and returns a named variable with `tf.random_normal_initializer`. Args: shape: Required. Shape of the variable name: Optional. Name of the variable. Defaults to 'weight' mean: Optional. Mean of the `random_normal_initializer`. Defaults to 0.0 stddev: Optional. Standard deviation of the `random_normal_initializer`. Defaults to 1e-3 dtype: Optional. Data type of the variable. Default to `tf.float32`. Returns: Weight variable with specified name and shape with random normal initialization. """ if stddev is None: raise ValueError('stddev not specified!') if initializer is None: initializer = tf.random_normal_initializer(mean=mean, stddev=stddev) weights = tf.get_variable(name, shape=shape, initializer=initializer, dtype=dtype) if constrain is not None: constrain(weights) return weights def bias_variable(shape, name='bias', value=0.0, initializer=None, constrain=None, dtype=tf.float32): """ Creates and returns a variable initialized with random_normal_initializer, suitable for use as a bias. In the current variable scope, creates (if necessary) and returns a named variable with `tf.random_normal_initializer`. Args: shape: Required. Shape of the variable name: Optional. Name of the variable. Defaults to 'bias' value: Optional. Constant value to which the variable is initialized. Defaults to 0.0 dtype: Optional. Data type of the variable. Default to `tf.float32`. Returns: Bias variable with specified name and shape initialized to a constant. """ if initializer is None: initializer = tf.constant_initializer(value=value) biases = tf.get_variable(name, shape=shape, initializer=initializer, dtype=dtype) if constrain is not None: constrain(biases) return biases def factorized_readout(inputs, n_outputs=100, constrain=True): width, height, n_features = inputs.get_shape()[1:].as_list() n_pixels = width * height with tf.variable_scope('readout'): # spatial readout w_spatial = weight_variable([n_pixels, 1, n_outputs], name='weight_spatial') if constrain: constraints.positive_constrain(w_spatial) w_spatial_norm = normalize_weights(w_spatial, dims=(0,)) # feature readout w_feature = weight_variable([1, n_features, n_outputs], name='weight_feature') if constrain: constraints.positive_constrain(w_feature) w_feature_norm = normalize_weights(w_feature, dims=(1,)) # scaling w_scale = bias_variable([n_outputs], name='weight_scale', value=1.
batch_size = tf.shape(inputs)[0] padded_output = [batch_size] + padded_output[1:] output = tf.nn.conv2d_transpose(inputs, weight, padded_output, strides, padding_type, name='transpose_convolution') if padding_type=='VALID' and np.sum(padding) > 0:
random_line_split
display.go
(win *winInfo) setTitle(position int, total int, path string) { win.window.SetTitle(winTitle + " - " + strconv.Itoa(position) + "/" + strconv.Itoa(total) + " - " + filepath.Base(path)) } // Create the texture from text using TTF func (win *winInfo) renderText(text string) (*sdl.Texture, error) { surface, err := win.font.RenderUTF8Shaded(text, sdlColorBlack, sdlColorWhite) defer surface.Free() if err != nil { return nil, err } texture, err := win.renderer.CreateTextureFromSurface(surface) if err != nil { return nil, err } return texture, nil } // displayLoading display loading on background func (win *winInfo) displayLoading() { texture, _ := win.renderText("Loading...") ww, wh := win.window.GetSize() win.renderer.Copy(texture, nil, &sdl.Rect{X: int32(ww/2 - 50), Y: int32(wh/2 - 1), W: 65, H: 20}) texture.Destroy() win.renderer.Present() } // displayPictureInfo display all information about the picture func (win *winInfo) displayPictureInfo() { /* Display exif information */ go func(path string) { exif.Open(path) }(curImg().path) msg := filepath.Base(curImg().path) texture, _ := win.renderText(msg) width := int32(len(msg) * 8) win.renderer.Copy(texture, nil, &sdl.Rect{X: 2, Y: 2, W: width, H: 20}) texture.Destroy() } func loadImg(win *winInfo, index int) { if slide.list[index].texture == nil { var err error var surface *sdl.Surface logger.Trace("load " + slide.list[index].path) surface, err = img.Load(slide.list[index].path) if err != nil { fmt.Printf("Failed to load: %s\n", err) } defer surface.Free() slide.list[index].H = surface.H slide.list[index].W = surface.W slide.list[index].texture, err = win.renderer.CreateTextureFromSurface(surface) if err != nil { fmt.Printf("Failed to create texture: %s\n", err) } } } func (win *winInfo) displayBar() { message := "\uF04A \uF04B \uF04E \uF0E2 \uF01E" surface, err := win.symbols.RenderUTF8Shaded(message, sdlColorWhite, sdlColorBlack) if err == nil { texture, err := win.renderer.CreateTextureFromSurface(surface) if err != nil { fmt.Println(err) } surface.Free() width := int32(len(message) * 18) win.renderer.Copy(texture, nil, &sdl.Rect{X: 120, Y: 500, W: width, H: 40}) texture.Destroy() } else { logger.Warning("OMG") } } func resetImg(index int) { if slide.list[index].texture != nil { slide.list[index].texture.Destroy() slide.list[index].texture = nil } } func (win *winInfo) loadAndFreeAround() { p1 := utils.Mod(slide.current-1, len(slide.list)) p2 := utils.Mod(slide.current-1, len(slide.list)) n1 := utils.Mod(slide.current+1, len(slide.list)) n2 := utils.Mod(slide.current+2, len(slide.list)) n3 := utils.Mod(slide.current+3, len(slide.list)) refresh := utils.IntList{p1, p2, n1, n2, n3} // preload the previous and next two images for _, idx := range refresh { loadImg(win, idx) } d1 := utils.Mod(slide.current-3, len(slide.list)) d2 := utils.Mod(slide.current+3, len(slide.list)) if !refresh.Find(d1) { resetImg(d1) } if !refresh.Find(d2) { resetImg(d2) } } func (win *winInfo) loadCurrentImage(render bool) { var src, dst sdl.Rect // load and display the current image loadImg(win, slide.current) // Display information of the image ww, wh := win.window.GetSize() src = sdl.Rect{X: 0, Y: 0, W: curImg().W, H: curImg().H} iw, ih := utils.ComputeFitImage(uint32(ww), uint32(wh), uint32(curImg().W), uint32(curImg().H))
if render { win.renderer.Clear() win.renderer.Copy(curImg().texture, &src, &dst) if window.displayInfo { window.displayPictureInfo() } //window.displayBar() win.renderer.Present() } // Update the window title win.setTitle(slide.current+1, len(slide.list), curImg().path) // Preload and free images from the list win.loadAndFreeAround() } // Arrange that main.main runs on main thread. func init() { runtime.LockOSThread() // Video only if err := sdl.Init(sdl.INIT_VIDEO); err != nil { logger.Warning(err.Error()) } } var window winInfo // MainLoop initializes the SDL package and run the main loop func MainLoop(fullScreen bool, slideshow bool) int { var event sdl.Event var src, dst sdl.Rect var err error var flags uint32 = sdl.WINDOW_SHOWN | sdl.WINDOW_RESIZABLE | sdl.WINDOW_ALLOW_HIGHDPI // Load the font library if err := ttf.Init(); err != nil { logger.Warning("Unable to open font lib") } window.window, err = sdl.CreateWindow(winTitle, sdl.WINDOWPOS_UNDEFINED, sdl.WINDOWPOS_UNDEFINED, winDefaultHeight, winDefaultWidth, flags) if err != nil { fmt.Fprintf(os.Stderr, "Failed to create window: %s\n", err) return 1 } defer window.window.Destroy() // Load resources if f, err := filepath.Abs(filepath.Dir(os.Args[0])); err == nil { icon := filepath.Join(f, "app", "icon.bmp") if i, err := sdl.LoadBMP(icon); err == nil { window.window.SetIcon(i) } font := filepath.Join(f, "app", "fonts", "opensans.ttf") window.font, err = ttf.OpenFont(font, 14) if err != nil { logger.Warning("Unable to load " + font) } window.font.SetKerning(false) font = filepath.Join(f, "app", "fonts", "fontawesome.ttf") window.symbols, err = ttf.OpenFont(font, 64) if err != nil { logger.Warning("Unable to load " + font) } } window.renderer, err = sdl.CreateRenderer(window.window, -1, sdl.RENDERER_ACCELERATED) if err != nil { fmt.Fprintf(os.Stderr, "Failed to create renderer: %s\n", err) return 2 } defer window.renderer.Destroy() window.displayInfo = false window.displayLoading() window.setTitle(slide.current+1, len(slide.list), curImg().path) window.loadCurrentImage(false) // Declare if the image needs to be updated var update = false var running = true for running { event = sdl.WaitEvent() switch t := event.(type) { case *sdl.QuitEvent: running = false case *sdl.DropEvent: fileName := t.File // Check if picture already in list found := false for i := range slide.list { if slide.list[i].path == fileName { found = true slide.current = i update = true break } } if !found { if err := addPic(fileName); err != nil { sdl.ShowSimpleMessageBox(sdl.MESSAGEBOX_INFORMATION, "File dropped on window", "Cannot add "+fileName, window.window) } else { slide.current = len(slide.list) - 1 update = true } } /*case *sdl.MouseMotionEvent: fmt.Printf("[%d ms] MouseMotion\ttype:%d\tid:%d\tx:%d\ty:%d\txrel:%d\tyrel:%d\n", t.Timestamp, t.Type, t.Which, t.X, t.Y, t.XRel, t.YRel) case *sdl.MouseButtonEvent: fmt.Printf("[%d ms] MouseButton\ttype:%d\tid:%d\tx:%d\ty:%d\tbutton:%d\tstate:%d\n", t.Timestamp, t.Type, t.Which, t.X, t.Y, t.Button, t.State)*/ case *
dst = sdl.Rect{X: int32(ww/2 - int32(iw)/2), Y: int32(wh/2 - int32(ih)/2), W: int32(iw), H: int32(ih)}
random_line_split
display.go
win *winInfo) setTitle(position int, total int, path string) { win.window.SetTitle(winTitle + " - " + strconv.Itoa(position) + "/" + strconv.Itoa(total) + " - " + filepath.Base(path)) } // Create the texture from text using TTF func (win *winInfo) renderText(text string) (*sdl.Texture, error) { surface, err := win.font.RenderUTF8Shaded(text, sdlColorBlack, sdlColorWhite) defer surface.Free() if err != nil { return nil, err } texture, err := win.renderer.CreateTextureFromSurface(surface) if err != nil { return nil, err } return texture, nil } // displayLoading display loading on background func (win *winInfo) displayLoading() { texture, _ := win.renderText("Loading...") ww, wh := win.window.GetSize() win.renderer.Copy(texture, nil, &sdl.Rect{X: int32(ww/2 - 50), Y: int32(wh/2 - 1), W: 65, H: 20}) texture.Destroy() win.renderer.Present() } // displayPictureInfo display all information about the picture func (win *winInfo) displayPictureInfo() { /* Display exif information */ go func(path string) { exif.Open(path) }(curImg().path) msg := filepath.Base(curImg().path) texture, _ := win.renderText(msg) width := int32(len(msg) * 8) win.renderer.Copy(texture, nil, &sdl.Rect{X: 2, Y: 2, W: width, H: 20}) texture.Destroy() } func loadImg(win *winInfo, index int) { if slide.list[index].texture == nil { var err error var surface *sdl.Surface logger.Trace("load " + slide.list[index].path) surface, err = img.Load(slide.list[index].path) if err != nil { fmt.Printf("Failed to load: %s\n", err) } defer surface.Free() slide.list[index].H = surface.H slide.list[index].W = surface.W slide.list[index].texture, err = win.renderer.CreateTextureFromSurface(surface) if err != nil { fmt.Printf("Failed to create texture: %s\n", err) } } } func (win *winInfo) displayBar() { message := "\uF04A \uF04B \uF04E \uF0E2 \uF01E" surface, err := win.symbols.RenderUTF8Shaded(message, sdlColorWhite, sdlColorBlack) if err == nil { texture, err := win.renderer.CreateTextureFromSurface(surface) if err != nil { fmt.Println(err) } surface.Free() width := int32(len(message) * 18) win.renderer.Copy(texture, nil, &sdl.Rect{X: 120, Y: 500, W: width, H: 40}) texture.Destroy() } else { logger.Warning("OMG") } } func resetImg(index int) { if slide.list[index].texture != nil { slide.list[index].texture.Destroy() slide.list[index].texture = nil } } func (win *winInfo) loadAndFreeAround() { p1 := utils.Mod(slide.current-1, len(slide.list)) p2 := utils.Mod(slide.current-1, len(slide.list)) n1 := utils.Mod(slide.current+1, len(slide.list)) n2 := utils.Mod(slide.current+2, len(slide.list)) n3 := utils.Mod(slide.current+3, len(slide.list)) refresh := utils.IntList{p1, p2, n1, n2, n3} // preload the previous and next two images for _, idx := range refresh { loadImg(win, idx) } d1 := utils.Mod(slide.current-3, len(slide.list)) d2 := utils.Mod(slide.current+3, len(slide.list)) if !refresh.Find(d1) { resetImg(d1) } if !refresh.Find(d2) { resetImg(d2) } } func (win *winInfo) loadCurrentImage(render bool) { var src, dst sdl.Rect // load and display the current image loadImg(win, slide.current) // Display information of the image ww, wh := win.window.GetSize() src = sdl.Rect{X: 0, Y: 0, W: curImg().W, H: curImg().H} iw, ih := utils.ComputeFitImage(uint32(ww), uint32(wh), uint32(curImg().W), uint32(curImg().H)) dst = sdl.Rect{X: int32(ww/2 - int32(iw)/2), Y: int32(wh/2 - int32(ih)/2), W: int32(iw), H: int32(ih)} if render { win.renderer.Clear() win.renderer.Copy(curImg().texture, &src, &dst) if window.displayInfo { window.displayPictureInfo() } //window.displayBar() win.renderer.Present() } // Update the window title win.setTitle(slide.current+1, len(slide.list), curImg().path) // Preload and free images from the list win.loadAndFreeAround() } // Arrange that main.main runs on main thread. func init() { runtime.LockOSThread() // Video only if err := sdl.Init(sdl.INIT_VIDEO); err != nil { logger.Warning(err.Error()) } } var window winInfo // MainLoop initializes the SDL package and run the main loop func
(fullScreen bool, slideshow bool) int { var event sdl.Event var src, dst sdl.Rect var err error var flags uint32 = sdl.WINDOW_SHOWN | sdl.WINDOW_RESIZABLE | sdl.WINDOW_ALLOW_HIGHDPI // Load the font library if err := ttf.Init(); err != nil { logger.Warning("Unable to open font lib") } window.window, err = sdl.CreateWindow(winTitle, sdl.WINDOWPOS_UNDEFINED, sdl.WINDOWPOS_UNDEFINED, winDefaultHeight, winDefaultWidth, flags) if err != nil { fmt.Fprintf(os.Stderr, "Failed to create window: %s\n", err) return 1 } defer window.window.Destroy() // Load resources if f, err := filepath.Abs(filepath.Dir(os.Args[0])); err == nil { icon := filepath.Join(f, "app", "icon.bmp") if i, err := sdl.LoadBMP(icon); err == nil { window.window.SetIcon(i) } font := filepath.Join(f, "app", "fonts", "opensans.ttf") window.font, err = ttf.OpenFont(font, 14) if err != nil { logger.Warning("Unable to load " + font) } window.font.SetKerning(false) font = filepath.Join(f, "app", "fonts", "fontawesome.ttf") window.symbols, err = ttf.OpenFont(font, 64) if err != nil { logger.Warning("Unable to load " + font) } } window.renderer, err = sdl.CreateRenderer(window.window, -1, sdl.RENDERER_ACCELERATED) if err != nil { fmt.Fprintf(os.Stderr, "Failed to create renderer: %s\n", err) return 2 } defer window.renderer.Destroy() window.displayInfo = false window.displayLoading() window.setTitle(slide.current+1, len(slide.list), curImg().path) window.loadCurrentImage(false) // Declare if the image needs to be updated var update = false var running = true for running { event = sdl.WaitEvent() switch t := event.(type) { case *sdl.QuitEvent: running = false case *sdl.DropEvent: fileName := t.File // Check if picture already in list found := false for i := range slide.list { if slide.list[i].path == fileName { found = true slide.current = i update = true break } } if !found { if err := addPic(fileName); err != nil { sdl.ShowSimpleMessageBox(sdl.MESSAGEBOX_INFORMATION, "File dropped on window", "Cannot add "+fileName, window.window) } else { slide.current = len(slide.list) - 1 update = true } } /*case *sdl.MouseMotionEvent: fmt.Printf("[%d ms] MouseMotion\ttype:%d\tid:%d\tx:%d\ty:%d\txrel:%d\tyrel:%d\n", t.Timestamp, t.Type, t.Which, t.X, t.Y, t.XRel, t.YRel) case *sdl.MouseButtonEvent: fmt.Printf("[%d ms] MouseButton\ttype:%d\tid:%d\tx:%d\ty:%d\tbutton:%d\tstate:%d\n", t.Timestamp, t.Type, t.Which, t.X, t.Y, t.Button, t.State)*/ case *
MainLoop
identifier_name
display.go
win *winInfo) setTitle(position int, total int, path string) { win.window.SetTitle(winTitle + " - " + strconv.Itoa(position) + "/" + strconv.Itoa(total) + " - " + filepath.Base(path)) } // Create the texture from text using TTF func (win *winInfo) renderText(text string) (*sdl.Texture, error) { surface, err := win.font.RenderUTF8Shaded(text, sdlColorBlack, sdlColorWhite) defer surface.Free() if err != nil { return nil, err } texture, err := win.renderer.CreateTextureFromSurface(surface) if err != nil { return nil, err } return texture, nil } // displayLoading display loading on background func (win *winInfo) displayLoading() { texture, _ := win.renderText("Loading...") ww, wh := win.window.GetSize() win.renderer.Copy(texture, nil, &sdl.Rect{X: int32(ww/2 - 50), Y: int32(wh/2 - 1), W: 65, H: 20}) texture.Destroy() win.renderer.Present() } // displayPictureInfo display all information about the picture func (win *winInfo) displayPictureInfo() { /* Display exif information */ go func(path string) { exif.Open(path) }(curImg().path) msg := filepath.Base(curImg().path) texture, _ := win.renderText(msg) width := int32(len(msg) * 8) win.renderer.Copy(texture, nil, &sdl.Rect{X: 2, Y: 2, W: width, H: 20}) texture.Destroy() } func loadImg(win *winInfo, index int) { if slide.list[index].texture == nil { var err error var surface *sdl.Surface logger.Trace("load " + slide.list[index].path) surface, err = img.Load(slide.list[index].path) if err != nil { fmt.Printf("Failed to load: %s\n", err) } defer surface.Free() slide.list[index].H = surface.H slide.list[index].W = surface.W slide.list[index].texture, err = win.renderer.CreateTextureFromSurface(surface) if err != nil { fmt.Printf("Failed to create texture: %s\n", err) } } } func (win *winInfo) displayBar() { message := "\uF04A \uF04B \uF04E \uF0E2 \uF01E" surface, err := win.symbols.RenderUTF8Shaded(message, sdlColorWhite, sdlColorBlack) if err == nil { texture, err := win.renderer.CreateTextureFromSurface(surface) if err != nil { fmt.Println(err) } surface.Free() width := int32(len(message) * 18) win.renderer.Copy(texture, nil, &sdl.Rect{X: 120, Y: 500, W: width, H: 40}) texture.Destroy() } else { logger.Warning("OMG") } } func resetImg(index int) { if slide.list[index].texture != nil { slide.list[index].texture.Destroy() slide.list[index].texture = nil } } func (win *winInfo) loadAndFreeAround() { p1 := utils.Mod(slide.current-1, len(slide.list)) p2 := utils.Mod(slide.current-1, len(slide.list)) n1 := utils.Mod(slide.current+1, len(slide.list)) n2 := utils.Mod(slide.current+2, len(slide.list)) n3 := utils.Mod(slide.current+3, len(slide.list)) refresh := utils.IntList{p1, p2, n1, n2, n3} // preload the previous and next two images for _, idx := range refresh { loadImg(win, idx) } d1 := utils.Mod(slide.current-3, len(slide.list)) d2 := utils.Mod(slide.current+3, len(slide.list)) if !refresh.Find(d1) { resetImg(d1) } if !refresh.Find(d2) { resetImg(d2) } } func (win *winInfo) loadCurrentImage(render bool) { var src, dst sdl.Rect // load and display the current image loadImg(win, slide.current) // Display information of the image ww, wh := win.window.GetSize() src = sdl.Rect{X: 0, Y: 0, W: curImg().W, H: curImg().H} iw, ih := utils.ComputeFitImage(uint32(ww), uint32(wh), uint32(curImg().W), uint32(curImg().H)) dst = sdl.Rect{X: int32(ww/2 - int32(iw)/2), Y: int32(wh/2 - int32(ih)/2), W: int32(iw), H: int32(ih)} if render { win.renderer.Clear() win.renderer.Copy(curImg().texture, &src, &dst) if window.displayInfo { window.displayPictureInfo() } //window.displayBar() win.renderer.Present() } // Update the window title win.setTitle(slide.current+1, len(slide.list), curImg().path) // Preload and free images from the list win.loadAndFreeAround() } // Arrange that main.main runs on main thread. func init()
var window winInfo // MainLoop initializes the SDL package and run the main loop func MainLoop(fullScreen bool, slideshow bool) int { var event sdl.Event var src, dst sdl.Rect var err error var flags uint32 = sdl.WINDOW_SHOWN | sdl.WINDOW_RESIZABLE | sdl.WINDOW_ALLOW_HIGHDPI // Load the font library if err := ttf.Init(); err != nil { logger.Warning("Unable to open font lib") } window.window, err = sdl.CreateWindow(winTitle, sdl.WINDOWPOS_UNDEFINED, sdl.WINDOWPOS_UNDEFINED, winDefaultHeight, winDefaultWidth, flags) if err != nil { fmt.Fprintf(os.Stderr, "Failed to create window: %s\n", err) return 1 } defer window.window.Destroy() // Load resources if f, err := filepath.Abs(filepath.Dir(os.Args[0])); err == nil { icon := filepath.Join(f, "app", "icon.bmp") if i, err := sdl.LoadBMP(icon); err == nil { window.window.SetIcon(i) } font := filepath.Join(f, "app", "fonts", "opensans.ttf") window.font, err = ttf.OpenFont(font, 14) if err != nil { logger.Warning("Unable to load " + font) } window.font.SetKerning(false) font = filepath.Join(f, "app", "fonts", "fontawesome.ttf") window.symbols, err = ttf.OpenFont(font, 64) if err != nil { logger.Warning("Unable to load " + font) } } window.renderer, err = sdl.CreateRenderer(window.window, -1, sdl.RENDERER_ACCELERATED) if err != nil { fmt.Fprintf(os.Stderr, "Failed to create renderer: %s\n", err) return 2 } defer window.renderer.Destroy() window.displayInfo = false window.displayLoading() window.setTitle(slide.current+1, len(slide.list), curImg().path) window.loadCurrentImage(false) // Declare if the image needs to be updated var update = false var running = true for running { event = sdl.WaitEvent() switch t := event.(type) { case *sdl.QuitEvent: running = false case *sdl.DropEvent: fileName := t.File // Check if picture already in list found := false for i := range slide.list { if slide.list[i].path == fileName { found = true slide.current = i update = true break } } if !found { if err := addPic(fileName); err != nil { sdl.ShowSimpleMessageBox(sdl.MESSAGEBOX_INFORMATION, "File dropped on window", "Cannot add "+fileName, window.window) } else { slide.current = len(slide.list) - 1 update = true } } /*case *sdl.MouseMotionEvent: fmt.Printf("[%d ms] MouseMotion\ttype:%d\tid:%d\tx:%d\ty:%d\txrel:%d\tyrel:%d\n", t.Timestamp, t.Type, t.Which, t.X, t.Y, t.XRel, t.YRel) case *sdl.MouseButtonEvent: fmt.Printf("[%d ms] MouseButton\ttype:%d\tid:%d\tx:%d\ty:%d\tbutton:%d\tstate:%d\n", t.Timestamp, t.Type, t.Which, t.X, t.Y, t.Button, t.State)*/ case
{ runtime.LockOSThread() // Video only if err := sdl.Init(sdl.INIT_VIDEO); err != nil { logger.Warning(err.Error()) } }
identifier_body
display.go
texture.Destroy() } else { logger.Warning("OMG") } } func resetImg(index int) { if slide.list[index].texture != nil { slide.list[index].texture.Destroy() slide.list[index].texture = nil } } func (win *winInfo) loadAndFreeAround() { p1 := utils.Mod(slide.current-1, len(slide.list)) p2 := utils.Mod(slide.current-1, len(slide.list)) n1 := utils.Mod(slide.current+1, len(slide.list)) n2 := utils.Mod(slide.current+2, len(slide.list)) n3 := utils.Mod(slide.current+3, len(slide.list)) refresh := utils.IntList{p1, p2, n1, n2, n3} // preload the previous and next two images for _, idx := range refresh { loadImg(win, idx) } d1 := utils.Mod(slide.current-3, len(slide.list)) d2 := utils.Mod(slide.current+3, len(slide.list)) if !refresh.Find(d1) { resetImg(d1) } if !refresh.Find(d2) { resetImg(d2) } } func (win *winInfo) loadCurrentImage(render bool) { var src, dst sdl.Rect // load and display the current image loadImg(win, slide.current) // Display information of the image ww, wh := win.window.GetSize() src = sdl.Rect{X: 0, Y: 0, W: curImg().W, H: curImg().H} iw, ih := utils.ComputeFitImage(uint32(ww), uint32(wh), uint32(curImg().W), uint32(curImg().H)) dst = sdl.Rect{X: int32(ww/2 - int32(iw)/2), Y: int32(wh/2 - int32(ih)/2), W: int32(iw), H: int32(ih)} if render { win.renderer.Clear() win.renderer.Copy(curImg().texture, &src, &dst) if window.displayInfo { window.displayPictureInfo() } //window.displayBar() win.renderer.Present() } // Update the window title win.setTitle(slide.current+1, len(slide.list), curImg().path) // Preload and free images from the list win.loadAndFreeAround() } // Arrange that main.main runs on main thread. func init() { runtime.LockOSThread() // Video only if err := sdl.Init(sdl.INIT_VIDEO); err != nil { logger.Warning(err.Error()) } } var window winInfo // MainLoop initializes the SDL package and run the main loop func MainLoop(fullScreen bool, slideshow bool) int { var event sdl.Event var src, dst sdl.Rect var err error var flags uint32 = sdl.WINDOW_SHOWN | sdl.WINDOW_RESIZABLE | sdl.WINDOW_ALLOW_HIGHDPI // Load the font library if err := ttf.Init(); err != nil { logger.Warning("Unable to open font lib") } window.window, err = sdl.CreateWindow(winTitle, sdl.WINDOWPOS_UNDEFINED, sdl.WINDOWPOS_UNDEFINED, winDefaultHeight, winDefaultWidth, flags) if err != nil { fmt.Fprintf(os.Stderr, "Failed to create window: %s\n", err) return 1 } defer window.window.Destroy() // Load resources if f, err := filepath.Abs(filepath.Dir(os.Args[0])); err == nil { icon := filepath.Join(f, "app", "icon.bmp") if i, err := sdl.LoadBMP(icon); err == nil { window.window.SetIcon(i) } font := filepath.Join(f, "app", "fonts", "opensans.ttf") window.font, err = ttf.OpenFont(font, 14) if err != nil { logger.Warning("Unable to load " + font) } window.font.SetKerning(false) font = filepath.Join(f, "app", "fonts", "fontawesome.ttf") window.symbols, err = ttf.OpenFont(font, 64) if err != nil { logger.Warning("Unable to load " + font) } } window.renderer, err = sdl.CreateRenderer(window.window, -1, sdl.RENDERER_ACCELERATED) if err != nil { fmt.Fprintf(os.Stderr, "Failed to create renderer: %s\n", err) return 2 } defer window.renderer.Destroy() window.displayInfo = false window.displayLoading() window.setTitle(slide.current+1, len(slide.list), curImg().path) window.loadCurrentImage(false) // Declare if the image needs to be updated var update = false var running = true for running { event = sdl.WaitEvent() switch t := event.(type) { case *sdl.QuitEvent: running = false case *sdl.DropEvent: fileName := t.File // Check if picture already in list found := false for i := range slide.list { if slide.list[i].path == fileName { found = true slide.current = i update = true break } } if !found { if err := addPic(fileName); err != nil { sdl.ShowSimpleMessageBox(sdl.MESSAGEBOX_INFORMATION, "File dropped on window", "Cannot add "+fileName, window.window) } else { slide.current = len(slide.list) - 1 update = true } } /*case *sdl.MouseMotionEvent: fmt.Printf("[%d ms] MouseMotion\ttype:%d\tid:%d\tx:%d\ty:%d\txrel:%d\tyrel:%d\n", t.Timestamp, t.Type, t.Which, t.X, t.Y, t.XRel, t.YRel) case *sdl.MouseButtonEvent: fmt.Printf("[%d ms] MouseButton\ttype:%d\tid:%d\tx:%d\ty:%d\tbutton:%d\tstate:%d\n", t.Timestamp, t.Type, t.Which, t.X, t.Y, t.Button, t.State)*/ case *sdl.WindowEvent: if t.Event == sdl.WINDOWEVENT_RESIZED || t.Event == sdl.WINDOWEVENT_EXPOSED { window.window.SetSize(t.Data1, t.Data2) // Display information of the image wWidth, wHeight := window.window.GetSize() src = sdl.Rect{X: 0, Y: 0, W: curImg().W, H: curImg().H} fitWidth, fitHeight := utils.ComputeFitImage(uint32(wWidth), uint32(wHeight), uint32(curImg().W), uint32(curImg().H)) dst = sdl.Rect{X: int32(wWidth/2 - int32(fitWidth)/2), Y: int32(wHeight/2 - int32(fitHeight)/2), W: int32(fitWidth), H: int32(fitHeight)} window.renderer.Clear() window.renderer.Copy(curImg().texture, &src, &dst) window.renderer.Present() if window.displayInfo { window.displayPictureInfo() window.renderer.Present() } } case *sdl.KeyboardEvent: if t.GetType() != sdl.KEYDOWN { break } // Get next or previous image if t.Repeat == 0 { if t.Keysym.Sym == sdl.K_LEFT { slide.current = utils.Mod((slide.current - 1), len(slide.list)) update = true } else if t.Keysym.Sym == sdl.K_RIGHT { slide.current = utils.Mod((slide.current + 1), len(slide.list)) update = true } else if t.Keysym.Sym == sdl.K_PAGEUP { if err := picture.RotateImage(curImg().path, picture.CounterClockwise); err != nil { logger.Warning(err.Error()) } else { resetImg(slide.current) } update = true } else if t.Keysym.Sym == sdl.K_PAGEDOWN { if err := picture.RotateImage(curImg().path, picture.Clockwise); err != nil { logger.Warning(err.Error()) } else { resetImg(slide.current) } update = true } else if t.Keysym.Sym == 102 { // F if window.fullscreen { window.window.SetFullscreen(0) } else { // Go fullscreen window.window.SetFullscreen(sdl.WINDOW_FULLSCREEN_DESKTOP) } window.fullscreen = !window.fullscreen } else if t.Keysym.Sym == 105 { // I window.displayInfo = !window.displayInfo if window.displayInfo { fmt.Println("Toggle info: on") window.displayPictureInfo() window.renderer.Present() } else { fmt.Println("Toggle info: off") update = true } } else if t.Keysym.Sym == sdl.K_ESCAPE { if window.fullscreen
{ window.window.SetFullscreen(0) window.fullscreen = false }
conditional_block
w_reGeorgSocksProxy.py
SocksProtocolNotImplemented(Exception): pass class RemoteConnectionFailed(Exception): pass class session(Thread): def __init__(self, pSocket, connectString): Thread.__init__(self) self.pSocket = pSocket self.connectString = connectString o = urlparse(connectString) try: self.httpPort = o.port except: if o.scheme == "https": self.httpPort = 443 else: self.httpPort = 80 else: if not o.port: if o.scheme == "https": self.httpPort = 443 else: self.httpPort = 80 self.httpScheme = o.scheme self.httpHost = o.netloc.split(":")[0] self.httpPath = o.path self.cookie = None def parseSocks5(self, sock): logger.debug("SocksVersion5 detected") # 02:00 nmethods, methods = (sock.recv(1), sock.recv(1)) # 05:00 sock.sendall(VER + METHOD) # :02 ver = sock.recv(1) if ver == "\x02": # this is a hack for proxychains # 05:01:00:01----:c0:a8:01:02:00:50 # '\x05', '\x01', '\x00', '\x01' ver, cmd, rsv, atyp = (sock.recv(1), sock.recv(1), sock.recv(1), sock.recv(1)) else: cmd, rsv, atyp = (sock.recv(1), sock.recv(1), sock.recv(1)) target = None targetPort = None if atyp == "\x01": # IPv4 # Reading 6 bytes for the IP and Port # c0:a8:01:02 target = sock.recv(4) # 00:50 targetPort = sock.recv(2) # 目标地址192.168.2.1 self.target = ".".join([str(ord(i)) for i in target]) elif atyp == "\x03": # Hostname targetLen = ord(sock.recv(1)) # hostname length (1 byte) target = sock.recv(targetLen) targetPort = sock.recv(2) target = "".join([unichr(ord(i)) for i in target]) elif atyp == "\x04": # IPv6 target = sock.recv(16) targetPort = sock.recv(2) tmp_addr = [] for i in xrange(len(target) / 2): tmp_addr.append(unichr(ord(target[2 * i]) * 256 + ord(target[2 * i + 1]))) target = ":".join(tmp_addr) # 80 self.targetPort = ord(targetPort[0]) * 256 + ord(targetPort[1]) if cmd == "\x02": # BIND raise SocksCmdNotImplemented("Socks5 - BIND not implemented") elif cmd == "\x03": # UDP raise SocksCmdNotImplemented("Socks5 - UDP not implemented") elif cmd == "\x01": # CONNECT serverIp = target try: serverIp = gethostbyname(self.target) except: logger.error("oeps") # 又转回来\xc0\xa8\x02\x01 serverIp = "".join([chr(int(i)) for i in serverIp.split(".")]) # 获取cookie,在服务端的脚本中,会执行相应端口探测 self.cookie = self.setupRemoteSession(target=self.target, targetPort=str(self.targetPort)) if self.cookie: sock.sendall(VER + SUCCESS + "\x00" + "\x01" + serverIp + chr(self.targetPort / 256) + chr( self.targetPort % 256)) return True else: sock.sendall(VER + REFUSED + "\x00" + "\x01" + serverIp + chr(self.targetPort / 256) + chr( self.targetPort % 256)) return False def handleSocks(self, sock): # 通过proxychain模拟客户端发送数据,第一个字节可以判断是socks5还是socks4 ver = sock.recv(1) # 05:02:00:02 if ver == "\x05": return self.parseSocks5(sock) def setupRemoteSession(self, target, targetPort): """探测端口存活""" header = ({"X-CMD": "CONNECT", "X-TARGET": target, "X-PORT": targetPort}) cookie = None try: response = requests.post(url=self.connectString, headers=header, data=None, timeout=TIMEOUT) except Exception, e: return else: if response: response_header = response.headers if response.status_code == 200 and response_header.get("X-STATUS") == "OK": cookie = response_header.get("Set-Cookie") logger.info("[%s:%s] HTTP [200]: cookie [%s]" % (target, targetPort, cookie)) elif response_header.get("X-ERROR"): logger.error(response_header.get("X-ERROR")) else: logger.error("[%s:%s] HTTP [%d]" % (target, targetPort, response.status_code)) return cookie def closeRemoteSession(self): header = {"X-CMD": "DISCONNECT", "Cookie": self.cookie} try: response = requests.post(url=self.connectString, headers=header, data=None, timeout=TIMEOUT) except Exception, e: logger.error("Close Connection Failure") else: if response.status_code == 200: logger.info("[%s:%d] Connection Terminated" % (self.httpHost, self.httpPort)) def reader(self): while True: try: if not self.pSocket: break header = {"X-CMD": "READ", "Cookie": self.cookie, "Connection": "Keep-Alive"} response = requests.post(url=self.connectString, headers=header, data=None) response_data = None if response.status_code == 200: response_header = response.headers status = response_header.get("x-status") if status == "OK": response_data = response.content else: logger.error("[%s:%d] HTTP [%d]: Status: [%s]: Message [%s] Shutting down" % ( self.target, self.targetPort, response.status_code, status, response_header.get("X-ERROR"))) else: logger.error( "[%s:%d] HTTP [%d]: Shutting down" % (self.target, self.targetPort, response.status_code)) if response_data is None: # Remote socket closed break if len(response_data) == 0: time.sleep(0.1) continue self.pSocket.send(response_data) except Exception, ex: print(format_exc()) raise ex self.closeRemoteSession() logger.debug("[%s:%d] Closing localsocket" % (self.target, self.targetPort)) try: self.pSocket.close() except: logger.debug("[%s:%d] Localsocket already closed" % (self.target, self.targetPort)) def writer(self): global READBUFSIZE while True: try: self.pSocket.settimeou
# 'GET / HTTP/1.1\r\nHost: 192.168.2.1\r\nUser-Agent: curl/7.58.0\r\nAccept: */*\r\n\r\n' data = self.pSocket.recv(READBUFSIZE) if not data: break header = {"X-CMD": "FORWARD", "Cookie": self.cookie, "Content-Type": "application/octet-stream", "Connection": "Keep-Alive"} # 携带数据 response = requests.post(url=self.connectString, headers=header, data=data) if response.status_code == 200: response_header = response.headers status = response_header.get("x-status") if status == "OK": if response_header.get("set-cookie") is not None: self.cookie = response_header.get("set-cookie") else: logger.error("[%s:%d] HTTP [%d]: Status: [%s]: Message [%s] Shutting down" % ( self.target, self.targetPort, response.status_code, status, response_header.get("x-error"))) break else: logger.error( "[%s:%d] HTTP [%d]: Shutting down" % (self.target, self.targetPort, response.status_code)) break # transferLog.info("[%s:%d] >>>> [%d]" % (self.target, self.port, len(data))) except timeout: continue except Exception, ex: raise ex self.closeRemoteSession() logger.debug("
t(1)
identifier_name
w_reGeorgSocksProxy.py
class SocksProtocolNotImplemented(Exception): pass class RemoteConnectionFailed(Exception): pass class session(Thread): def __init__(self, pSocket, connectString): Thread.__init__(self) self.pSocket = pSocket self.connectString = connectString o = urlparse(connectString) try: self.httpPort = o.port except: if o.scheme == "https": self.httpPort = 443 else: self.httpPort = 80 else: if not o.port: if o.scheme == "https": self.httpPort = 443 else: self.httpPort = 80 self.httpScheme = o.scheme self.httpHost = o.netloc.split(":")[0] self.httpPath = o.path self.cookie = None def parseSocks5(self, sock): logger.debug("SocksVersion5 detected") # 02:00 nmethods, methods = (sock.recv(1), sock.recv(1)) # 05:00 sock.sendall(VER + METHOD) # :02 ver = sock.recv(1) if ver == "\x02": # this is a hack for proxychains # 05:01:00:01----:c0:a8:01:02:00:50 # '\x05', '\x01', '\x00', '\x01' ver, cmd, rsv, atyp = (sock.recv(1), sock.recv(1), sock.recv(1), sock.recv(1)) else: cmd, rsv, atyp = (sock.recv(1), sock.recv(1), sock.recv(1)) target = None targetPort = None if atyp == "\x01": # IPv4 # Reading 6 bytes for the IP and Port # c0:a8:01:02 target = sock.recv(4) # 00:50 targetPort = sock.recv(2) # 目标地址192.168.2.1 self.target = ".".join([str(ord(i)) for i in target]) elif atyp == "\x03": # Hostname targetLen = ord(sock.recv(1)) # hostname length (1 byte) target = sock.recv(targetLen) targetPort = sock.recv(2) target = "".join([unichr(ord(i)) for i in target]) elif atyp == "\x04": # IPv6 target = sock.recv(16) targetPort = sock.recv(2) tmp_addr = [] for i in xrange(len(target) / 2): tmp_addr.append(unichr(ord(target[2 * i]) * 256 + ord(target[2 * i + 1]))) target = ":".join(tmp_addr) # 80 self.targetPort = ord(targetPort[0]) * 256 + ord(targetPort[1]) if cmd == "\x02": # BIND raise SocksCmdNotImplemented("Socks5 - BIND not implemented") elif cmd == "\x03": # UDP raise SocksCmdNotImplemented("Socks5 - UDP not implemented") elif cmd == "\x01": # CONNECT serverIp = target try: serverIp = gethostbyname(self.target) except: logger.error("oeps") # 又转回来\xc0\xa8\x02\x01 serverIp = "".join([chr(int(i)) for i in serverIp.split(".")]) # 获取cookie,在服务端的脚本中,会执行相应端口探测 self.cookie = self.setupRemoteSession(target=self.target, targetPort=str(self.targetPort)) if self.cookie: sock.sendall(VER + SUCCESS + "\x00" + "\x01" + serverIp + chr(self.targetPort / 256) + chr( self.targetPort % 256)) return True else: sock.sendall(VER + REFUSED + "\x00" + "\x01" + serverIp + chr(self.targetPort / 256) + chr( self.targetPort % 256)) return False def handleSocks(self, sock): # 通过proxychain模拟客户端发送数据,第一个字节可以判断是socks5还是socks4 ver = sock.recv(1) # 05:02:00:02 if ver == "\x05": return self.parseSocks5(sock) def setupRemoteSession(self, target, targetPort): """探测端口存活""" header = ({"X-CMD": "CONNECT", "X-TARGET": target, "X-PORT": targetPort}) cookie = None try: response = requests.post(url=self.connectString, headers=header, data=None, timeout=TIMEOUT) except Exception, e: return else: if response:
elif response_header.get("X-ERROR"): logger.error(response_header.get("X-ERROR")) else: logger.error("[%s:%s] HTTP [%d]" % (target, targetPort, response.status_code)) return cookie def closeRemoteSession(self): header = {"X-CMD": "DISCONNECT", "Cookie": self.cookie} try: response = requests.post(url=self.connectString, headers=header, data=None, timeout=TIMEOUT) except Exception, e: logger.error("Close Connection Failure") else: if response.status_code == 200: logger.info("[%s:%d] Connection Terminated" % (self.httpHost, self.httpPort)) def reader(self): while True: try: if not self.pSocket: break header = {"X-CMD": "READ", "Cookie": self.cookie, "Connection": "Keep-Alive"} response = requests.post(url=self.connectString, headers=header, data=None) response_data = None if response.status_code == 200: response_header = response.headers status = response_header.get("x-status") if status == "OK": response_data = response.content else: logger.error("[%s:%d] HTTP [%d]: Status: [%s]: Message [%s] Shutting down" % ( self.target, self.targetPort, response.status_code, status, response_header.get("X-ERROR"))) else: logger.error( "[%s:%d] HTTP [%d]: Shutting down" % (self.target, self.targetPort, response.status_code)) if response_data is None: # Remote socket closed break if len(response_data) == 0: time.sleep(0.1) continue self.pSocket.send(response_data) except Exception, ex: print(format_exc()) raise ex self.closeRemoteSession() logger.debug("[%s:%d] Closing localsocket" % (self.target, self.targetPort)) try: self.pSocket.close() except: logger.debug("[%s:%d] Localsocket already closed" % (self.target, self.targetPort)) def writer(self): global READBUFSIZE while True: try: self.pSocket.settimeout(1) # 'GET / HTTP/1.1\r\nHost: 192.168.2.1\r\nUser-Agent: curl/7.58.0\r\nAccept: */*\r\n\r\n' data = self.pSocket.recv(READBUFSIZE) if not data: break header = {"X-CMD": "FORWARD", "Cookie": self.cookie, "Content-Type": "application/octet-stream", "Connection": "Keep-Alive"} # 携带数据 response = requests.post(url=self.connectString, headers=header, data=data) if response.status_code == 200: response_header = response.headers status = response_header.get("x-status") if status == "OK": if response_header.get("set-cookie") is not None: self.cookie = response_header.get("set-cookie") else: logger.error("[%s:%d] HTTP [%d]: Status: [%s]: Message [%s] Shutting down" % ( self.target, self.targetPort, response.status_code, status, response_header.get("x-error"))) break else: logger.error( "[%s:%d] HTTP [%d]: Shutting down" % (self.target, self.targetPort, response.status_code)) break # transferLog.info("[%s:%d] >>>> [%d]" % (self.target, self.port, len(data))) except timeout: continue except Exception, ex: raise ex self.closeRemoteSession() logger.debug("Closing local
response_header = response.headers if response.status_code == 200 and response_header.get("X-STATUS") == "OK": cookie = response_header.get("Set-Cookie") logger.info("[%s:%s] HTTP [200]: cookie [%s]" % (target, targetPort, cookie))
random_line_split
w_reGeorgSocksProxy.py
SocksProtocolNotImplemented(Exception): pass class RemoteConnectionFailed(Exception): pass class session(Thread): def __init__(self, pSocket, connectString): Thread.__init__(self) self.pSocket = pSocket self.connectString = connectString o = urlparse(connectString) try: self.httpPort = o.port except: if o.scheme == "https": self.httpPort = 443 else: self.httpPort = 80 else: if not o.port: if o.scheme == "https": self.httpPort = 443 else: self.httpPort = 80 self.httpScheme = o.scheme self.httpHost = o.netloc.split(":")[0] self.httpPath = o.path self.cookie = None def parseSocks5(self, sock): logger.debug("SocksVersion5 detected") # 02:00 nmethods, methods = (sock.recv(1), sock.recv(1)) # 05:00 sock.sendall(VER + METHOD) # :02 ver = sock.recv(1) if ver == "\x02": # this is a hack for proxychains # 05:01:00:01----:c0:a8:01:02:00:50 # '\x05', '\x01', '\x00', '\x01' ver, cmd, rsv, atyp = (sock.recv(1), sock.recv(1), sock.recv(1), sock.recv(1)) else: cmd, rsv, atyp = (sock.recv(1), sock.recv(1), sock.recv(1)) target = None targetPort = None if atyp == "\x01": # IPv4 # Reading 6 bytes for the IP and Port # c0:a8:01:02 target = sock.recv(4) # 00:50 targetPort = sock.recv(2) # 目标地址192.168.2.1 self.target = ".".join([str(ord(i)) for i in target]) elif atyp == "\x03": # Hostname targetLen = ord(sock.recv(1)) # hostname length (1 byte) target = sock.recv(targetLen) targetPort = sock.recv(2) target = "".join([unichr(ord(i)) for i in target]) elif atyp == "\x04": # IPv6 target = sock.recv(16) targetPort = sock.recv(2) tmp_addr = [] for i in xrange(len(target) / 2): tmp_addr.append(unichr(ord(target[2 * i]) * 256 + ord(target[2 * i + 1]))) target = ":".join(tmp_addr) # 80 self.targetPort = ord(targetPort[0]) * 256 + ord(targetPort[1]) if cmd == "\x02": # BIND raise SocksCmdNotImplemented("Socks5 - BIND not implemented") elif cmd == "\x03": # UDP raise SocksCmdNotImplemented("Socks5 - UDP not implemented") elif cmd == "\x01": # CONNECT serverIp = target try: serverIp = gethostbyname(self.target) except: logger.error("oeps") # 又转回来\xc0\xa8\x02\x01 serverIp = "".join([chr(int(i)) for i in serverIp.split(".")]) # 获取cookie,在服务端的脚本中,会执行相应端口探测 self.cookie = self.setupRemoteSession(target=self.target, targetPort=str(self.targetPort)) if self.cookie: sock.sendall(VER + SUCCESS + "\x00" + "\x01" + serverIp + chr(self.targetPort / 256) + chr( self.targetPort % 256)) return True else: sock.sendall(VER + REFUSED + "\x00" + "\x01" + serverIp + chr(self.targetPort / 256) + chr( self.targetPort % 256)) return False def handleSocks(self, sock): # 通过proxychain模拟客户端发送数据,第一个字节可以判断是socks5还是socks4 ver = sock.recv(1) # 05:02:00:02 if ver == "\x05": return self.parseSocks5(sock) def setupRemoteSession(self, target, targetPort): """探测端口存活""" header = ({"X-CMD": "CONNECT", "X-TARGET": target, "X-PORT": targetPort}) cookie = None try: response = requests.post(url=self.connectString, headers=header, data=None, timeout=TIMEOUT) except Exception, e: return else: if response: response_header = response.headers if response.status_code == 200 and response_header.get("X-STATUS") == "OK": cookie = response_header.get("Set-Cookie") logger.info("[%s:%s] HTTP [200]: cookie [%s]" % (target, targetPort, cookie)) elif response_header.get("X-ERROR"): logger.error(response_header.get("X-ERROR")) else: logger.error("[%s:%s] HTTP [%d]" % (target, targetPort, response.status_code)) return cookie def closeRemoteSession(self): header = {"X-CMD": "DISCONNECT", "Cookie": self.cookie} try: response = requests.post(url=self.connectString, headers=header, data=None, timeout=TIMEOUT) except Exception, e: logger.error("Close Connection Failure") else: if response.status_code == 200: logger.info("[%s:%d] Connection Terminated" % (self.httpHost, self.httpPort)) def reader(self): while True: try: if not self.pSocket: break header = {"X-CMD": "READ", "Cookie": self.cookie, "Connection": "Keep-Alive"} response = requests.post(url=self.connectString, headers=header, data=None) response_data = None if response.status_code == 200: response_header = response.headers status = response_header.get("x-status") if status == "OK": response_data = response.content else: logger.error("[%s:%d] HTTP [%d]: Status: [%s]: Message [%s] Shutting down" % ( self.target, self.targetPort, response.status_code, status, response_header.get("X-ERROR"))) else: logger.error( "[%s:%d] HTTP [%d]: Shutting down" % (self.target, self.targetPort, response.status_code)) if response_data is None: # Remote socket closed break if len(response_data) == 0: time.sleep(0.1) continue self.pSocket.send(response_data) except Exception, ex: print(format_exc()) raise ex self.closeRemoteSession() logger.debug("[%s:%d] Closing localsocket" % (self.target, self.targetPort)) try: self.pSocket.close() except: logger.debug("[%s:%d] Localsocket already closed" % (self.target, self.targetPort)) def writer(self): global READBUFSIZE while True: try: self.pSocket.settimeout(1) # 'GET / HTTP/1.1\r\nHost: 192.168.2.1\r\nUser-Agent: curl/7.58.0\r\nAccept: */*\r\n\r\n' data = self.pSocket.recv(READBUFSIZE) if not data: break header = {"X-CMD": "FORWARD", "Cookie": self.cookie, "Content-Type": "application/octet-stream", "Connection": "Keep-Alive"} # 携带数据 response = requests.post(url=self.connectString, headers=header, data=data) if response.status_code == 200: response_header = response.headers status = response_header.get("x-status") if status == "OK": if response_header.get("set-cookie") is not None: self.cookie = response_header.get("set-cookie") else: logger.error("[%s:%d] HTTP [%d]: Status: [%s]: Message [%s] Shutting down" % ( self.target, self.tar
target, self.targetPort, response.status_code)) break # transferLog.info("[%s:%d] >>>> [%d]" % (self.target, self.port, len(data))) except timeout: continue except Exception, ex: raise ex self.closeRemoteSession() logger.debug("Closing
getPort, response.status_code, status, response_header.get("x-error"))) break else: logger.error( "[%s:%d] HTTP [%d]: Shutting down" % (self.
conditional_block
w_reGeorgSocksProxy.py
80 self.httpScheme = o.scheme self.httpHost = o.netloc.split(":")[0] self.httpPath = o.path self.cookie = None def parseSocks5(self, sock): logger.debug("SocksVersion5 detected") # 02:00 nmethods, methods = (sock.recv(1), sock.recv(1)) # 05:00 sock.sendall(VER + METHOD) # :02 ver = sock.recv(1) if ver == "\x02": # this is a hack for proxychains # 05:01:00:01----:c0:a8:01:02:00:50 # '\x05', '\x01', '\x00', '\x01' ver, cmd, rsv, atyp = (sock.recv(1), sock.recv(1), sock.recv(1), sock.recv(1)) else: cmd, rsv, atyp = (sock.recv(1), sock.recv(1), sock.recv(1)) target = None targetPort = None if atyp == "\x01": # IPv4 # Reading 6 bytes for the IP and Port # c0:a8:01:02 target = sock.recv(4) # 00:50 targetPort = sock.recv(2) # 目标地址192.168.2.1 self.target = ".".join([str(ord(i)) for i in target]) elif atyp == "\x03": # Hostname targetLen = ord(sock.recv(1)) # hostname length (1 byte) target = sock.recv(targetLen) targetPort = sock.recv(2) target = "".join([unichr(ord(i)) for i in target]) elif atyp == "\x04": # IPv6 target = sock.recv(16) targetPort = sock.recv(2) tmp_addr = [] for i in xrange(len(target) / 2): tmp_addr.append(unichr(ord(target[2 * i]) * 256 + ord(target[2 * i + 1]))) target = ":".join(tmp_addr) # 80 self.targetPort = ord(targetPort[0]) * 256 + ord(targetPort[1]) if cmd == "\x02": # BIND raise SocksCmdNotImplemented("Socks5 - BIND not implemented") elif cmd == "\x03": # UDP raise SocksCmdNotImplemented("Socks5 - UDP not implemented") elif cmd == "\x01": # CONNECT serverIp = target try: serverIp = gethostbyname(self.target) except: logger.error("oeps") # 又转回来\xc0\xa8\x02\x01 serverIp = "".join([chr(int(i)) for i in serverIp.split(".")]) # 获取cookie,在服务端的脚本中,会执行相应端口探测 self.cookie = self.setupRemoteSession(target=self.target, targetPort=str(self.targetPort)) if self.cookie: sock.sendall(VER + SUCCESS + "\x00" + "\x01" + serverIp + chr(self.targetPort / 256) + chr( self.targetPort % 256)) return True else: sock.sendall(VER + REFUSED + "\x00" + "\x01" + serverIp + chr(self.targetPort / 256) + chr( self.targetPort % 256)) return False def handleSocks(self, sock): # 通过proxychain模拟客户端发送数据,第一个字节可以判断是socks5还是socks4 ver = sock.recv(1) # 05:02:00:02 if ver == "\x05": return self.parseSocks5(sock) def setupRemoteSession(self, target, targetPort): """探测端口存活""" header = ({"X-CMD": "CONNECT", "X-TARGET": target, "X-PORT": targetPort}) cookie = None try: response = requests.post(url=self.connectString, headers=header, data=None, timeout=TIMEOUT) except Exception, e: return else: if response: response_header = response.headers if response.status_code == 200 and response_header.get("X-STATUS") == "OK": cookie = response_header.get("Set-Cookie") logger.info("[%s:%s] HTTP [200]: cookie [%s]" % (target, targetPort, cookie)) elif response_header.get("X-ERROR"): logger.error(response_header.get("X-ERROR")) else: logger.error("[%s:%s] HTTP [%d]" % (target, targetPort, response.status_code)) return cookie def closeRemoteSession(self): header = {"X-CMD": "DISCONNECT", "Cookie": self.cookie} try: response = requests.post(url=self.connectString, headers=header, data=None, timeout=TIMEOUT) except Exception, e: logger.error("Close Connection Failure") else: if response.status_code == 200: logger.info("[%s:%d] Connection Terminated" % (self.httpHost, self.httpPort)) def reader(self): while True: try: if not self.pSocket: break header = {"X-CMD": "READ", "Cookie": self.cookie, "Connection": "Keep-Alive"} response = requests.post(url=self.connectString, headers=header, data=None) response_data = None if response.status_code == 200: response_header = response.headers status = response_header.get("x-status") if status == "OK": response_data = response.content else: logger.error("[%s:%d] HTTP [%d]: Status: [%s]: Message [%s] Shutting down" % ( self.target, self.targetPort, response.status_code, status, response_header.get("X-ERROR"))) else: logger.error( "[%s:%d] HTTP [%d]: Shutting down" % (self.target, self.targetPort, response.status_code)) if response_data is None: # Remote socket closed break if len(response_data) == 0: time.sleep(0.1) continue self.pSocket.send(response_data) except Exception, ex: print(format_exc()) raise ex self.closeRemoteSession() logger.debug("[%s:%d] Closing localsocket" % (self.target, self.targetPort)) try: self.pSocket.close() except: logger.debug("[%s:%d] Localsocket already closed" % (self.target, self.targetPort)) def writer(self): global READBUFSIZE while True: try: self.pSocket.settimeout(1) # 'GET / HTTP/1.1\r\nHost: 192.168.2.1\r\nUser-Agent: curl/7.58.0\r\nAccept: */*\r\n\r\n' data = self.pSocket.recv(READBUFSIZE) if not data: break header = {"X-CMD": "FORWARD", "Cookie": self.cookie, "Content-Type": "application/octet-stream", "Connection": "Keep-Alive"} # 携带数据 response = requests.post(url=self.connectString, headers=header, data=data) if response.status_code == 200: response_header = response.headers status = response_header.get("x-status") if status == "OK": if response_header.get("set-cookie") is not None: self.cookie = response_header.get("set-cookie") else: logger.error("[%s:%d] HTTP [%d]: Status: [%s]: Message [%s] Shutting down" % ( self.target, self.targetPort, response.status_code, status, response_header.get("x-error"))) break else: logger.error( "[%s:%d] HTTP [%d]: Shutting down" % (self.target, self.targetPort, response.status_code)) break # transferLog.info("[%s:%d] >>>> [%d]" % (self.target, self.port, len(data))) except timeout: continue except Exception, ex: raise ex self.closeRemoteSession() logger.debug("Closing localsocket") try: self.pSocket.close() except: logger.debug("Localsocket already closed") def run(self): try: if self.handleSocks(self.pSocket): r = Thread(target=self.reader, args=())
r.start() w = Thread(target=self.writer, args=()) w.start() w.join() r.join() except Exception, e: # 报错关闭连接 logger.error(format_exc()) self.closeRemoteSession() self.pSocket.close() def askgeorg(url): """检测reg连接方法""" try: response = requests.get(url=url, headers=HEADER, timeout=TIMEOUT) except
identifier_body
activity_heartbeat_manager.rs
<JoinHandle<()>>>, } /// Used to supply heartbeat details to the heartbeat processor, which periodically sends them to /// the server. struct ActivityHeartbeatProcessorHandle { heartbeat_tx: Sender<Vec<common::Payload>>, join_handle: JoinHandle<()>, } /// Heartbeat processor, that aggregates and periodically sends heartbeat requests for a single /// activity to the server. struct ActivityHeartbeatProcessor<SG> { task_token: TaskToken, delay: time::Duration, /// Used to receive heartbeat events. heartbeat_rx: Receiver<Vec<common::Payload>>, /// Used to receive shutdown notifications. shutdown_rx: Receiver<bool>, /// Used to send CleanupProcessor event at the end of the processor loop. events_tx: UnboundedSender<LifecycleEvent>, /// Used to send cancellation notices that we learned about when heartbeating back up to core cancels_tx: UnboundedSender<TaskToken>, server_gateway: Arc<SG>, } #[derive(Debug)] pub enum LifecycleEvent { Heartbeat(ValidActivityHeartbeat), CleanupProcessor(TaskToken), Shutdown, } #[derive(Debug)] pub struct ValidActivityHeartbeat { pub task_token: TaskToken, pub details: Vec<common::Payload>, pub delay: time::Duration, } /// Handle that is used by the core for all interactions with the manager, allows sending new /// heartbeats or requesting and awaiting for the shutdown. When shutdown is requested, signal gets /// sent to all processors, which allows them to complete gracefully. impl ActivityHeartbeatManagerHandle { /// Records a new heartbeat, note that first call would result in an immediate call to the /// server, while rapid successive calls would accumulate for up to `delay` /// and then latest heartbeat details will be sent to the server. If there is no activity for /// `delay` then heartbeat processor will be reset and process would start /// over again, meaning that next heartbeat will be sent immediately, creating a new processor. pub fn record( &self, details: ActivityHeartbeat, delay: Duration, ) -> Result<(), ActivityHeartbeatError> { if self.shutting_down.load(Ordering::Relaxed) { return Err(ActivityHeartbeatError::ShuttingDown); } self.events .send(LifecycleEvent::Heartbeat(ValidActivityHeartbeat { task_token: TaskToken(details.task_token), details: details.details, delay, })) .expect("Receive half of the heartbeats event channel must not be dropped"); Ok(()) } /// Returns a future that resolves any time there is a new activity cancel that must be /// dispatched to lang pub async fn next_pending_cancel(&self) -> Option<TaskToken> { self.incoming_cancels.lock().await.recv().await } /// Initiates shutdown procedure by stopping lifecycle loop and awaiting for all heartbeat
if !self.shutting_down.load(Ordering::Relaxed) { self.events .send(LifecycleEvent::Shutdown) .expect("should be able to send shutdown event"); self.shutting_down.store(true, Ordering::Relaxed); } let mut handle = self.join_handle.lock().await; if let Some(h) = handle.take() { h.await.expect("shutdown should exit cleanly"); } } } impl<SG: ServerGatewayApis + Send + Sync + 'static> ActivityHeartbeatManager<SG> { #![allow(clippy::new_ret_no_self)] /// Creates a new instance of an activity heartbeat manager and returns a handle to the user, /// which allows to send new heartbeats and initiate the shutdown. pub fn new(sg: Arc<SG>) -> ActivityHeartbeatManagerHandle { let (shutdown_tx, shutdown_rx) = channel(false); let (events_tx, events_rx) = unbounded_channel(); let (cancels_tx, cancels_rx) = unbounded_channel(); let s = Self { heartbeat_processors: Default::default(), events_tx: events_tx.clone(), events_rx, shutdown_tx, shutdown_rx, cancels_tx, server_gateway: sg, }; let join_handle = tokio::spawn(s.lifecycle()); ActivityHeartbeatManagerHandle { shutting_down: AtomicBool::new(false), events: events_tx, incoming_cancels: Mutex::new(cancels_rx), join_handle: Mutex::new(Some(join_handle)), } } /// Main loop, that handles all heartbeat requests and dispatches them to processors. async fn lifecycle(mut self) { while let Some(event) = self.events_rx.recv().await { match event { LifecycleEvent::Heartbeat(heartbeat) => self.record(heartbeat), LifecycleEvent::Shutdown => break, LifecycleEvent::CleanupProcessor(task_token) => { self.heartbeat_processors.remove(&task_token); } } } self.shutdown().await.expect("shutdown should exit cleanly") } /// Records heartbeat, by sending it to the processor. /// New processor is created if one doesn't exist, otherwise new event is dispatched to the /// existing processor's receiver channel. fn record(&mut self, heartbeat: ValidActivityHeartbeat) { match self.heartbeat_processors.get(&heartbeat.task_token) { Some(handle) => { handle .heartbeat_tx .send(heartbeat.details) .expect("heartbeat channel can't be dropped if we are inside this method"); } None => { let (heartbeat_tx, heartbeat_rx) = channel(heartbeat.details); let processor = ActivityHeartbeatProcessor { task_token: heartbeat.task_token.clone(), delay: heartbeat.delay, heartbeat_rx, shutdown_rx: self.shutdown_rx.clone(), events_tx: self.events_tx.clone(), cancels_tx: self.cancels_tx.clone(), server_gateway: self.server_gateway.clone(), }; let join_handle = tokio::spawn(processor.run()); let handle = ActivityHeartbeatProcessorHandle { heartbeat_tx, join_handle, }; self.heartbeat_processors .insert(heartbeat.task_token, handle); } } } /// Initiates termination of all heartbeat processors by sending a signal and awaits termination pub async fn shutdown(mut self) -> Result<(), JoinError> { self.shutdown_tx .send(true) .expect("shutdown channel can't be dropped before shutdown is complete"); for v in self.heartbeat_processors.drain() { v.1.join_handle.await?; } Ok(()) } } impl<SG: ServerGatewayApis + Send + Sync + 'static> ActivityHeartbeatProcessor<SG> { async fn run(mut self) { // Each processor is initialized with heartbeat payloads, first thing we need to do is send // it out. self.record_heartbeat().await; loop { sleep(self.delay).await; select! { biased; _ = self.heartbeat_rx.changed() => { self.record_heartbeat().await; } _ = self.shutdown_rx.changed() => { break; } _ = sleep(self.delay) => { // Timed out while waiting for the next heartbeat. We waited 2 * delay in total, // where delay is 1/2 of the activity heartbeat timeout. This means that // activity has either timed out or completed by now. break; } }; } self.events_tx .send(LifecycleEvent::CleanupProcessor(self.task_token)) .expect("cleanup requests should not be dropped"); } async fn record_heartbeat(&mut self) { let details = self.heartbeat_rx.borrow().clone(); match self .server_gateway .record_activity_heartbeat(self.task_token.clone(), details.into_payloads()) .await { Ok(RecordActivityTaskHeartbeatResponse { cancel_requested }) => { if cancel_requested { self.cancels_tx .send(self.task_token.clone()) .expect("Receive half of heartbeat cancels not blocked"); } } // Send cancels for any activity that learns its workflow already finished (which is // one thing not found implies - other reasons would seem equally valid). Err(s) if s.code() == tonic::Code::NotFound => { self.cancels_tx .send(self.task_token.clone()) .expect("Receive half of heartbeat cancels not blocked"); } Err(e) => { warn!("Error when recording heartbeat: {:?}", e) } } } } #[cfg(test)] mod test { use super::*; use crate::pollers::MockServerGatewayApis; use crate::protos::coresdk::common::Payload; use crate::protos::temporal::api::workflowservice::v1::RecordActivityTaskHeartbeatResponse; use std::time::Duration; /// Ensure that hearbeats that are sent with a small delay are aggregated and sent roughly once /// every 1/2 of the heartbeat timeout. #[tokio::test] async fn process_heartbeats_and_shutdown() { let mut mock_gateway = MockServerGatewayApis::new(); mock_gateway .expect_record_activity_heartbeat() .returning(|_, _| Ok(RecordActivityTaskHeart
/// processors to terminate gracefully. pub async fn shutdown(&self) { // If shutdown was called multiple times, shutdown signal has been sent already and consumer // might have been dropped already, meaning that sending to the channel may fail. // All we need to do is to simply await on handle for the completion.
random_line_split
activity_heartbeat_manager.rs
_self)] /// Creates a new instance of an activity heartbeat manager and returns a handle to the user, /// which allows to send new heartbeats and initiate the shutdown. pub fn new(sg: Arc<SG>) -> ActivityHeartbeatManagerHandle { let (shutdown_tx, shutdown_rx) = channel(false); let (events_tx, events_rx) = unbounded_channel(); let (cancels_tx, cancels_rx) = unbounded_channel(); let s = Self { heartbeat_processors: Default::default(), events_tx: events_tx.clone(), events_rx, shutdown_tx, shutdown_rx, cancels_tx, server_gateway: sg, }; let join_handle = tokio::spawn(s.lifecycle()); ActivityHeartbeatManagerHandle { shutting_down: AtomicBool::new(false), events: events_tx, incoming_cancels: Mutex::new(cancels_rx), join_handle: Mutex::new(Some(join_handle)), } } /// Main loop, that handles all heartbeat requests and dispatches them to processors. async fn lifecycle(mut self) { while let Some(event) = self.events_rx.recv().await { match event { LifecycleEvent::Heartbeat(heartbeat) => self.record(heartbeat), LifecycleEvent::Shutdown => break, LifecycleEvent::CleanupProcessor(task_token) => { self.heartbeat_processors.remove(&task_token); } } } self.shutdown().await.expect("shutdown should exit cleanly") } /// Records heartbeat, by sending it to the processor. /// New processor is created if one doesn't exist, otherwise new event is dispatched to the /// existing processor's receiver channel. fn record(&mut self, heartbeat: ValidActivityHeartbeat) { match self.heartbeat_processors.get(&heartbeat.task_token) { Some(handle) => { handle .heartbeat_tx .send(heartbeat.details) .expect("heartbeat channel can't be dropped if we are inside this method"); } None => { let (heartbeat_tx, heartbeat_rx) = channel(heartbeat.details); let processor = ActivityHeartbeatProcessor { task_token: heartbeat.task_token.clone(), delay: heartbeat.delay, heartbeat_rx, shutdown_rx: self.shutdown_rx.clone(), events_tx: self.events_tx.clone(), cancels_tx: self.cancels_tx.clone(), server_gateway: self.server_gateway.clone(), }; let join_handle = tokio::spawn(processor.run()); let handle = ActivityHeartbeatProcessorHandle { heartbeat_tx, join_handle, }; self.heartbeat_processors .insert(heartbeat.task_token, handle); } } } /// Initiates termination of all heartbeat processors by sending a signal and awaits termination pub async fn shutdown(mut self) -> Result<(), JoinError> { self.shutdown_tx .send(true) .expect("shutdown channel can't be dropped before shutdown is complete"); for v in self.heartbeat_processors.drain() { v.1.join_handle.await?; } Ok(()) } } impl<SG: ServerGatewayApis + Send + Sync + 'static> ActivityHeartbeatProcessor<SG> { async fn run(mut self) { // Each processor is initialized with heartbeat payloads, first thing we need to do is send // it out. self.record_heartbeat().await; loop { sleep(self.delay).await; select! { biased; _ = self.heartbeat_rx.changed() => { self.record_heartbeat().await; } _ = self.shutdown_rx.changed() => { break; } _ = sleep(self.delay) => { // Timed out while waiting for the next heartbeat. We waited 2 * delay in total, // where delay is 1/2 of the activity heartbeat timeout. This means that // activity has either timed out or completed by now. break; } }; } self.events_tx .send(LifecycleEvent::CleanupProcessor(self.task_token)) .expect("cleanup requests should not be dropped"); } async fn record_heartbeat(&mut self) { let details = self.heartbeat_rx.borrow().clone(); match self .server_gateway .record_activity_heartbeat(self.task_token.clone(), details.into_payloads()) .await { Ok(RecordActivityTaskHeartbeatResponse { cancel_requested }) => { if cancel_requested { self.cancels_tx .send(self.task_token.clone()) .expect("Receive half of heartbeat cancels not blocked"); } } // Send cancels for any activity that learns its workflow already finished (which is // one thing not found implies - other reasons would seem equally valid). Err(s) if s.code() == tonic::Code::NotFound => { self.cancels_tx .send(self.task_token.clone()) .expect("Receive half of heartbeat cancels not blocked"); } Err(e) => { warn!("Error when recording heartbeat: {:?}", e) } } } } #[cfg(test)] mod test { use super::*; use crate::pollers::MockServerGatewayApis; use crate::protos::coresdk::common::Payload; use crate::protos::temporal::api::workflowservice::v1::RecordActivityTaskHeartbeatResponse; use std::time::Duration; /// Ensure that hearbeats that are sent with a small delay are aggregated and sent roughly once /// every 1/2 of the heartbeat timeout. #[tokio::test] async fn process_heartbeats_and_shutdown() { let mut mock_gateway = MockServerGatewayApis::new(); mock_gateway .expect_record_activity_heartbeat() .returning(|_, _| Ok(RecordActivityTaskHeartbeatResponse::default())) .times(2); let hm = ActivityHeartbeatManager::new(Arc::new(mock_gateway)); let fake_task_token = vec![1, 2, 3]; // Sending heartbeat requests for 400ms, this should send first hearbeat right away, and all other // requests should be aggregated and last one should be sent to the server in 500ms (1/2 of heartbeat timeout). for i in 0u8..40 { sleep(Duration::from_millis(10)).await; record_heartbeat(&hm, fake_task_token.clone(), i, Duration::from_millis(1000)); } hm.shutdown().await; } /// Ensure that heartbeat can be called from a tight loop without any delays, resulting in two /// interactions with the server - one immediately and one after 500ms after the delay. #[tokio::test] async fn process_tight_loop_and_shutdown() { let mut mock_gateway = MockServerGatewayApis::new(); mock_gateway .expect_record_activity_heartbeat() .returning(|_, _| Ok(RecordActivityTaskHeartbeatResponse::default())) .times(2); let hm = ActivityHeartbeatManager::new(Arc::new(mock_gateway)); let fake_task_token = vec![1, 2, 3]; // Sending heartbeat requests for 400ms, this should send first hearbeat right away, and all other // requests should be aggregated and last one should be sent to the server in 500ms (1/2 of heartbeat timeout). for i in 0u8..u8::MAX { record_heartbeat(&hm, fake_task_token.clone(), i, Duration::from_millis(1000)); } hm.shutdown().await; } /// This test reports one heartbeat and waits until processor times out and exits then sends another one. /// Expectation is that new processor should be spawned and heartbeat shouldn't get lost. #[tokio::test] async fn report_heartbeat_after_timeout() { let mut mock_gateway = MockServerGatewayApis::new(); mock_gateway .expect_record_activity_heartbeat() .returning(|_, _| Ok(RecordActivityTaskHeartbeatResponse::default())) .times(2); let hm = ActivityHeartbeatManager::new(Arc::new(mock_gateway)); let fake_task_token = vec![1, 2, 3]; record_heartbeat(&hm, fake_task_token.clone(), 0, Duration::from_millis(100)); sleep(Duration::from_millis(500)).await; record_heartbeat(&hm, fake_task_token.clone(), 1, Duration::from_millis(100)); hm.shutdown().await; } /// Recording new heartbeats after shutdown is not allowed, and will result in error. #[tokio::test] async fn record_after_shutdown()
{ let mut mock_gateway = MockServerGatewayApis::new(); mock_gateway .expect_record_activity_heartbeat() .returning(|_, _| Ok(RecordActivityTaskHeartbeatResponse::default())) .times(0); let hm = ActivityHeartbeatManager::new(Arc::new(mock_gateway)); hm.shutdown().await; match hm.record( ActivityHeartbeat { task_token: vec![1, 2, 3], details: vec![Payload { // payload doesn't matter in this case, as it shouldn't get sent anyways. ..Default::default() }], }, Duration::from_millis(1000), ) { Ok(_) => { unreachable!("heartbeat should not be recorded after the shutdown")
identifier_body
activity_heartbeat_manager.rs
JoinHandle<()>>>, } /// Used to supply heartbeat details to the heartbeat processor, which periodically sends them to /// the server. struct ActivityHeartbeatProcessorHandle { heartbeat_tx: Sender<Vec<common::Payload>>, join_handle: JoinHandle<()>, } /// Heartbeat processor, that aggregates and periodically sends heartbeat requests for a single /// activity to the server. struct ActivityHeartbeatProcessor<SG> { task_token: TaskToken, delay: time::Duration, /// Used to receive heartbeat events. heartbeat_rx: Receiver<Vec<common::Payload>>, /// Used to receive shutdown notifications. shutdown_rx: Receiver<bool>, /// Used to send CleanupProcessor event at the end of the processor loop. events_tx: UnboundedSender<LifecycleEvent>, /// Used to send cancellation notices that we learned about when heartbeating back up to core cancels_tx: UnboundedSender<TaskToken>, server_gateway: Arc<SG>, } #[derive(Debug)] pub enum LifecycleEvent { Heartbeat(ValidActivityHeartbeat), CleanupProcessor(TaskToken), Shutdown, } #[derive(Debug)] pub struct ValidActivityHeartbeat { pub task_token: TaskToken, pub details: Vec<common::Payload>, pub delay: time::Duration, } /// Handle that is used by the core for all interactions with the manager, allows sending new /// heartbeats or requesting and awaiting for the shutdown. When shutdown is requested, signal gets /// sent to all processors, which allows them to complete gracefully. impl ActivityHeartbeatManagerHandle { /// Records a new heartbeat, note that first call would result in an immediate call to the /// server, while rapid successive calls would accumulate for up to `delay` /// and then latest heartbeat details will be sent to the server. If there is no activity for /// `delay` then heartbeat processor will be reset and process would start /// over again, meaning that next heartbeat will be sent immediately, creating a new processor. pub fn record( &self, details: ActivityHeartbeat, delay: Duration, ) -> Result<(), ActivityHeartbeatError> { if self.shutting_down.load(Ordering::Relaxed) { return Err(ActivityHeartbeatError::ShuttingDown); } self.events .send(LifecycleEvent::Heartbeat(ValidActivityHeartbeat { task_token: TaskToken(details.task_token), details: details.details, delay, })) .expect("Receive half of the heartbeats event channel must not be dropped"); Ok(()) } /// Returns a future that resolves any time there is a new activity cancel that must be /// dispatched to lang pub async fn next_pending_cancel(&self) -> Option<TaskToken> { self.incoming_cancels.lock().await.recv().await } /// Initiates shutdown procedure by stopping lifecycle loop and awaiting for all heartbeat /// processors to terminate gracefully. pub async fn shutdown(&self) { // If shutdown was called multiple times, shutdown signal has been sent already and consumer // might have been dropped already, meaning that sending to the channel may fail. // All we need to do is to simply await on handle for the completion. if !self.shutting_down.load(Ordering::Relaxed) { self.events .send(LifecycleEvent::Shutdown) .expect("should be able to send shutdown event"); self.shutting_down.store(true, Ordering::Relaxed); } let mut handle = self.join_handle.lock().await; if let Some(h) = handle.take() { h.await.expect("shutdown should exit cleanly"); } } } impl<SG: ServerGatewayApis + Send + Sync + 'static> ActivityHeartbeatManager<SG> { #![allow(clippy::new_ret_no_self)] /// Creates a new instance of an activity heartbeat manager and returns a handle to the user, /// which allows to send new heartbeats and initiate the shutdown. pub fn new(sg: Arc<SG>) -> ActivityHeartbeatManagerHandle { let (shutdown_tx, shutdown_rx) = channel(false); let (events_tx, events_rx) = unbounded_channel(); let (cancels_tx, cancels_rx) = unbounded_channel(); let s = Self { heartbeat_processors: Default::default(), events_tx: events_tx.clone(), events_rx, shutdown_tx, shutdown_rx, cancels_tx, server_gateway: sg, }; let join_handle = tokio::spawn(s.lifecycle()); ActivityHeartbeatManagerHandle { shutting_down: AtomicBool::new(false), events: events_tx, incoming_cancels: Mutex::new(cancels_rx), join_handle: Mutex::new(Some(join_handle)), } } /// Main loop, that handles all heartbeat requests and dispatches them to processors. async fn lifecycle(mut self) { while let Some(event) = self.events_rx.recv().await { match event { LifecycleEvent::Heartbeat(heartbeat) => self.record(heartbeat), LifecycleEvent::Shutdown => break, LifecycleEvent::CleanupProcessor(task_token) => { self.heartbeat_processors.remove(&task_token); } } } self.shutdown().await.expect("shutdown should exit cleanly") } /// Records heartbeat, by sending it to the processor. /// New processor is created if one doesn't exist, otherwise new event is dispatched to the /// existing processor's receiver channel. fn record(&mut self, heartbeat: ValidActivityHeartbeat) { match self.heartbeat_processors.get(&heartbeat.task_token) { Some(handle) => { handle .heartbeat_tx .send(heartbeat.details) .expect("heartbeat channel can't be dropped if we are inside this method"); } None => { let (heartbeat_tx, heartbeat_rx) = channel(heartbeat.details); let processor = ActivityHeartbeatProcessor { task_token: heartbeat.task_token.clone(), delay: heartbeat.delay, heartbeat_rx, shutdown_rx: self.shutdown_rx.clone(), events_tx: self.events_tx.clone(), cancels_tx: self.cancels_tx.clone(), server_gateway: self.server_gateway.clone(), }; let join_handle = tokio::spawn(processor.run()); let handle = ActivityHeartbeatProcessorHandle { heartbeat_tx, join_handle, }; self.heartbeat_processors .insert(heartbeat.task_token, handle); } } } /// Initiates termination of all heartbeat processors by sending a signal and awaits termination pub async fn shutdown(mut self) -> Result<(), JoinError> { self.shutdown_tx .send(true) .expect("shutdown channel can't be dropped before shutdown is complete"); for v in self.heartbeat_processors.drain() { v.1.join_handle.await?; } Ok(()) } } impl<SG: ServerGatewayApis + Send + Sync + 'static> ActivityHeartbeatProcessor<SG> { async fn run(mut self) { // Each processor is initialized with heartbeat payloads, first thing we need to do is send // it out. self.record_heartbeat().await; loop { sleep(self.delay).await; select! { biased; _ = self.heartbeat_rx.changed() => { self.record_heartbeat().await; } _ = self.shutdown_rx.changed() => { break; } _ = sleep(self.delay) => { // Timed out while waiting for the next heartbeat. We waited 2 * delay in total, // where delay is 1/2 of the activity heartbeat timeout. This means that // activity has either timed out or completed by now. break; } }; } self.events_tx .send(LifecycleEvent::CleanupProcessor(self.task_token)) .expect("cleanup requests should not be dropped"); } async fn record_heartbeat(&mut self) { let details = self.heartbeat_rx.borrow().clone(); match self .server_gateway .record_activity_heartbeat(self.task_token.clone(), details.into_payloads()) .await { Ok(RecordActivityTaskHeartbeatResponse { cancel_requested }) => { if cancel_requested { self.cancels_tx .send(self.task_token.clone()) .expect("Receive half of heartbeat cancels not blocked"); } } // Send cancels for any activity that learns its workflow already finished (which is // one thing not found implies - other reasons would seem equally valid). Err(s) if s.code() == tonic::Code::NotFound =>
Err(e) => { warn!("Error when recording heartbeat: {:?}", e) } } } } #[cfg(test)] mod test { use super::*; use crate::pollers::MockServerGatewayApis; use crate::protos::coresdk::common::Payload; use crate::protos::temporal::api::workflowservice::v1::RecordActivityTaskHeartbeatResponse; use std::time::Duration; /// Ensure that hearbeats that are sent with a small delay are aggregated and sent roughly once /// every 1/2 of the heartbeat timeout. #[tokio::test] async fn process_heartbeats_and_shutdown() { let mut mock_gateway = MockServerGatewayApis::new(); mock_gateway .expect_record_activity_heartbeat() .returning(|_, _| Ok(RecordActivity
{ self.cancels_tx .send(self.task_token.clone()) .expect("Receive half of heartbeat cancels not blocked"); }
conditional_block
activity_heartbeat_manager.rs
JoinHandle<()>>>, } /// Used to supply heartbeat details to the heartbeat processor, which periodically sends them to /// the server. struct ActivityHeartbeatProcessorHandle { heartbeat_tx: Sender<Vec<common::Payload>>, join_handle: JoinHandle<()>, } /// Heartbeat processor, that aggregates and periodically sends heartbeat requests for a single /// activity to the server. struct ActivityHeartbeatProcessor<SG> { task_token: TaskToken, delay: time::Duration, /// Used to receive heartbeat events. heartbeat_rx: Receiver<Vec<common::Payload>>, /// Used to receive shutdown notifications. shutdown_rx: Receiver<bool>, /// Used to send CleanupProcessor event at the end of the processor loop. events_tx: UnboundedSender<LifecycleEvent>, /// Used to send cancellation notices that we learned about when heartbeating back up to core cancels_tx: UnboundedSender<TaskToken>, server_gateway: Arc<SG>, } #[derive(Debug)] pub enum LifecycleEvent { Heartbeat(ValidActivityHeartbeat), CleanupProcessor(TaskToken), Shutdown, } #[derive(Debug)] pub struct ValidActivityHeartbeat { pub task_token: TaskToken, pub details: Vec<common::Payload>, pub delay: time::Duration, } /// Handle that is used by the core for all interactions with the manager, allows sending new /// heartbeats or requesting and awaiting for the shutdown. When shutdown is requested, signal gets /// sent to all processors, which allows them to complete gracefully. impl ActivityHeartbeatManagerHandle { /// Records a new heartbeat, note that first call would result in an immediate call to the /// server, while rapid successive calls would accumulate for up to `delay` /// and then latest heartbeat details will be sent to the server. If there is no activity for /// `delay` then heartbeat processor will be reset and process would start /// over again, meaning that next heartbeat will be sent immediately, creating a new processor. pub fn record( &self, details: ActivityHeartbeat, delay: Duration, ) -> Result<(), ActivityHeartbeatError> { if self.shutting_down.load(Ordering::Relaxed) { return Err(ActivityHeartbeatError::ShuttingDown); } self.events .send(LifecycleEvent::Heartbeat(ValidActivityHeartbeat { task_token: TaskToken(details.task_token), details: details.details, delay, })) .expect("Receive half of the heartbeats event channel must not be dropped"); Ok(()) } /// Returns a future that resolves any time there is a new activity cancel that must be /// dispatched to lang pub async fn next_pending_cancel(&self) -> Option<TaskToken> { self.incoming_cancels.lock().await.recv().await } /// Initiates shutdown procedure by stopping lifecycle loop and awaiting for all heartbeat /// processors to terminate gracefully. pub async fn shutdown(&self) { // If shutdown was called multiple times, shutdown signal has been sent already and consumer // might have been dropped already, meaning that sending to the channel may fail. // All we need to do is to simply await on handle for the completion. if !self.shutting_down.load(Ordering::Relaxed) { self.events .send(LifecycleEvent::Shutdown) .expect("should be able to send shutdown event"); self.shutting_down.store(true, Ordering::Relaxed); } let mut handle = self.join_handle.lock().await; if let Some(h) = handle.take() { h.await.expect("shutdown should exit cleanly"); } } } impl<SG: ServerGatewayApis + Send + Sync + 'static> ActivityHeartbeatManager<SG> { #![allow(clippy::new_ret_no_self)] /// Creates a new instance of an activity heartbeat manager and returns a handle to the user, /// which allows to send new heartbeats and initiate the shutdown. pub fn new(sg: Arc<SG>) -> ActivityHeartbeatManagerHandle { let (shutdown_tx, shutdown_rx) = channel(false); let (events_tx, events_rx) = unbounded_channel(); let (cancels_tx, cancels_rx) = unbounded_channel(); let s = Self { heartbeat_processors: Default::default(), events_tx: events_tx.clone(), events_rx, shutdown_tx, shutdown_rx, cancels_tx, server_gateway: sg, }; let join_handle = tokio::spawn(s.lifecycle()); ActivityHeartbeatManagerHandle { shutting_down: AtomicBool::new(false), events: events_tx, incoming_cancels: Mutex::new(cancels_rx), join_handle: Mutex::new(Some(join_handle)), } } /// Main loop, that handles all heartbeat requests and dispatches them to processors. async fn lifecycle(mut self) { while let Some(event) = self.events_rx.recv().await { match event { LifecycleEvent::Heartbeat(heartbeat) => self.record(heartbeat), LifecycleEvent::Shutdown => break, LifecycleEvent::CleanupProcessor(task_token) => { self.heartbeat_processors.remove(&task_token); } } } self.shutdown().await.expect("shutdown should exit cleanly") } /// Records heartbeat, by sending it to the processor. /// New processor is created if one doesn't exist, otherwise new event is dispatched to the /// existing processor's receiver channel. fn
(&mut self, heartbeat: ValidActivityHeartbeat) { match self.heartbeat_processors.get(&heartbeat.task_token) { Some(handle) => { handle .heartbeat_tx .send(heartbeat.details) .expect("heartbeat channel can't be dropped if we are inside this method"); } None => { let (heartbeat_tx, heartbeat_rx) = channel(heartbeat.details); let processor = ActivityHeartbeatProcessor { task_token: heartbeat.task_token.clone(), delay: heartbeat.delay, heartbeat_rx, shutdown_rx: self.shutdown_rx.clone(), events_tx: self.events_tx.clone(), cancels_tx: self.cancels_tx.clone(), server_gateway: self.server_gateway.clone(), }; let join_handle = tokio::spawn(processor.run()); let handle = ActivityHeartbeatProcessorHandle { heartbeat_tx, join_handle, }; self.heartbeat_processors .insert(heartbeat.task_token, handle); } } } /// Initiates termination of all heartbeat processors by sending a signal and awaits termination pub async fn shutdown(mut self) -> Result<(), JoinError> { self.shutdown_tx .send(true) .expect("shutdown channel can't be dropped before shutdown is complete"); for v in self.heartbeat_processors.drain() { v.1.join_handle.await?; } Ok(()) } } impl<SG: ServerGatewayApis + Send + Sync + 'static> ActivityHeartbeatProcessor<SG> { async fn run(mut self) { // Each processor is initialized with heartbeat payloads, first thing we need to do is send // it out. self.record_heartbeat().await; loop { sleep(self.delay).await; select! { biased; _ = self.heartbeat_rx.changed() => { self.record_heartbeat().await; } _ = self.shutdown_rx.changed() => { break; } _ = sleep(self.delay) => { // Timed out while waiting for the next heartbeat. We waited 2 * delay in total, // where delay is 1/2 of the activity heartbeat timeout. This means that // activity has either timed out or completed by now. break; } }; } self.events_tx .send(LifecycleEvent::CleanupProcessor(self.task_token)) .expect("cleanup requests should not be dropped"); } async fn record_heartbeat(&mut self) { let details = self.heartbeat_rx.borrow().clone(); match self .server_gateway .record_activity_heartbeat(self.task_token.clone(), details.into_payloads()) .await { Ok(RecordActivityTaskHeartbeatResponse { cancel_requested }) => { if cancel_requested { self.cancels_tx .send(self.task_token.clone()) .expect("Receive half of heartbeat cancels not blocked"); } } // Send cancels for any activity that learns its workflow already finished (which is // one thing not found implies - other reasons would seem equally valid). Err(s) if s.code() == tonic::Code::NotFound => { self.cancels_tx .send(self.task_token.clone()) .expect("Receive half of heartbeat cancels not blocked"); } Err(e) => { warn!("Error when recording heartbeat: {:?}", e) } } } } #[cfg(test)] mod test { use super::*; use crate::pollers::MockServerGatewayApis; use crate::protos::coresdk::common::Payload; use crate::protos::temporal::api::workflowservice::v1::RecordActivityTaskHeartbeatResponse; use std::time::Duration; /// Ensure that hearbeats that are sent with a small delay are aggregated and sent roughly once /// every 1/2 of the heartbeat timeout. #[tokio::test] async fn process_heartbeats_and_shutdown() { let mut mock_gateway = MockServerGatewayApis::new(); mock_gateway .expect_record_activity_heartbeat() .returning(|_, _| Ok(RecordActivityTask
record
identifier_name
msg_test.go
Price: sdk.NewInt(100000), expectPass: true}, {amount: sdk.NewInt(0), gasPrice: sdk.NewInt(100000), expectPass: true}, {amount: sdk.NewInt(-1), gasPrice: sdk.NewInt(100000), expectPass: false}, {amount: sdk.NewInt(100), gasPrice: sdk.NewInt(-1), expectPass: false}, {amount: sdk.NewInt(100), gasPrice: sdk.NewInt(0), expectPass: false}, } for i, tc := range testCases { msg := NewMsgEthermint(tc.nonce, tc.to, tc.amount, tc.gasLimit, tc.gasPrice, tc.payload, tc.from) if tc.expectPass { require.Nil(t, msg.ValidateBasic(), "test: %v", i) } else { require.NotNil(t, msg.ValidateBasic(), "test: %v", i) } } } func TestMsgEthermintEncodingAndDecoding(t *testing.T) { addr := newSdkAddress() fromAddr := newSdkAddress() msg := NewMsgEthermint(0, &addr, sdk.NewInt(1), 100000, sdk.NewInt(2), []byte("test"), fromAddr) raw, err := ModuleCdc.MarshalBinaryBare(msg) require.NoError(t, err) var msg2 MsgEthermint err = ModuleCdc.UnmarshalBinaryBare(raw, &msg2) require.NoError(t, err) require.Equal(t, msg.AccountNonce, msg2.AccountNonce) require.Equal(t, msg.Recipient, msg2.Recipient) require.Equal(t, msg.Amount, msg2.Amount) require.Equal(t, msg.GasLimit, msg2.GasLimit) require.Equal(t, msg.Price, msg2.Price) require.Equal(t, msg.Payload, msg2.Payload) require.Equal(t, msg.From, msg2.From) } func newSdkAddress() sdk.AccAddress { tmpKey := secp256k1.GenPrivKey().PubKey() return sdk.AccAddress(tmpKey.Address().Bytes()) } func TestMsgEthereumTx(t *testing.T) { addr := GenerateEthAddress() msg := NewMsgEthereumTx(0, &addr, nil, 100000, nil, []byte("test")) require.NotNil(t, msg) require.Equal(t, *msg.Data.Recipient, addr) require.Equal(t, msg.Route(), RouterKey) require.Equal(t, msg.Type(), TypeMsgEthereumTx) require.NotNil(t, msg.To()) require.Equal(t, msg.GetMsgs(), []sdk.Msg{msg}) require.Panics(t, func() { msg.GetSigners() }) require.Panics(t, func() { msg.GetSignBytes() }) msg = NewMsgEthereumTxContract(0, nil, 100000, nil, []byte("test")) require.NotNil(t, msg) require.Nil(t, msg.Data.Recipient) require.Nil(t, msg.To()) } func TestMsgEthereumTxValidation(t *testing.T) { testCases := []struct { msg string amount *big.Int gasPrice *big.Int expectPass bool }{ {msg: "pass", amount: big.NewInt(100), gasPrice: big.NewInt(100000), expectPass: true}, {msg: "invalid amount", amount: big.NewInt(-1), gasPrice: big.NewInt(100000), expectPass: false}, {msg: "invalid gas price", amount: big.NewInt(100), gasPrice: big.NewInt(-1), expectPass: false}, {msg: "invalid gas price", amount: big.NewInt(100), gasPrice: big.NewInt(0), expectPass: false}, } for i, tc := range testCases { msg := NewMsgEthereumTx(0, nil, tc.amount, 0, tc.gasPrice, nil) if tc.expectPass { require.Nil(t, msg.ValidateBasic(), "valid test %d failed: %s", i, tc.msg) } else { require.NotNil(t, msg.ValidateBasic(), "invalid test %d passed: %s", i, tc.msg) } } } func TestMsgEthereumTxRLPSignBytes(t *testing.T) { addr := ethcmn.BytesToAddress([]byte("test_address")) chainID := big.NewInt(3) msg := NewMsgEthereumTx(0, &addr, nil, 100000, nil, []byte("test")) hash := msg.RLPSignBytes(chainID) require.Equal(t, "5BD30E35AD27449390B14C91E6BCFDCAADF8FE44EF33680E3BC200FC0DC083C7", fmt.Sprintf("%X", hash)) } func TestMsgEthereumTxRLPEncode(t *testing.T) { addr := ethcmn.BytesToAddress([]byte("test_address")) msg := NewMsgEthereumTx(0, &addr, nil, 100000, nil, []byte("test")) raw, err := rlp.EncodeToBytes(&msg) require.NoError(t, err) require.Equal(t, ethcmn.FromHex("E48080830186A0940000000000000000746573745F61646472657373808474657374808080"), raw) } func TestMsgEthereumTxRLPDecode(t *testing.T) { var msg MsgEthereumTx raw := ethcmn.FromHex("E48080830186A0940000000000000000746573745F61646472657373808474657374808080") addr := ethcmn.BytesToAddress([]byte("test_address")) expectedMsg := NewMsgEthereumTx(0, &addr, nil, 100000, nil, []byte("test")) err := rlp.Decode(bytes.NewReader(raw), &msg) require.NoError(t, err) require.Equal(t, expectedMsg.Data, msg.Data) // value size exceeds available input length of stream mockStream := rlp.NewStream(bytes.NewReader(raw), 1) require.Error(t, msg.DecodeRLP(mockStream)) } func TestMsgEthereumTxSig(t *testing.T)
require.Equal(t, addr1, signer) require.NotEqual(t, addr2, signer) // msg atomic load signer, err = msg.VerifySig(chainID) require.NoError(t, err) require.Equal(t, addr1, signer) signers := msg.GetSigners() require.Equal(t, 1, len(signers)) require.True(t, addrSDKAddr1.Equals(signers[0])) // zero chainID err = msg.Sign(zeroChainID, priv1.ToECDSA()) require.Nil(t, err) _, err = msg.VerifySig(zeroChainID) require.Nil(t, err) // require invalid chain ID fail validation msg = NewMsgEthereumTx(0, &addr1, nil, 100000, nil, []byte("test")) err = msg.Sign(chainID, priv1.ToECDSA()) require.Nil(t, err) signer, err = msg.VerifySig(big.NewInt(4)) require.Error(t, err) require.Equal(t, ethcmn.Address{}, signer) } func TestMsgEthereumTx_ChainID(t *testing.T) { chainID := big.NewInt(3) priv, _ := ethsecp256k1.GenerateKey() addr := ethcmn.BytesToAddress(priv.PubKey().Address().Bytes()) msg := NewMsgEthereumTx(0, &addr, nil, 100000, nil, []byte("test")) err := msg.Sign(chainID, priv.ToECDSA()) require.Nil(t, err) require.True(t,
{ chainID, zeroChainID := big.NewInt(3), big.NewInt(0) priv1, _ := ethsecp256k1.GenerateKey() priv2, _ := ethsecp256k1.GenerateKey() addr1 := ethcmn.BytesToAddress(priv1.PubKey().Address().Bytes()) trimed := strings.TrimPrefix(addr1.Hex(), "0x") fmt.Printf("%s\n", trimed) addrSDKAddr1, err := sdk.AccAddressFromHex(trimed) require.NoError(t, err) addr2 := ethcmn.BytesToAddress(priv2.PubKey().Address().Bytes()) // require valid signature passes validation msg := NewMsgEthereumTx(0, &addr1, nil, 100000, nil, []byte("test")) err = msg.Sign(chainID, priv1.ToECDSA()) require.Nil(t, err) signer, err := msg.VerifySig(chainID) require.NoError(t, err)
identifier_body
msg_test.go
Price: sdk.NewInt(100000), expectPass: true}, {amount: sdk.NewInt(0), gasPrice: sdk.NewInt(100000), expectPass: true}, {amount: sdk.NewInt(-1), gasPrice: sdk.NewInt(100000), expectPass: false}, {amount: sdk.NewInt(100), gasPrice: sdk.NewInt(-1), expectPass: false}, {amount: sdk.NewInt(100), gasPrice: sdk.NewInt(0), expectPass: false}, } for i, tc := range testCases { msg := NewMsgEthermint(tc.nonce, tc.to, tc.amount, tc.gasLimit, tc.gasPrice, tc.payload, tc.from) if tc.expectPass { require.Nil(t, msg.ValidateBasic(), "test: %v", i) } else { require.NotNil(t, msg.ValidateBasic(), "test: %v", i) } } } func TestMsgEthermintEncodingAndDecoding(t *testing.T) { addr := newSdkAddress() fromAddr := newSdkAddress() msg := NewMsgEthermint(0, &addr, sdk.NewInt(1), 100000, sdk.NewInt(2), []byte("test"), fromAddr) raw, err := ModuleCdc.MarshalBinaryBare(msg) require.NoError(t, err) var msg2 MsgEthermint err = ModuleCdc.UnmarshalBinaryBare(raw, &msg2) require.NoError(t, err) require.Equal(t, msg.AccountNonce, msg2.AccountNonce) require.Equal(t, msg.Recipient, msg2.Recipient) require.Equal(t, msg.Amount, msg2.Amount) require.Equal(t, msg.GasLimit, msg2.GasLimit) require.Equal(t, msg.Price, msg2.Price) require.Equal(t, msg.Payload, msg2.Payload) require.Equal(t, msg.From, msg2.From) } func newSdkAddress() sdk.AccAddress { tmpKey := secp256k1.GenPrivKey().PubKey() return sdk.AccAddress(tmpKey.Address().Bytes()) } func TestMsgEthereumTx(t *testing.T) { addr := GenerateEthAddress() msg := NewMsgEthereumTx(0, &addr, nil, 100000, nil, []byte("test")) require.NotNil(t, msg) require.Equal(t, *msg.Data.Recipient, addr) require.Equal(t, msg.Route(), RouterKey) require.Equal(t, msg.Type(), TypeMsgEthereumTx) require.NotNil(t, msg.To()) require.Equal(t, msg.GetMsgs(), []sdk.Msg{msg}) require.Panics(t, func() { msg.GetSigners() }) require.Panics(t, func() { msg.GetSignBytes() }) msg = NewMsgEthereumTxContract(0, nil, 100000, nil, []byte("test")) require.NotNil(t, msg) require.Nil(t, msg.Data.Recipient) require.Nil(t, msg.To()) } func TestMsgEthereumTxValidation(t *testing.T) { testCases := []struct { msg string amount *big.Int gasPrice *big.Int expectPass bool }{ {msg: "pass", amount: big.NewInt(100), gasPrice: big.NewInt(100000), expectPass: true}, {msg: "invalid amount", amount: big.NewInt(-1), gasPrice: big.NewInt(100000), expectPass: false}, {msg: "invalid gas price", amount: big.NewInt(100), gasPrice: big.NewInt(-1), expectPass: false}, {msg: "invalid gas price", amount: big.NewInt(100), gasPrice: big.NewInt(0), expectPass: false}, } for i, tc := range testCases { msg := NewMsgEthereumTx(0, nil, tc.amount, 0, tc.gasPrice, nil) if tc.expectPass { require.Nil(t, msg.ValidateBasic(), "valid test %d failed: %s", i, tc.msg) } else { require.NotNil(t, msg.ValidateBasic(), "invalid test %d passed: %s", i, tc.msg) } } } func TestMsgEthereumTxRLPSignBytes(t *testing.T) { addr := ethcmn.BytesToAddress([]byte("test_address")) chainID := big.NewInt(3) msg := NewMsgEthereumTx(0, &addr, nil, 100000, nil, []byte("test")) hash := msg.RLPSignBytes(chainID) require.Equal(t, "5BD30E35AD27449390B14C91E6BCFDCAADF8FE44EF33680E3BC200FC0DC083C7", fmt.Sprintf("%X", hash)) } func TestMsgEthereumTxRLPEncode(t *testing.T) { addr := ethcmn.BytesToAddress([]byte("test_address")) msg := NewMsgEthereumTx(0, &addr, nil, 100000, nil, []byte("test")) raw, err := rlp.EncodeToBytes(&msg) require.NoError(t, err) require.Equal(t, ethcmn.FromHex("E48080830186A0940000000000000000746573745F61646472657373808474657374808080"), raw) } func TestMsgEthereumTxRLPDecode(t *testing.T) { var msg MsgEthereumTx raw := ethcmn.FromHex("E48080830186A0940000000000000000746573745F61646472657373808474657374808080") addr := ethcmn.BytesToAddress([]byte("test_address")) expectedMsg := NewMsgEthereumTx(0, &addr, nil, 100000, nil, []byte("test")) err := rlp.Decode(bytes.NewReader(raw), &msg) require.NoError(t, err) require.Equal(t, expectedMsg.Data, msg.Data) // value size exceeds available input length of stream mockStream := rlp.NewStream(bytes.NewReader(raw), 1) require.Error(t, msg.DecodeRLP(mockStream)) } func TestMsgEthereumTxSig(t *testing.T) { chainID, zeroChainID := big.NewInt(3), big.NewInt(0) priv1, _ := ethsecp256k1.GenerateKey() priv2, _ := ethsecp256k1.GenerateKey() addr1 := ethcmn.BytesToAddress(priv1.PubKey().Address().Bytes()) trimed := strings.TrimPrefix(addr1.Hex(), "0x") fmt.Printf("%s\n", trimed) addrSDKAddr1, err := sdk.AccAddressFromHex(trimed) require.NoError(t, err) addr2 := ethcmn.BytesToAddress(priv2.PubKey().Address().Bytes()) // require valid signature passes validation msg := NewMsgEthereumTx(0, &addr1, nil, 100000, nil, []byte("test")) err = msg.Sign(chainID, priv1.ToECDSA()) require.Nil(t, err) signer, err := msg.VerifySig(chainID) require.NoError(t, err) require.Equal(t, addr1, signer) require.NotEqual(t, addr2, signer) // msg atomic load signer, err = msg.VerifySig(chainID) require.NoError(t, err) require.Equal(t, addr1, signer) signers := msg.GetSigners() require.Equal(t, 1, len(signers)) require.True(t, addrSDKAddr1.Equals(signers[0])) // zero chainID err = msg.Sign(zeroChainID, priv1.ToECDSA()) require.Nil(t, err) _, err = msg.VerifySig(zeroChainID) require.Nil(t, err) // require invalid chain ID fail validation msg = NewMsgEthereumTx(0, &addr1, nil, 100000, nil, []byte("test")) err = msg.Sign(chainID, priv1.ToECDSA()) require.Nil(t, err) signer, err = msg.VerifySig(big.NewInt(4)) require.Error(t, err) require.Equal(t, ethcmn.Address{}, signer) } func
(t *testing.T) { chainID := big.NewInt(3) priv, _ := ethsecp256k1.GenerateKey() addr := ethcmn.BytesToAddress(priv.PubKey().Address().Bytes()) msg := NewMsgEthereumTx(0, &addr, nil, 100000, nil, []byte("test")) err := msg.Sign(chainID, priv.ToECDSA()) require.Nil(t, err) require.True(t,
TestMsgEthereumTx_ChainID
identifier_name
msg_test.go
Price: sdk.NewInt(100000), expectPass: true}, {amount: sdk.NewInt(0), gasPrice: sdk.NewInt(100000), expectPass: true}, {amount: sdk.NewInt(-1), gasPrice: sdk.NewInt(100000), expectPass: false}, {amount: sdk.NewInt(100), gasPrice: sdk.NewInt(-1), expectPass: false}, {amount: sdk.NewInt(100), gasPrice: sdk.NewInt(0), expectPass: false}, } for i, tc := range testCases { msg := NewMsgEthermint(tc.nonce, tc.to, tc.amount, tc.gasLimit, tc.gasPrice, tc.payload, tc.from) if tc.expectPass { require.Nil(t, msg.ValidateBasic(), "test: %v", i) } else { require.NotNil(t, msg.ValidateBasic(), "test: %v", i) } } } func TestMsgEthermintEncodingAndDecoding(t *testing.T) { addr := newSdkAddress() fromAddr := newSdkAddress() msg := NewMsgEthermint(0, &addr, sdk.NewInt(1), 100000, sdk.NewInt(2), []byte("test"), fromAddr) raw, err := ModuleCdc.MarshalBinaryBare(msg) require.NoError(t, err) var msg2 MsgEthermint err = ModuleCdc.UnmarshalBinaryBare(raw, &msg2) require.NoError(t, err) require.Equal(t, msg.AccountNonce, msg2.AccountNonce) require.Equal(t, msg.Recipient, msg2.Recipient) require.Equal(t, msg.Amount, msg2.Amount) require.Equal(t, msg.GasLimit, msg2.GasLimit) require.Equal(t, msg.Price, msg2.Price) require.Equal(t, msg.Payload, msg2.Payload) require.Equal(t, msg.From, msg2.From) } func newSdkAddress() sdk.AccAddress { tmpKey := secp256k1.GenPrivKey().PubKey() return sdk.AccAddress(tmpKey.Address().Bytes()) } func TestMsgEthereumTx(t *testing.T) { addr := GenerateEthAddress() msg := NewMsgEthereumTx(0, &addr, nil, 100000, nil, []byte("test")) require.NotNil(t, msg) require.Equal(t, *msg.Data.Recipient, addr) require.Equal(t, msg.Route(), RouterKey) require.Equal(t, msg.Type(), TypeMsgEthereumTx) require.NotNil(t, msg.To()) require.Equal(t, msg.GetMsgs(), []sdk.Msg{msg}) require.Panics(t, func() { msg.GetSigners() }) require.Panics(t, func() { msg.GetSignBytes() }) msg = NewMsgEthereumTxContract(0, nil, 100000, nil, []byte("test")) require.NotNil(t, msg) require.Nil(t, msg.Data.Recipient) require.Nil(t, msg.To()) } func TestMsgEthereumTxValidation(t *testing.T) { testCases := []struct { msg string amount *big.Int gasPrice *big.Int expectPass bool }{ {msg: "pass", amount: big.NewInt(100), gasPrice: big.NewInt(100000), expectPass: true}, {msg: "invalid amount", amount: big.NewInt(-1), gasPrice: big.NewInt(100000), expectPass: false}, {msg: "invalid gas price", amount: big.NewInt(100), gasPrice: big.NewInt(-1), expectPass: false}, {msg: "invalid gas price", amount: big.NewInt(100), gasPrice: big.NewInt(0), expectPass: false}, } for i, tc := range testCases
} func TestMsgEthereumTxRLPSignBytes(t *testing.T) { addr := ethcmn.BytesToAddress([]byte("test_address")) chainID := big.NewInt(3) msg := NewMsgEthereumTx(0, &addr, nil, 100000, nil, []byte("test")) hash := msg.RLPSignBytes(chainID) require.Equal(t, "5BD30E35AD27449390B14C91E6BCFDCAADF8FE44EF33680E3BC200FC0DC083C7", fmt.Sprintf("%X", hash)) } func TestMsgEthereumTxRLPEncode(t *testing.T) { addr := ethcmn.BytesToAddress([]byte("test_address")) msg := NewMsgEthereumTx(0, &addr, nil, 100000, nil, []byte("test")) raw, err := rlp.EncodeToBytes(&msg) require.NoError(t, err) require.Equal(t, ethcmn.FromHex("E48080830186A0940000000000000000746573745F61646472657373808474657374808080"), raw) } func TestMsgEthereumTxRLPDecode(t *testing.T) { var msg MsgEthereumTx raw := ethcmn.FromHex("E48080830186A0940000000000000000746573745F61646472657373808474657374808080") addr := ethcmn.BytesToAddress([]byte("test_address")) expectedMsg := NewMsgEthereumTx(0, &addr, nil, 100000, nil, []byte("test")) err := rlp.Decode(bytes.NewReader(raw), &msg) require.NoError(t, err) require.Equal(t, expectedMsg.Data, msg.Data) // value size exceeds available input length of stream mockStream := rlp.NewStream(bytes.NewReader(raw), 1) require.Error(t, msg.DecodeRLP(mockStream)) } func TestMsgEthereumTxSig(t *testing.T) { chainID, zeroChainID := big.NewInt(3), big.NewInt(0) priv1, _ := ethsecp256k1.GenerateKey() priv2, _ := ethsecp256k1.GenerateKey() addr1 := ethcmn.BytesToAddress(priv1.PubKey().Address().Bytes()) trimed := strings.TrimPrefix(addr1.Hex(), "0x") fmt.Printf("%s\n", trimed) addrSDKAddr1, err := sdk.AccAddressFromHex(trimed) require.NoError(t, err) addr2 := ethcmn.BytesToAddress(priv2.PubKey().Address().Bytes()) // require valid signature passes validation msg := NewMsgEthereumTx(0, &addr1, nil, 100000, nil, []byte("test")) err = msg.Sign(chainID, priv1.ToECDSA()) require.Nil(t, err) signer, err := msg.VerifySig(chainID) require.NoError(t, err) require.Equal(t, addr1, signer) require.NotEqual(t, addr2, signer) // msg atomic load signer, err = msg.VerifySig(chainID) require.NoError(t, err) require.Equal(t, addr1, signer) signers := msg.GetSigners() require.Equal(t, 1, len(signers)) require.True(t, addrSDKAddr1.Equals(signers[0])) // zero chainID err = msg.Sign(zeroChainID, priv1.ToECDSA()) require.Nil(t, err) _, err = msg.VerifySig(zeroChainID) require.Nil(t, err) // require invalid chain ID fail validation msg = NewMsgEthereumTx(0, &addr1, nil, 100000, nil, []byte("test")) err = msg.Sign(chainID, priv1.ToECDSA()) require.Nil(t, err) signer, err = msg.VerifySig(big.NewInt(4)) require.Error(t, err) require.Equal(t, ethcmn.Address{}, signer) } func TestMsgEthereumTx_ChainID(t *testing.T) { chainID := big.NewInt(3) priv, _ := ethsecp256k1.GenerateKey() addr := ethcmn.BytesToAddress(priv.PubKey().Address().Bytes()) msg := NewMsgEthereumTx(0, &addr, nil, 100000, nil, []byte("test")) err := msg.Sign(chainID, priv.ToECDSA()) require.Nil(t, err) require.True(t
{ msg := NewMsgEthereumTx(0, nil, tc.amount, 0, tc.gasPrice, nil) if tc.expectPass { require.Nil(t, msg.ValidateBasic(), "valid test %d failed: %s", i, tc.msg) } else { require.NotNil(t, msg.ValidateBasic(), "invalid test %d passed: %s", i, tc.msg) } }
conditional_block
msg_test.go
Price: sdk.NewInt(100000), expectPass: true}, {amount: sdk.NewInt(0), gasPrice: sdk.NewInt(100000), expectPass: true}, {amount: sdk.NewInt(-1), gasPrice: sdk.NewInt(100000), expectPass: false}, {amount: sdk.NewInt(100), gasPrice: sdk.NewInt(-1), expectPass: false}, {amount: sdk.NewInt(100), gasPrice: sdk.NewInt(0), expectPass: false}, } for i, tc := range testCases { msg := NewMsgEthermint(tc.nonce, tc.to, tc.amount, tc.gasLimit, tc.gasPrice, tc.payload, tc.from) if tc.expectPass { require.Nil(t, msg.ValidateBasic(), "test: %v", i) } else { require.NotNil(t, msg.ValidateBasic(), "test: %v", i) } } } func TestMsgEthermintEncodingAndDecoding(t *testing.T) { addr := newSdkAddress() fromAddr := newSdkAddress() msg := NewMsgEthermint(0, &addr, sdk.NewInt(1), 100000, sdk.NewInt(2), []byte("test"), fromAddr) raw, err := ModuleCdc.MarshalBinaryBare(msg) require.NoError(t, err) var msg2 MsgEthermint err = ModuleCdc.UnmarshalBinaryBare(raw, &msg2) require.NoError(t, err) require.Equal(t, msg.AccountNonce, msg2.AccountNonce) require.Equal(t, msg.Recipient, msg2.Recipient) require.Equal(t, msg.Amount, msg2.Amount) require.Equal(t, msg.GasLimit, msg2.GasLimit) require.Equal(t, msg.Price, msg2.Price) require.Equal(t, msg.Payload, msg2.Payload) require.Equal(t, msg.From, msg2.From) } func newSdkAddress() sdk.AccAddress { tmpKey := secp256k1.GenPrivKey().PubKey() return sdk.AccAddress(tmpKey.Address().Bytes()) } func TestMsgEthereumTx(t *testing.T) { addr := GenerateEthAddress() msg := NewMsgEthereumTx(0, &addr, nil, 100000, nil, []byte("test")) require.NotNil(t, msg) require.Equal(t, *msg.Data.Recipient, addr) require.Equal(t, msg.Route(), RouterKey) require.Equal(t, msg.Type(), TypeMsgEthereumTx) require.NotNil(t, msg.To()) require.Equal(t, msg.GetMsgs(), []sdk.Msg{msg}) require.Panics(t, func() { msg.GetSigners() }) require.Panics(t, func() { msg.GetSignBytes() }) msg = NewMsgEthereumTxContract(0, nil, 100000, nil, []byte("test")) require.NotNil(t, msg) require.Nil(t, msg.Data.Recipient) require.Nil(t, msg.To()) } func TestMsgEthereumTxValidation(t *testing.T) { testCases := []struct { msg string amount *big.Int gasPrice *big.Int expectPass bool }{ {msg: "pass", amount: big.NewInt(100), gasPrice: big.NewInt(100000), expectPass: true}, {msg: "invalid amount", amount: big.NewInt(-1), gasPrice: big.NewInt(100000), expectPass: false}, {msg: "invalid gas price", amount: big.NewInt(100), gasPrice: big.NewInt(-1), expectPass: false}, {msg: "invalid gas price", amount: big.NewInt(100), gasPrice: big.NewInt(0), expectPass: false},
} for i, tc := range testCases { msg := NewMsgEthereumTx(0, nil, tc.amount, 0, tc.gasPrice, nil) if tc.expectPass { require.Nil(t, msg.ValidateBasic(), "valid test %d failed: %s", i, tc.msg) } else { require.NotNil(t, msg.ValidateBasic(), "invalid test %d passed: %s", i, tc.msg) } } } func TestMsgEthereumTxRLPSignBytes(t *testing.T) { addr := ethcmn.BytesToAddress([]byte("test_address")) chainID := big.NewInt(3) msg := NewMsgEthereumTx(0, &addr, nil, 100000, nil, []byte("test")) hash := msg.RLPSignBytes(chainID) require.Equal(t, "5BD30E35AD27449390B14C91E6BCFDCAADF8FE44EF33680E3BC200FC0DC083C7", fmt.Sprintf("%X", hash)) } func TestMsgEthereumTxRLPEncode(t *testing.T) { addr := ethcmn.BytesToAddress([]byte("test_address")) msg := NewMsgEthereumTx(0, &addr, nil, 100000, nil, []byte("test")) raw, err := rlp.EncodeToBytes(&msg) require.NoError(t, err) require.Equal(t, ethcmn.FromHex("E48080830186A0940000000000000000746573745F61646472657373808474657374808080"), raw) } func TestMsgEthereumTxRLPDecode(t *testing.T) { var msg MsgEthereumTx raw := ethcmn.FromHex("E48080830186A0940000000000000000746573745F61646472657373808474657374808080") addr := ethcmn.BytesToAddress([]byte("test_address")) expectedMsg := NewMsgEthereumTx(0, &addr, nil, 100000, nil, []byte("test")) err := rlp.Decode(bytes.NewReader(raw), &msg) require.NoError(t, err) require.Equal(t, expectedMsg.Data, msg.Data) // value size exceeds available input length of stream mockStream := rlp.NewStream(bytes.NewReader(raw), 1) require.Error(t, msg.DecodeRLP(mockStream)) } func TestMsgEthereumTxSig(t *testing.T) { chainID, zeroChainID := big.NewInt(3), big.NewInt(0) priv1, _ := ethsecp256k1.GenerateKey() priv2, _ := ethsecp256k1.GenerateKey() addr1 := ethcmn.BytesToAddress(priv1.PubKey().Address().Bytes()) trimed := strings.TrimPrefix(addr1.Hex(), "0x") fmt.Printf("%s\n", trimed) addrSDKAddr1, err := sdk.AccAddressFromHex(trimed) require.NoError(t, err) addr2 := ethcmn.BytesToAddress(priv2.PubKey().Address().Bytes()) // require valid signature passes validation msg := NewMsgEthereumTx(0, &addr1, nil, 100000, nil, []byte("test")) err = msg.Sign(chainID, priv1.ToECDSA()) require.Nil(t, err) signer, err := msg.VerifySig(chainID) require.NoError(t, err) require.Equal(t, addr1, signer) require.NotEqual(t, addr2, signer) // msg atomic load signer, err = msg.VerifySig(chainID) require.NoError(t, err) require.Equal(t, addr1, signer) signers := msg.GetSigners() require.Equal(t, 1, len(signers)) require.True(t, addrSDKAddr1.Equals(signers[0])) // zero chainID err = msg.Sign(zeroChainID, priv1.ToECDSA()) require.Nil(t, err) _, err = msg.VerifySig(zeroChainID) require.Nil(t, err) // require invalid chain ID fail validation msg = NewMsgEthereumTx(0, &addr1, nil, 100000, nil, []byte("test")) err = msg.Sign(chainID, priv1.ToECDSA()) require.Nil(t, err) signer, err = msg.VerifySig(big.NewInt(4)) require.Error(t, err) require.Equal(t, ethcmn.Address{}, signer) } func TestMsgEthereumTx_ChainID(t *testing.T) { chainID := big.NewInt(3) priv, _ := ethsecp256k1.GenerateKey() addr := ethcmn.BytesToAddress(priv.PubKey().Address().Bytes()) msg := NewMsgEthereumTx(0, &addr, nil, 100000, nil, []byte("test")) err := msg.Sign(chainID, priv.ToECDSA()) require.Nil(t, err) require.True(t, chain
random_line_split
process_vm.rs
Drop for ProcessVM { fn drop(&mut self) { let mut mem_chunks = self.mem_chunks.write().unwrap(); // There are two cases when this drop is called: // (1) Process exits normally and in the end, drop process VM // (2) During creating process stage, process VM is ready but there are some other errors when creating the process, e.g. spawn_attribute is set // to a wrong value // // For the first case, the process VM is cleaned in the exit procedure and nothing is needed. For the second cases, mem_chunks is not empty and should // be cleaned here. mem_chunks .drain_filter(|chunk| chunk.is_single_vma()) .for_each(|chunk| { USER_SPACE_VM_MANAGER .internal() .munmap_chunk(&chunk, None, false); }); assert!(mem_chunks.len() == 0); info!("Process VM dropped"); } } impl ProcessVM { pub fn mem_chunks(&self) -> &MemChunks { &self.mem_chunks } pub fn stack_range(&self) -> &VMRange { &self.stack_range } pub fn heap_range(&self) -> &VMRange { &self.heap_range } pub fn add_mem_chunk(&self, chunk: ChunkRef) { let mut mem_chunks = self.mem_chunks.write().unwrap(); mem_chunks.insert(chunk); } pub fn remove_mem_chunk(&self, chunk: &ChunkRef) { let mut mem_chunks = self.mem_chunks.write().unwrap(); mem_chunks.remove(chunk); } pub fn replace_mem_chunk(&self, old_chunk: &ChunkRef, new_chunk: ChunkRef) { self.remove_mem_chunk(old_chunk); self.add_mem_chunk(new_chunk) } // Try merging all connecting single VMAs of the process. // This is a very expensive operation. pub fn merge_all_single_vma_chunks( mem_chunks: &mut RwLockWriteGuard<HashSet<ChunkRef>>, ) -> Result<Vec<VMArea>> { // Get all single VMA chunks // Shared chunks shouldn't be merged since they are managed by shm manager and shared by multi processes let mut single_vma_chunks = mem_chunks .drain_filter(|chunk| chunk.is_single_vma() && !chunk.is_shared()) .collect::<Vec<ChunkRef>>(); single_vma_chunks.sort_unstable_by(|chunk_a, chunk_b| { chunk_a .range() .start() .partial_cmp(&chunk_b.range().start()) .unwrap() }); // Try merging connecting single VMA chunks for chunks in single_vma_chunks.windows(2) { let chunk_a = &chunks[0]; let chunk_b = &chunks[1]; let mut vma_a = match chunk_a.internal() { ChunkType::MultiVMA(_) => { unreachable!(); } ChunkType::SingleVMA(vma) => vma.lock().unwrap(), }; let mut vma_b = match chunk_b.internal() { ChunkType::MultiVMA(_) => { unreachable!(); } ChunkType::SingleVMA(vma) => vma.lock().unwrap(), }; if VMArea::can_merge_vmas(&vma_a, &vma_b) { let new_start = vma_a.start(); vma_b.set_start(new_start); // set vma_a to zero vma_a.set_end(new_start); } } // Collect merged vmas which will be the output of this function let mut merged_vmas = Vec::new(); // Insert unchanged chunks back to mem_chunks list and collect merged vmas for output for chunk in single_vma_chunks.into_iter().filter_map(|chunk| { if !chunk.is_single_dummy_vma() { if chunk.is_single_vma_with_conflict_size() { let new_vma = chunk.get_vma_for_single_vma_chunk().clone(); merged_vmas.push(new_vma); // Don't insert the merged chunks to mem_chunk list here. It should be updated later. None } else { Some(chunk) } } else { None } }) { mem_chunks.insert(chunk); } Ok(merged_vmas) } pub fn get_process_range(&self) -> &VMRange { USER_SPACE_VM_MANAGER.range() } pub fn get_elf_ranges(&self) -> &[VMRange] { &self.elf_ranges } pub fn get_heap_range(&self) -> &VMRange { &self.heap_range } pub fn get_stack_range(&self) -> &VMRange { &self.stack_range } pub fn get_base_addr(&self) -> usize { self.get_process_range().start() } pub fn get_stack_base(&self) -> usize { self.get_stack_range().end() } pub fn get_stack_limit(&self) -> usize { self.get_stack_range().start() } pub fn get_brk(&self) -> usize { *self.brk.read().unwrap() } pub fn brk(&self, brk: usize) -> Result<usize> { let heap_start = self.heap_range.start(); let heap_end = self.heap_range.end(); // Acquire lock first to avoid data-race. let mut brk_guard = self.brk.write().unwrap(); if brk >= heap_start && brk <= heap_end { // Get page-aligned brk address. let new_brk = align_up(brk, PAGE_SIZE); // Get page-aligned old brk address. let old_brk = align_up(*brk_guard, PAGE_SIZE); // Reset the memory when brk shrinks. if new_brk < old_brk { let shrink_brk_range = VMRange::new(new_brk, old_brk).expect("shrink brk range must be valid"); USER_SPACE_VM_MANAGER.reset_memory(shrink_brk_range)?; } // Return the user-specified brk address without page aligned. This is same as Linux. *brk_guard = brk; Ok(brk) } else { if brk < heap_start { error!("New brk address is too low"); } else if brk > heap_end { error!("New brk address is too high"); } Ok(*brk_guard) } } // Get a NON-accurate free size for current process pub fn get_free_size(&self) -> usize { let chunk_free_size = { let process_chunks = self.mem_chunks.read().unwrap(); process_chunks .iter() .fold(0, |acc, chunks| acc + chunks.free_size()) }; let free_size = chunk_free_size + USER_SPACE_VM_MANAGER.free_size(); free_size } pub fn mmap( &self, addr: usize, size: usize, perms: VMPerms, flags: MMapFlags, fd: FileDesc, offset: usize, ) -> Result<usize> { let addr_option = { if flags.contains(MMapFlags::MAP_FIXED) { VMMapAddr::Force(addr) } else { if addr == 0 { VMMapAddr::Any } else { VMMapAddr::Hint(addr) } } }; let initializer = { if flags.contains(MMapFlags::MAP_ANONYMOUS) { // There is no need to fill zeros in mmap. Cleaning is done after munmap. VMInitializer::DoNothing() } else { let file_ref = current!().file(fd)?; // Only shared, file-backed memory mappings have write-back files let need_write_back = if flags.contains(MMapFlags::MAP_SHARED) { true } else { false }; VMInitializer::FileBacked { file: FileBacked::new(file_ref, offset, need_write_back), } } }; let mmap_options = VMMapOptionsBuilder::default() .size(size) .addr(addr_option) .perms(perms) .initializer(initializer) .build()?; let mmap_addr = USER_SPACE_VM_MANAGER.mmap(&mmap_options)?; Ok(mmap_addr) } pub fn mremap( &self, old_addr: usize, old_size: usize, new_size: usize, flags: MRemapFlags, ) -> Result<usize> { let mremap_option = VMRemapOptions::new(old_addr, old_size, new_size, flags)?; USER_SPACE_VM_MANAGER.mremap(&mremap_option) } pub fn munmap(&self, addr: usize, size: usize) -> Result<()> { USER_SPACE_VM_MANAGER.munmap(addr, size) } pub fn mprotect(&self, addr: usize, size: usize, perms: VMPerms) -> Result<()> { let size = { if size == 0 { return Ok(()); } align_up(size, PAGE_SIZE) }; let protect_range = VMRange::new_with_size(addr, size)?; return USER_SPACE_VM_MANAGER.mprotect(addr, size, perms); } pub fn msync(&self, addr: usize, size: usize) -> Result<()>
{ return USER_SPACE_VM_MANAGER.msync(addr, size); }
identifier_body
process_vm.rs
elf_layout.extend(&segment_layout); elf_layout }) }) .collect(); // Make heap and stack 16-byte aligned let other_layouts = vec![ VMLayout::new(heap_size, 16)?, VMLayout::new(stack_size, 16)?, ]; let process_layout = elf_layouts.iter().chain(other_layouts.iter()).fold( VMLayout::new_empty(), |mut process_layout, sub_layout| { process_layout.add(&sub_layout); process_layout }, ); // Now that we end up with the memory layout required by the process, // let's allocate the memory for the process let mut chunks = HashSet::new(); // Init the memory for ELFs in the process let mut elf_ranges = Vec::with_capacity(2); elf_layouts .iter() .zip(self.elfs.iter()) .map(|(elf_layout, elf_file)| { let vm_option = VMMapOptionsBuilder::default() .size(elf_layout.size()) .align(elf_layout.align()) .perms(VMPerms::ALL) // set it to read | write | exec for simplicity .initializer(VMInitializer::ElfSpecific { elf_file: elf_file.file_ref().clone(), }) .build() .map_err(|e| { &self.handle_error_when_init(&chunks); e })?; let (elf_range, chunk_ref) = USER_SPACE_VM_MANAGER.alloc(&vm_option).map_err(|e| { &self.handle_error_when_init(&chunks); e })?; debug_assert!(elf_range.start() % elf_layout.align() == 0); chunks.insert(chunk_ref); Self::init_elf_memory(&elf_range, elf_file).map_err(|e| { &self.handle_error_when_init(&chunks); e })?; trace!("elf range = {:?}", elf_range); elf_ranges.push(elf_range); Ok(()) }) .collect::<Result<()>>()?; // Init the heap memory in the process let heap_layout = &other_layouts[0]; let vm_option = VMMapOptionsBuilder::default() .size(heap_layout.size()) .align(heap_layout.align()) .perms(VMPerms::READ | VMPerms::WRITE) .build() .map_err(|e| { &self.handle_error_when_init(&chunks); e })?; let (heap_range, chunk_ref) = USER_SPACE_VM_MANAGER.alloc(&vm_option).map_err(|e| { &self.handle_error_when_init(&chunks); e })?; debug_assert!(heap_range.start() % heap_layout.align() == 0); trace!("heap range = {:?}", heap_range); let brk = RwLock::new(heap_range.start()); chunks.insert(chunk_ref); // Init the stack memory in the process let stack_layout = &other_layouts[1]; let vm_option = VMMapOptionsBuilder::default() .size(stack_layout.size()) .align(heap_layout.align()) .perms(VMPerms::READ | VMPerms::WRITE) .build() .map_err(|e| { &self.handle_error_when_init(&chunks); e })?; let (stack_range, chunk_ref) = USER_SPACE_VM_MANAGER.alloc(&vm_option).map_err(|e| { &self.handle_error_when_init(&chunks); e })?; debug_assert!(stack_range.start() % stack_layout.align() == 0); chunks.insert(chunk_ref); trace!("stack range = {:?}", stack_range); let mem_chunks = Arc::new(RwLock::new(chunks)); Ok(ProcessVM { elf_ranges, heap_range, stack_range, brk, mem_chunks, }) } fn validate(&self) -> Result<()> { let validate_size = |size_opt| -> Result<()> { if let Some(size) = size_opt { if size == 0 || size % PAGE_SIZE != 0 { return_errno!(EINVAL, "invalid size"); } } Ok(()) }; validate_size(self.heap_size)?; validate_size(self.stack_size)?; validate_size(self.mmap_size)?; Ok(()) } fn handle_error_when_init(&self, chunks: &HashSet<Arc<Chunk>>) { chunks.iter().for_each(|chunk| { USER_SPACE_VM_MANAGER .internal() .munmap_chunk(chunk, None, false); }); } fn init_elf_memory(elf_range: &VMRange, elf_file: &ElfFile) -> Result<()> { // Destination buffer: ELF appeared in the process let elf_proc_buf = unsafe { elf_range.as_slice_mut() }; // Source buffer: ELF stored in the ELF file let elf_file_buf = elf_file.as_slice(); let base_load_address_offset = elf_file.base_load_address_offset() as usize; // Offsets to track zerolized range let mut empty_start_offset = 0; let mut empty_end_offset = 0; // Init all loadable segments elf_file .program_headers() .filter(|segment| segment.loadable()) .for_each(|segment| { let file_size = segment.p_filesz as usize; let file_offset = segment.p_offset as usize; let mem_addr = segment.p_vaddr as usize; let mem_size = segment.p_memsz as usize; let alignment = segment.p_align as usize; debug_assert!(file_size <= mem_size); let mem_start_offset = mem_addr - base_load_address_offset; // Initialize empty part to zero based on alignment empty_start_offset = align_down(mem_start_offset, alignment); for b in &mut elf_proc_buf[empty_start_offset..mem_start_offset] { *b = 0; } // Bytes of file_size length are loaded from the ELF file elf_file.file_ref().read_at( file_offset, &mut elf_proc_buf[mem_start_offset..mem_start_offset + file_size], ); // Set the remaining part to zero based on alignment debug_assert!(file_size <= mem_size); empty_end_offset = align_up(mem_start_offset + mem_size, alignment); for b in &mut elf_proc_buf[mem_start_offset + file_size..empty_end_offset] { *b = 0; } }); Ok(()) } } // MemChunks is the structure to track all the chunks which are used by this process. type MemChunks = Arc<RwLock<HashSet<ChunkRef>>>; /// The per-process virtual memory #[derive(Debug)] pub struct ProcessVM { elf_ranges: Vec<VMRange>, heap_range: VMRange, stack_range: VMRange, brk: RwLock<usize>, // Memory safety notes: the mem_chunks field must be the last one. // // Rust drops fields in the same order as they are declared. So by making // mem_chunks the last field, we ensure that when all other fields are // dropped, their drop methods (if provided) can still access the memory // region represented by the mem_chunks field. mem_chunks: MemChunks, } impl Default for ProcessVM { fn default() -> ProcessVM { ProcessVM { elf_ranges: Default::default(), heap_range: Default::default(), stack_range: Default::default(), brk: Default::default(), mem_chunks: Arc::new(RwLock::new(HashSet::new())), } } } impl Drop for ProcessVM { fn drop(&mut self) { let mut mem_chunks = self.mem_chunks.write().unwrap(); // There are two cases when this drop is called: // (1) Process exits normally and in the end, drop process VM // (2) During creating process stage, process VM is ready but there are some other errors when creating the process, e.g. spawn_attribute is set // to a wrong value // // For the first case, the process VM is cleaned in the exit procedure and nothing is needed. For the second cases, mem_chunks is not empty and should // be cleaned here. mem_chunks .drain_filter(|chunk| chunk.is_single_vma()) .for_each(|chunk| { USER_SPACE_VM_MANAGER .internal() .munmap_chunk(&chunk, None, false); }); assert!(mem_chunks.len() == 0); info!("Process VM dropped"); } } impl ProcessVM { pub fn mem_chunks(&self) -> &MemChunks { &self.mem_chunks } pub fn stack_range(&self) -> &VMRange { &self.stack_range } pub fn heap_range(&self) -> &VMRange { &self.heap_range } pub fn add_mem_chunk(&self, chunk: ChunkRef) { let mut mem_chunks = self.mem_chunks.write().unwrap(); mem_chunks.insert(chunk); } pub fn remove_mem_chunk(&self, chunk: &ChunkRef) { let mut mem_chunks = self.mem_chunks.write().unwrap(); mem_chunks.remove(chunk); } pub fn replace_mem_chunk(&self, old_chunk: &Chunk
let segment_size = (segment.p_vaddr + segment.p_memsz) as usize; let segment_align = segment.p_align as usize; let segment_layout = VMLayout::new(segment_size, segment_align).unwrap();
random_line_split
process_vm.rs
{ let vm_option = VMMapOptionsBuilder::default() .size(elf_layout.size()) .align(elf_layout.align()) .perms(VMPerms::ALL) // set it to read | write | exec for simplicity .initializer(VMInitializer::ElfSpecific { elf_file: elf_file.file_ref().clone(), }) .build() .map_err(|e| { &self.handle_error_when_init(&chunks); e })?; let (elf_range, chunk_ref) = USER_SPACE_VM_MANAGER.alloc(&vm_option).map_err(|e| { &self.handle_error_when_init(&chunks); e })?; debug_assert!(elf_range.start() % elf_layout.align() == 0); chunks.insert(chunk_ref); Self::init_elf_memory(&elf_range, elf_file).map_err(|e| { &self.handle_error_when_init(&chunks); e })?; trace!("elf range = {:?}", elf_range); elf_ranges.push(elf_range); Ok(()) }) .collect::<Result<()>>()?; // Init the heap memory in the process let heap_layout = &other_layouts[0]; let vm_option = VMMapOptionsBuilder::default() .size(heap_layout.size()) .align(heap_layout.align()) .perms(VMPerms::READ | VMPerms::WRITE) .build() .map_err(|e| { &self.handle_error_when_init(&chunks); e })?; let (heap_range, chunk_ref) = USER_SPACE_VM_MANAGER.alloc(&vm_option).map_err(|e| { &self.handle_error_when_init(&chunks); e })?; debug_assert!(heap_range.start() % heap_layout.align() == 0); trace!("heap range = {:?}", heap_range); let brk = RwLock::new(heap_range.start()); chunks.insert(chunk_ref); // Init the stack memory in the process let stack_layout = &other_layouts[1]; let vm_option = VMMapOptionsBuilder::default() .size(stack_layout.size()) .align(heap_layout.align()) .perms(VMPerms::READ | VMPerms::WRITE) .build() .map_err(|e| { &self.handle_error_when_init(&chunks); e })?; let (stack_range, chunk_ref) = USER_SPACE_VM_MANAGER.alloc(&vm_option).map_err(|e| { &self.handle_error_when_init(&chunks); e })?; debug_assert!(stack_range.start() % stack_layout.align() == 0); chunks.insert(chunk_ref); trace!("stack range = {:?}", stack_range); let mem_chunks = Arc::new(RwLock::new(chunks)); Ok(ProcessVM { elf_ranges, heap_range, stack_range, brk, mem_chunks, }) } fn validate(&self) -> Result<()> { let validate_size = |size_opt| -> Result<()> { if let Some(size) = size_opt { if size == 0 || size % PAGE_SIZE != 0 { return_errno!(EINVAL, "invalid size"); } } Ok(()) }; validate_size(self.heap_size)?; validate_size(self.stack_size)?; validate_size(self.mmap_size)?; Ok(()) } fn handle_error_when_init(&self, chunks: &HashSet<Arc<Chunk>>) { chunks.iter().for_each(|chunk| { USER_SPACE_VM_MANAGER .internal() .munmap_chunk(chunk, None, false); }); } fn init_elf_memory(elf_range: &VMRange, elf_file: &ElfFile) -> Result<()> { // Destination buffer: ELF appeared in the process let elf_proc_buf = unsafe { elf_range.as_slice_mut() }; // Source buffer: ELF stored in the ELF file let elf_file_buf = elf_file.as_slice(); let base_load_address_offset = elf_file.base_load_address_offset() as usize; // Offsets to track zerolized range let mut empty_start_offset = 0; let mut empty_end_offset = 0; // Init all loadable segments elf_file .program_headers() .filter(|segment| segment.loadable()) .for_each(|segment| { let file_size = segment.p_filesz as usize; let file_offset = segment.p_offset as usize; let mem_addr = segment.p_vaddr as usize; let mem_size = segment.p_memsz as usize; let alignment = segment.p_align as usize; debug_assert!(file_size <= mem_size); let mem_start_offset = mem_addr - base_load_address_offset; // Initialize empty part to zero based on alignment empty_start_offset = align_down(mem_start_offset, alignment); for b in &mut elf_proc_buf[empty_start_offset..mem_start_offset] { *b = 0; } // Bytes of file_size length are loaded from the ELF file elf_file.file_ref().read_at( file_offset, &mut elf_proc_buf[mem_start_offset..mem_start_offset + file_size], ); // Set the remaining part to zero based on alignment debug_assert!(file_size <= mem_size); empty_end_offset = align_up(mem_start_offset + mem_size, alignment); for b in &mut elf_proc_buf[mem_start_offset + file_size..empty_end_offset] { *b = 0; } }); Ok(()) } } // MemChunks is the structure to track all the chunks which are used by this process. type MemChunks = Arc<RwLock<HashSet<ChunkRef>>>; /// The per-process virtual memory #[derive(Debug)] pub struct ProcessVM { elf_ranges: Vec<VMRange>, heap_range: VMRange, stack_range: VMRange, brk: RwLock<usize>, // Memory safety notes: the mem_chunks field must be the last one. // // Rust drops fields in the same order as they are declared. So by making // mem_chunks the last field, we ensure that when all other fields are // dropped, their drop methods (if provided) can still access the memory // region represented by the mem_chunks field. mem_chunks: MemChunks, } impl Default for ProcessVM { fn default() -> ProcessVM { ProcessVM { elf_ranges: Default::default(), heap_range: Default::default(), stack_range: Default::default(), brk: Default::default(), mem_chunks: Arc::new(RwLock::new(HashSet::new())), } } } impl Drop for ProcessVM { fn
(&mut self) { let mut mem_chunks = self.mem_chunks.write().unwrap(); // There are two cases when this drop is called: // (1) Process exits normally and in the end, drop process VM // (2) During creating process stage, process VM is ready but there are some other errors when creating the process, e.g. spawn_attribute is set // to a wrong value // // For the first case, the process VM is cleaned in the exit procedure and nothing is needed. For the second cases, mem_chunks is not empty and should // be cleaned here. mem_chunks .drain_filter(|chunk| chunk.is_single_vma()) .for_each(|chunk| { USER_SPACE_VM_MANAGER .internal() .munmap_chunk(&chunk, None, false); }); assert!(mem_chunks.len() == 0); info!("Process VM dropped"); } } impl ProcessVM { pub fn mem_chunks(&self) -> &MemChunks { &self.mem_chunks } pub fn stack_range(&self) -> &VMRange { &self.stack_range } pub fn heap_range(&self) -> &VMRange { &self.heap_range } pub fn add_mem_chunk(&self, chunk: ChunkRef) { let mut mem_chunks = self.mem_chunks.write().unwrap(); mem_chunks.insert(chunk); } pub fn remove_mem_chunk(&self, chunk: &ChunkRef) { let mut mem_chunks = self.mem_chunks.write().unwrap(); mem_chunks.remove(chunk); } pub fn replace_mem_chunk(&self, old_chunk: &ChunkRef, new_chunk: ChunkRef) { self.remove_mem_chunk(old_chunk); self.add_mem_chunk(new_chunk) } // Try merging all connecting single VMAs of the process. // This is a very expensive operation. pub fn merge_all_single_vma_chunks( mem_chunks: &mut RwLockWriteGuard<HashSet<ChunkRef>>, ) -> Result<Vec<VMArea>> { // Get all single VMA chunks // Shared chunks shouldn't be merged since they are managed by shm manager and shared by multi processes let mut single_vma_chunks = mem_chunks .drain_filter(|chunk| chunk.is_single_vma() && !chunk.is_shared()) .collect::<Vec<ChunkRef>>(); single_vma_chunks.sort_unstable_by(|chunk_a, chunk_b| { chunk_a .range() .start() .partial_cmp(&chunk_b.range().start()) .unwrap() }); // Try merging connecting single VMA chunks for chunks in single_vma_chunks.windows(2) { let chunk_a = &chunks[0]; let chunk_b = &chunks[1]; let mut vma_a = match chunk_a.internal()
drop
identifier_name
KR_256_5x5_32_Team.py
2) def
(): time_str = dt.now().strftime('%Y-%m-%d-%H-%M-%S') experiment_id = 'base_{}'.format(time_str) return experiment_id def scheduler(epoch): if epoch == 150: K.set_value(model.optimizer.lr, 0.00014339) return K.get_value(model.optimizer.lr) def strong_aug(p=1): return Compose([ RandomRotate90(), Flip(), Transpose(), OneOf([ IAAAdditiveGaussianNoise(), GaussNoise(), ], p=0.2), OneOf([ MotionBlur(p=0.2), MedianBlur(blur_limit=3, p=0.1), Blur(blur_limit=3, p=0.1), ], p=0.2), OneOf([ CLAHE(clip_limit=2), IAASharpen(), IAAEmboss(), RandomBrightnessContrast(), ], p=0.3), HueSaturationValue(p=0.3), ], p=p) AUGMENTATIONS = strong_aug(p=0.9) seed_everything() model = Sequential() """128 Layers""" model.add(Conv2D(32, (5, 5), padding="same", strides=(2, 2), activation='linear', input_shape=(256, 256, 3), data_format="channels_last", name='conv2d_1',kernel_regularizer=regularizers.l2(0.0005))) model.add(BatchNormalization(name='batch_normalization_1')) model.add(LeakyReLU(alpha=0.01,name='leaky_re_lu_1')) model.add(Conv2D(32, (3, 3),padding="same", activation='linear', name='conv2d_2',kernel_regularizer=regularizers.l2(0.0005))) model.add(BatchNormalization(name='batch_normalization_2')) model.add(LeakyReLU(alpha=0.01,name='leaky_re_lu_2')) model.add(MaxPooling2D(pool_size=3, strides=(2, 2),name='max_pooling2d_1')) model.add(Conv2D(64, (5, 5),padding="same", strides=(2, 2), activation='linear', name='conv2d_3',kernel_regularizer=regularizers.l2(0.0005))) model.add(BatchNormalization(name='batch_normalization_3')) model.add(LeakyReLU(alpha=0.01,name='leaky_re_lu_3')) model.add(Conv2D(64, (3, 3),padding="same", activation='linear', name='conv2d_4',kernel_regularizer=regularizers.l2(0.0005))) model.add(BatchNormalization(name='batch_normalization_4')) model.add(LeakyReLU(alpha=0.01,name='leaky_re_lu_4')) model.add(Conv2D(64, (3, 3),padding="same", activation='linear', name='conv2d_5',kernel_regularizer=regularizers.l2(0.0005))) model.add(BatchNormalization(name='batch_normalization_5')) model.add(LeakyReLU(alpha=0.01,name='leaky_re_lu_5')) model.add(MaxPooling2D(pool_size=3, strides=(2, 2),name='max_pooling2d_2')) #model.add(Dropout(0.25)) model.add(Conv2D(128, (3, 3),padding="same", activation='linear', name='conv2d_6',kernel_regularizer=regularizers.l2(0.0005))) model.add(BatchNormalization(name='batch_normalization_6')) model.add(LeakyReLU(alpha=0.01,name='leaky_re_lu_6')) model.add(Conv2D(128, (3, 3),padding="same", activation='linear', name='conv2d_7',kernel_regularizer=regularizers.l2(0.0005))) model.add(BatchNormalization(name='batch_normalization_7')) model.add(LeakyReLU(alpha=0.01,name='leaky_re_lu_7')) model.add(Conv2D(128, (3, 3),padding="same", activation='linear', name='conv2d_8',kernel_regularizer=regularizers.l2(0.0005))) model.add(BatchNormalization(name='batch_normalization_8')) model.add(LeakyReLU(alpha=0.01,name='leaky_re_lu_8')) """256 Layers""" model.add(MaxPooling2D(pool_size=3, strides=(2, 2),name='max_pooling2d_3')) #model.add(Dropout(0.25)) model.add(Conv2D(256, (3, 3),padding="same", activation='linear', kernel_initializer=Orthogonal(gain=1.0), name='conv2d_9',bias_initializer=Constant(value=0.05),kernel_regularizer=regularizers.l2(0.0005))) model.add(BatchNormalization(name='batch_normalization_11')) model.add(LeakyReLU(alpha=0.01,name='leaky_re_lu_12')) model.add(Conv2D(256, (3, 3),padding="same", activation='linear', kernel_initializer=Orthogonal(gain=1.0), name='conv2d_10',bias_initializer=Constant(value=0.05),kernel_regularizer=regularizers.l2(0.0005))) model.add(BatchNormalization(name='batch_normalization_12')) model.add(LeakyReLU(alpha=0.01,name='leaky_re_lu_13')) model.add(Conv2D(256, (3, 3),padding="same", activation='linear', kernel_initializer=Orthogonal(gain=1.0), name='conv2d_11',bias_initializer=Constant(value=0.05),kernel_regularizer=regularizers.l2(0.0005))) model.add(BatchNormalization(name='batch_normalization_13')) model.add(LeakyReLU(alpha=0.01,name='leaky_re_lu_14')) """512 Layers""" #model.add(MaxPooling2D(pool_size=3, strides=(2, 2),name='max_pooling2d_4')) # #model.add(Conv2D(512, (3, 3),padding="same", activation='linear', kernel_initializer=Orthogonal(gain=1.0), # name='conv2d_12',bias_initializer=Constant(value=0.05),kernel_regularizer=regularizers.l2(0.0005))) #model.add(BatchNormalization(name='batch_normalization_14')) #model.add(LeakyReLU(alpha=0.01,name='leaky_re_lu_15')) # #model.add(Conv2D(512, (3, 3),padding="same", activation='linear', kernel_initializer=Orthogonal(gain=1.0), # name='conv2d_13',bias_initializer=Constant(value=0.05),kernel_regularizer=regularizers.l2(0.0005))) #model.add(BatchNormalization(name='batch_normalization_15')) #model.add(LeakyReLU(alpha=0.01,name='leaky_re_lu_16')) model.add(RMSPooling2D(pool_size=3,strides=(3, 3),name='rms_pooling2d_1')) model.add(Dropout(0.5)) model.add(Flatten(name='flatten_1')) model.add(Dense(1024, activation='linear', kernel_initializer=Orthogonal(gain=1.0), name='dense_1',kernel_regularizer=regularizers.l2(0.0005))) model.add(BatchNormalization(name='batch_normalization_9')) model.add(LeakyReLU(alpha=0.01,name='leaky_re_lu_9')) model.add(Lambda(Maxout1, name='lambda_1')) #instead of FeaturePoolLayer model.add(Dropout(0.5)) model.add(Dense(1024, activation='linear', kernel_initializer=Orthogonal(gain=1.0), name='dense_2',kernel_regularizer=regularizers.l2(0.0005))) model.add(BatchNormalization(name='batch_normalization_10')) model.add(LeakyReLU(alpha=0.01,name='leaky_re_lu_10')) model.add(Lambda(Maxout2, name='lambda_2'))#instead of FeaturePoolLayer model.add(Dense(1, activation='relu', kernel_initializer=Orthogonal(gain=1.0), name='dense_3',kernel_regularizer=regularizers.l2(0.0005))) pre_256_model = load_model('./Experiments/PH_1/EX_40/EX_40.hdf5', custom_objects={'RMSPooling2D': RMSPooling2D,'Lambda':Lambda,'tf':tf}) # copy weights
get_experiment_id
identifier_name
KR_256_5x5_32_Team.py
def Maxout1(x): return tf.contrib.layers.maxout(x, 512) def Maxout2(x): return tf.contrib.layers.maxout(x, 512) def get_experiment_id(): time_str = dt.now().strftime('%Y-%m-%d-%H-%M-%S') experiment_id = 'base_{}'.format(time_str) return experiment_id def scheduler(epoch): if epoch == 150: K.set_value(model.optimizer.lr, 0.00014339) return K.get_value(model.optimizer.lr) def strong_aug(p=1): return Compose([ RandomRotate90(), Flip(), Transpose(), OneOf([ IAAAdditiveGaussianNoise(), GaussNoise(), ], p=0.2), OneOf([ MotionBlur(p=0.2), MedianBlur(blur_limit=3, p=0.1), Blur(blur_limit=3, p=0.1), ], p=0.2), OneOf([ CLAHE(clip_limit=2), IAASharpen(), IAAEmboss(), RandomBrightnessContrast(), ], p=0.3), HueSaturationValue(p=0.3), ], p=p) AUGMENTATIONS = strong_aug(p=0.9) seed_everything() model = Sequential() """128 Layers""" model.add(Conv2D(32, (5, 5), padding="same", strides=(2, 2), activation='linear', input_shape=(256, 256, 3), data_format="channels_last", name='conv2d_1',kernel_regularizer=regularizers.l2(0.0005))) model.add(BatchNormalization(name='batch_normalization_1')) model.add(LeakyReLU(alpha=0.01,name='leaky_re_lu_1')) model.add(Conv2D(32, (3, 3),padding="same", activation='linear', name='conv2d_2',kernel_regularizer=regularizers.l2(0.0005))) model.add(BatchNormalization(name='batch_normalization_2')) model.add(LeakyReLU(alpha=0.01,name='leaky_re_lu_2')) model.add(MaxPooling2D(pool_size=3, strides=(2, 2),name='max_pooling2d_1')) model.add(Conv2D(64, (5, 5),padding="same", strides=(2, 2), activation='linear', name='conv2d_3',kernel_regularizer=regularizers.l2(0.0005))) model.add(BatchNormalization(name='batch_normalization_3')) model.add(LeakyReLU(alpha=0.01,name='leaky_re_lu_3')) model.add(Conv2D(64, (3, 3),padding="same", activation='linear', name='conv2d_4',kernel_regularizer=regularizers.l2(0.0005))) model.add(BatchNormalization(name='batch_normalization_4')) model.add(LeakyReLU(alpha=0.01,name='leaky_re_lu_4')) model.add(Conv2D(64, (3, 3),padding="same", activation='linear', name='conv2d_5',kernel_regularizer=regularizers.l2(0.0005))) model.add(BatchNormalization(name='batch_normalization_5')) model.add(LeakyReLU(alpha=0.01,name='leaky_re_lu_5')) model.add(MaxPooling2D(pool_size=3, strides=(2, 2),name='max_pooling2d_2')) #model.add(Dropout(0.25)) model.add(Conv2D(128, (3, 3),padding="same", activation='linear', name='conv2d_6',kernel_regularizer=regularizers.l2(0.0005))) model.add(BatchNormalization(name='batch_normalization_6')) model.add(LeakyReLU(alpha=0.01,name='leaky_re_lu_6')) model.add(Conv2D(128, (3, 3),padding="same", activation='linear', name='conv2d_7',kernel_regularizer=regularizers.l2(0.0005))) model.add(BatchNormalization(name='batch_normalization_7')) model.add(LeakyReLU(alpha=0.01,name='leaky_re_lu_7')) model.add(Conv2D(128, (3, 3),padding="same", activation='linear', name='conv2d_8',kernel_regularizer=regularizers.l2(0.0005))) model.add(BatchNormalization(name='batch_normalization_8')) model.add(LeakyReLU(alpha=0.01,name='leaky_re_lu_8')) """256 Layers""" model.add(MaxPooling2D(pool_size=3, strides=(2, 2),name='max_pooling2d_3')) #model.add(Dropout(0.25)) model.add(Conv2D(256, (3, 3),padding="same", activation='linear', kernel_initializer=Orthogonal(gain=1.0), name='conv2d_9',bias_initializer=Constant(value=0.05),kernel_regularizer=regularizers.l2(0.0005))) model.add(BatchNormalization(name='batch_normalization_11')) model.add(LeakyReLU(alpha=0.01,name='leaky_re_lu_12')) model.add(Conv2D(256, (3, 3),padding="same", activation='linear', kernel_initializer=Orthogonal(gain=1.0), name='conv2d_10',bias_initializer=Constant(value=0.05),kernel_regularizer=regularizers.l2(0.0005))) model.add(BatchNormalization(name='batch_normalization_12')) model.add(LeakyReLU(alpha=0.01,name='leaky_re_lu_13')) model.add(Conv2D(256, (3, 3),padding="same", activation='linear', kernel_initializer=Orthogonal(gain=1.0), name='conv2d_11',bias_initializer=Constant(value=0.05),kernel_regularizer=regularizers.l2(0.0005))) model.add(BatchNormalization(name='batch_normalization_13')) model.add(LeakyReLU(alpha=0.01,name='leaky_re_lu_14')) """512 Layers""" #model.add(MaxPooling2D(pool_size=3, strides=(2, 2),name='max_pooling2d_4')) # #model.add(Conv2D(512, (3, 3),padding="same", activation='linear', kernel_initializer=Orthogonal(gain=1.0), # name='conv2d_12',bias_initializer=Constant(value=0.05),kernel_regularizer=regularizers.l2(0.0005))) #model.add(BatchNormalization(name='batch_normalization_14')) #model.add(LeakyReLU(alpha=0.01,name='leaky_re_lu_15')) # #model.add(Conv2D(512, (3, 3),padding="same", activation='linear', kernel_initializer=Orthogonal(gain=1.0), # name='conv2d_13',bias_initializer=Constant(value=0.05),kernel_regularizer=regularizers.l2(0.0005))) #model.add(BatchNormalization(name='batch_normalization_15')) #model.add(LeakyReLU(alpha=0.01,name='leaky_re_lu_16')) model.add(RMSPooling2D(pool_size=3,strides=(3, 3),name='rms_pooling2d_1')) model.add(Dropout(0.5)) model.add(Flatten(name='flatten_1')) model.add(Dense(1024, activation='linear', kernel_initializer=Orthogonal(gain=1.0), name='dense_1',kernel_regularizer=regularizers.l2(0.0005))) model.add(BatchNormalization(name='batch_normalization_9')) model.add(LeakyReLU(alpha=0.01,name='leaky_re_lu_9')) model.add(Lambda(Maxout1, name='lambda_1')) #instead of FeaturePoolLayer model.add(Dropout(0.5)) model.add(Dense(1024, activation='linear', kernel_initializer=Orthogonal(gain=1.0), name='dense_2',kernel_regularizer=regularizers.l2(0.0005))) model.add(BatchNormalization(name='batch_normalization_10')) model.add(LeakyReLU(alpha=0.01,name='leaky_re_lu_10')) model.add(Lambda(Maxout2, name='lambda_2'))#instead of FeaturePoolLayer model.add(Dense(1, activation='relu', kernel_initializer=Orthogonal(gain=1.0), name='dense_3
print('Random seeds initialized') random.seed(seed) os.environ['PYTHONHASHSEED'] = str(seed) np.random.seed(seed) tf.set_random_seed(seed)
identifier_body
KR_256_5x5_32_Team.py
same", activation='linear', kernel_initializer=Orthogonal(gain=1.0), name='conv2d_9',bias_initializer=Constant(value=0.05),kernel_regularizer=regularizers.l2(0.0005))) model.add(BatchNormalization(name='batch_normalization_11')) model.add(LeakyReLU(alpha=0.01,name='leaky_re_lu_12')) model.add(Conv2D(256, (3, 3),padding="same", activation='linear', kernel_initializer=Orthogonal(gain=1.0), name='conv2d_10',bias_initializer=Constant(value=0.05),kernel_regularizer=regularizers.l2(0.0005))) model.add(BatchNormalization(name='batch_normalization_12')) model.add(LeakyReLU(alpha=0.01,name='leaky_re_lu_13')) model.add(Conv2D(256, (3, 3),padding="same", activation='linear', kernel_initializer=Orthogonal(gain=1.0), name='conv2d_11',bias_initializer=Constant(value=0.05),kernel_regularizer=regularizers.l2(0.0005))) model.add(BatchNormalization(name='batch_normalization_13')) model.add(LeakyReLU(alpha=0.01,name='leaky_re_lu_14')) """512 Layers""" #model.add(MaxPooling2D(pool_size=3, strides=(2, 2),name='max_pooling2d_4')) # #model.add(Conv2D(512, (3, 3),padding="same", activation='linear', kernel_initializer=Orthogonal(gain=1.0), # name='conv2d_12',bias_initializer=Constant(value=0.05),kernel_regularizer=regularizers.l2(0.0005))) #model.add(BatchNormalization(name='batch_normalization_14')) #model.add(LeakyReLU(alpha=0.01,name='leaky_re_lu_15')) # #model.add(Conv2D(512, (3, 3),padding="same", activation='linear', kernel_initializer=Orthogonal(gain=1.0), # name='conv2d_13',bias_initializer=Constant(value=0.05),kernel_regularizer=regularizers.l2(0.0005))) #model.add(BatchNormalization(name='batch_normalization_15')) #model.add(LeakyReLU(alpha=0.01,name='leaky_re_lu_16')) model.add(RMSPooling2D(pool_size=3,strides=(3, 3),name='rms_pooling2d_1')) model.add(Dropout(0.5)) model.add(Flatten(name='flatten_1')) model.add(Dense(1024, activation='linear', kernel_initializer=Orthogonal(gain=1.0), name='dense_1',kernel_regularizer=regularizers.l2(0.0005))) model.add(BatchNormalization(name='batch_normalization_9')) model.add(LeakyReLU(alpha=0.01,name='leaky_re_lu_9')) model.add(Lambda(Maxout1, name='lambda_1')) #instead of FeaturePoolLayer model.add(Dropout(0.5)) model.add(Dense(1024, activation='linear', kernel_initializer=Orthogonal(gain=1.0), name='dense_2',kernel_regularizer=regularizers.l2(0.0005))) model.add(BatchNormalization(name='batch_normalization_10')) model.add(LeakyReLU(alpha=0.01,name='leaky_re_lu_10')) model.add(Lambda(Maxout2, name='lambda_2'))#instead of FeaturePoolLayer model.add(Dense(1, activation='relu', kernel_initializer=Orthogonal(gain=1.0), name='dense_3',kernel_regularizer=regularizers.l2(0.0005))) pre_256_model = load_model('./Experiments/PH_1/EX_40/EX_40.hdf5', custom_objects={'RMSPooling2D': RMSPooling2D,'Lambda':Lambda,'tf':tf}) # copy weights from old model to new one for layer in model.layers: try: layer.set_weights( pre_256_model.get_layer(name=layer.name).get_weights()) print("Succesfully transfered weights for layer {}".format(layer.name)) except: print("Could not transfer weights for layer {}".format(layer.name)) #model.add_loss( sample_loss(y_true, y_pred, Homotopy ) ) adam = Adam(lr=0.0014339, beta_1=0.9, beta_2=0.999, epsilon=0.1, decay=1e-6, amsgrad=True) model.compile(loss='mse', optimizer=adam, metrics=['mae', 'acc']) model.summary() #weight checking #a = pre_128_model.layers[0].get_weights()[0] #b = model.layers[0].get_weights()[0] #if np.array_equal(a, b): # print('equal') #else: # print('not equal') experiment_id = get_experiment_id() train_df=pd.read_csv(("/data1/visionlab/Thesis/labels/EyePACS_2015_new_train.csv"), dtype={'image': str, 'level': float}) val_df=pd.read_csv(("/data1/visionlab/Thesis/labels/EyePACS_2015_new_val.csv"), dtype={'image': str, 'level': float}) train_datagen = ImageDataAugmentor(augment=AUGMENTATIONS, preprocess_input=None, rescale=1./255.) val_datagen = ImageDataGenerator(rescale=1./255.) def append_ext(fn): return fn+".tiff" #def append_ext1(fn): # return fn+".png" train_df["image"]=train_df["image"].apply(append_ext) val_df["image"]=val_df["image"].apply(append_ext) train_generator = train_datagen.flow_from_dataframe( dataframe=train_df, directory='/data1/visionlab/data/EyePACS_2015/256-EyePACS-all', x_col="image", y_col="level", has_ext=False, batch_size=32, seed=42, shuffle=True, class_mode="other", target_size=(256,256)) valid_generator = val_datagen.flow_from_dataframe( dataframe=val_df, directory='/data1/visionlab/data/EyePACS_2015/256-EyePACS-all', x_col="image", y_col="level", has_ext=False, batch_size=32, seed=42, shuffle=True, class_mode="other", target_size=(256,256)) STEP_SIZE_TRAIN=train_generator.n//train_generator.batch_size STEP_SIZE_VALID=valid_generator.n//valid_generator.batch_size # #balance_weights = K.variable(class_weight.compute_class_weight( # 'balanced', # np.unique(train_df.level), # train_df.level)) callbacks = [ # Resample_Iterator(balance_weights), Val_QWK(valid_generator, STEP_SIZE_VALID), ModelCheckpoint(experiment_id + "-val_kappa_checkpoint.hdf5", monitor='val_kappa', verbose=1, save_best_only=True, mode='max'), # EarlyStopping(monitor='val_kappa', patience=200, min_delta=0.001, verbose=1, restore_best_weights=True, mode='max',baseline=None), Val_CM(valid_generator, STEP_SIZE_VALID), LearningRateScheduler(scheduler, verbose=1), # ReduceLROnPlateau(monitor='val_loss', factor=0.1, patience=10, verbose=1, mode='min', min_delta=0.01, cooldown=0, min_lr=0), TensorBoard(log_dir='./Graph', histogram_freq=0, write_graph=True, write_images=True), WandbCallback(), ] history=model.fit_generator(generator=train_generator, steps_per_epoch=STEP_SIZE_TRAIN, validation_data=valid_generator, validation_steps=STEP_SIZE_VALID, # class_weight=[balance_weights], callbacks=callbacks, workers=8, use_multiprocessing=False, epochs=200 ) #model.save(os.path.join(wandb.run.dir, "model.h5")) # list all data in history print(history.history.keys()) # summarize history for accuracy plt.plot(history.history['acc']) plt.plot(history.history['val_acc']) plt.title('model accuracy') plt.ylabel('accuracy') plt.xlabel('epoch') plt.legend(['train', 'validation'], loc='upper left') plt.show() #plt.savefig('model accuracy.png') # summarize history for loss plt.plot(history.history['loss']) plt.plot(history.history['val_loss']) plt.title('model loss') plt.ylabel('loss') plt.xlabel('epoch') plt.legend(['train', 'validation'], loc='upper left') plt.show() #plt.savefig('model loss.png') # summarize history for val_kappa plt.plot(history.history['val_kappa']) plt.title('Validation_Kappa') plt.ylabel('val_kappa')
plt.xlabel('epoch')
random_line_split
KR_256_5x5_32_Team.py
0014339) return K.get_value(model.optimizer.lr) def strong_aug(p=1): return Compose([ RandomRotate90(), Flip(), Transpose(), OneOf([ IAAAdditiveGaussianNoise(), GaussNoise(), ], p=0.2), OneOf([ MotionBlur(p=0.2), MedianBlur(blur_limit=3, p=0.1), Blur(blur_limit=3, p=0.1), ], p=0.2), OneOf([ CLAHE(clip_limit=2), IAASharpen(), IAAEmboss(), RandomBrightnessContrast(), ], p=0.3), HueSaturationValue(p=0.3), ], p=p) AUGMENTATIONS = strong_aug(p=0.9) seed_everything() model = Sequential() """128 Layers""" model.add(Conv2D(32, (5, 5), padding="same", strides=(2, 2), activation='linear', input_shape=(256, 256, 3), data_format="channels_last", name='conv2d_1',kernel_regularizer=regularizers.l2(0.0005))) model.add(BatchNormalization(name='batch_normalization_1')) model.add(LeakyReLU(alpha=0.01,name='leaky_re_lu_1')) model.add(Conv2D(32, (3, 3),padding="same", activation='linear', name='conv2d_2',kernel_regularizer=regularizers.l2(0.0005))) model.add(BatchNormalization(name='batch_normalization_2')) model.add(LeakyReLU(alpha=0.01,name='leaky_re_lu_2')) model.add(MaxPooling2D(pool_size=3, strides=(2, 2),name='max_pooling2d_1')) model.add(Conv2D(64, (5, 5),padding="same", strides=(2, 2), activation='linear', name='conv2d_3',kernel_regularizer=regularizers.l2(0.0005))) model.add(BatchNormalization(name='batch_normalization_3')) model.add(LeakyReLU(alpha=0.01,name='leaky_re_lu_3')) model.add(Conv2D(64, (3, 3),padding="same", activation='linear', name='conv2d_4',kernel_regularizer=regularizers.l2(0.0005))) model.add(BatchNormalization(name='batch_normalization_4')) model.add(LeakyReLU(alpha=0.01,name='leaky_re_lu_4')) model.add(Conv2D(64, (3, 3),padding="same", activation='linear', name='conv2d_5',kernel_regularizer=regularizers.l2(0.0005))) model.add(BatchNormalization(name='batch_normalization_5')) model.add(LeakyReLU(alpha=0.01,name='leaky_re_lu_5')) model.add(MaxPooling2D(pool_size=3, strides=(2, 2),name='max_pooling2d_2')) #model.add(Dropout(0.25)) model.add(Conv2D(128, (3, 3),padding="same", activation='linear', name='conv2d_6',kernel_regularizer=regularizers.l2(0.0005))) model.add(BatchNormalization(name='batch_normalization_6')) model.add(LeakyReLU(alpha=0.01,name='leaky_re_lu_6')) model.add(Conv2D(128, (3, 3),padding="same", activation='linear', name='conv2d_7',kernel_regularizer=regularizers.l2(0.0005))) model.add(BatchNormalization(name='batch_normalization_7')) model.add(LeakyReLU(alpha=0.01,name='leaky_re_lu_7')) model.add(Conv2D(128, (3, 3),padding="same", activation='linear', name='conv2d_8',kernel_regularizer=regularizers.l2(0.0005))) model.add(BatchNormalization(name='batch_normalization_8')) model.add(LeakyReLU(alpha=0.01,name='leaky_re_lu_8')) """256 Layers""" model.add(MaxPooling2D(pool_size=3, strides=(2, 2),name='max_pooling2d_3')) #model.add(Dropout(0.25)) model.add(Conv2D(256, (3, 3),padding="same", activation='linear', kernel_initializer=Orthogonal(gain=1.0), name='conv2d_9',bias_initializer=Constant(value=0.05),kernel_regularizer=regularizers.l2(0.0005))) model.add(BatchNormalization(name='batch_normalization_11')) model.add(LeakyReLU(alpha=0.01,name='leaky_re_lu_12')) model.add(Conv2D(256, (3, 3),padding="same", activation='linear', kernel_initializer=Orthogonal(gain=1.0), name='conv2d_10',bias_initializer=Constant(value=0.05),kernel_regularizer=regularizers.l2(0.0005))) model.add(BatchNormalization(name='batch_normalization_12')) model.add(LeakyReLU(alpha=0.01,name='leaky_re_lu_13')) model.add(Conv2D(256, (3, 3),padding="same", activation='linear', kernel_initializer=Orthogonal(gain=1.0), name='conv2d_11',bias_initializer=Constant(value=0.05),kernel_regularizer=regularizers.l2(0.0005))) model.add(BatchNormalization(name='batch_normalization_13')) model.add(LeakyReLU(alpha=0.01,name='leaky_re_lu_14')) """512 Layers""" #model.add(MaxPooling2D(pool_size=3, strides=(2, 2),name='max_pooling2d_4')) # #model.add(Conv2D(512, (3, 3),padding="same", activation='linear', kernel_initializer=Orthogonal(gain=1.0), # name='conv2d_12',bias_initializer=Constant(value=0.05),kernel_regularizer=regularizers.l2(0.0005))) #model.add(BatchNormalization(name='batch_normalization_14')) #model.add(LeakyReLU(alpha=0.01,name='leaky_re_lu_15')) # #model.add(Conv2D(512, (3, 3),padding="same", activation='linear', kernel_initializer=Orthogonal(gain=1.0), # name='conv2d_13',bias_initializer=Constant(value=0.05),kernel_regularizer=regularizers.l2(0.0005))) #model.add(BatchNormalization(name='batch_normalization_15')) #model.add(LeakyReLU(alpha=0.01,name='leaky_re_lu_16')) model.add(RMSPooling2D(pool_size=3,strides=(3, 3),name='rms_pooling2d_1')) model.add(Dropout(0.5)) model.add(Flatten(name='flatten_1')) model.add(Dense(1024, activation='linear', kernel_initializer=Orthogonal(gain=1.0), name='dense_1',kernel_regularizer=regularizers.l2(0.0005))) model.add(BatchNormalization(name='batch_normalization_9')) model.add(LeakyReLU(alpha=0.01,name='leaky_re_lu_9')) model.add(Lambda(Maxout1, name='lambda_1')) #instead of FeaturePoolLayer model.add(Dropout(0.5)) model.add(Dense(1024, activation='linear', kernel_initializer=Orthogonal(gain=1.0), name='dense_2',kernel_regularizer=regularizers.l2(0.0005))) model.add(BatchNormalization(name='batch_normalization_10')) model.add(LeakyReLU(alpha=0.01,name='leaky_re_lu_10')) model.add(Lambda(Maxout2, name='lambda_2'))#instead of FeaturePoolLayer model.add(Dense(1, activation='relu', kernel_initializer=Orthogonal(gain=1.0), name='dense_3',kernel_regularizer=regularizers.l2(0.0005))) pre_256_model = load_model('./Experiments/PH_1/EX_40/EX_40.hdf5', custom_objects={'RMSPooling2D': RMSPooling2D,'Lambda':Lambda,'tf':tf}) # copy weights from old model to new one for layer in model.layers:
try: layer.set_weights( pre_256_model.get_layer(name=layer.name).get_weights()) print("Succesfully transfered weights for layer {}".format(layer.name)) except: print("Could not transfer weights for layer {}".format(layer.name))
conditional_block
mongodb.go
StringIDToObjID(parsedFilter["_id"].(string)) if err != nil { /* log */ goto Done } } itemsMap, err = mh.getDocs(dbname,collectionname,parsedFilter) if err != nil { //log } Done: return itemsMap, err } func (mh *mongoDBHandler) InsertOne(dbname string, collectionname string, doc interface{}) ([]map[string]interface{}, error) { var itemsMap []map[string]interface{} var err error db := mh.MongoClient.Database(dbname) collection := db.Collection(collectionname) var insDoc = map[string]interface{}{} insDoc = c.ConvertInterfaceToMapStringInterface(doc) // check if "_id" is part of request, convert accordingly if _, ok := insDoc["_id"]; ok { insDoc["_id"], err = ConvertStringIDToObjID(insDoc["_id"].(string)) if err != nil { /* log */ return itemsMap, err } } var res *mongo.InsertOneResult if err == nil { now := time.Now() insDoc["createdAt"] = now insDoc["updatedAt"] = now // insDoc["owner"] = now res, err = collection.InsertOne(context.TODO(), insDoc) if err == nil { fmt.Printf("inserted document with ID %v\n", res.InsertedID.(primitive.ObjectID).Hex()) itemsMap, err = mh.getDocs(dbname,collectionname,bson.M{"_id":res.InsertedID}) fmt.Println("Inserted doc: ",itemsMap) } else { //v, _ := err.(type) hasDupEntry, msgToPass := containsWriteErrDupEntry(err) if hasDupEntry { err = c.ItemAlreadyExistsError{msgToPass} } } } fmt.Printf("InsertOne: ItemsMap: %s\n InsertOne: err: %s\n",itemsMap,err) return itemsMap, err } func (mh *mongoDBHandler) getDocs(dbname string, collectionname string, filter interface{}) ([]map[string]interface{}, error) { fmt.Println("== getDocs") db := mh.MongoClient.Database(dbname) collection := db.Collection(collectionname) var result bson.M var results []bson.M var itemsMap []map[string]interface{} fmt.Println("=== filter: ",filter) cursor, err := collection.Find(context.TODO(),filter) if err != nil { goto Done } if err = cursor.All(context.TODO(), &results); err != nil { goto Done } if len(results) <= 0 { fmt.Println("No doc found") } else { fmt.Println("Doc(s) found:") for _, result = range results { var itemMap map[string]interface{} b, _ := bson.Marshal(result) bson.Unmarshal(b, &itemMap) itemMap["_id"] = itemMap["_id"].(primitive.ObjectID).Hex() fmt.Printf("itemMap after id: %v\n",itemMap) itemsMap = append(itemsMap, itemMap) } } Done: fmt.Printf("itemsMap: %v\n",itemsMap) for k, v := range itemsMap { fmt.Println("itemsMap[",k,"]=",v) } fmt.Println("== /getDocs") return itemsMap, err } func (mh *mongoDBHandler) DeleteDoc(dbname string, collectionname string, filter interface{}) ([]map[string]interface{}, error) { var itemsMap []map[string]interface{} var err error db := mh.MongoClient.Database(dbname) collection := db.Collection(collectionname) var res *mongo.DeleteResult //unify filter before passing to actual query var parsedFilter = map[string]interface{}{} parsedFilter=c.ConvertInterfaceToMapStringInterface(filter) // check if "_id" is part of filter, convert accordingly if _, ok := parsedFilter["_id"]; ok { parsedFilter["_id"], err = ConvertStringIDToObjID(parsedFilter["_id"].(string)) if err != nil { goto Done } } // grab doc to be deleted, so it can be provided in response for reference itemsMap, err = mh.getDocs(dbname,collectionname,parsedFilter) if err != nil { goto Done } fmt.Println("== DeleteDoc, doc to be deleted: ", itemsMap) res, err = collection.DeleteOne(context.TODO(), parsedFilter) fmt.Printf("== DeleteDoc, deleted %v documents\n", res.DeletedCount) if res.DeletedCount == 0 { err = c.NotFoundError{ fmt.Sprintf( "not found: %s", parsedFilter) } } Done: return itemsMap, err } func (mh *mongoDBHandler) UpdateDoc(dbname string, collectionname string, filter interface{}, doc interface{}) ([]map[string]interface{}, error) { var itemsMap []map[string]interface{} var err error db := mh.MongoClient.Database(dbname) collection := db.Collection(collectionname) var res *mongo.UpdateResult //unify filter before passing to actual query var parsedFilter = map[string]interface{}{} parsedFilter=c.ConvertInterfaceToMapStringInterface(filter) // check if "_id" is part of filter, convert accordingly if _, ok := parsedFilter["_id"]; ok { parsedFilter["_id"], err = ConvertStringIDToObjID(parsedFilter["_id"].(string)) if err != nil { return nil, err } } // parse doc into proper MongoDB update specification // basically: { "$set" : {doc}} var updateDoc = map[string]interface{}{} updateDoc["$set"]=doc updateDoc["$set"].(map[string]interface{})["updatedAt"]=time.Now() res, err = collection.UpdateOne(context.TODO(), parsedFilter, updateDoc) if err != nil { goto Done } if res.MatchedCount != 0 { itemsMap, err = mh.getDocs(dbname,collectionname,parsedFilter) fmt.Printf("Updated existing document %v\n for filter %v\n", itemsMap, parsedFilter) } else { err = c.NotFoundError{ fmt.Sprintf( "not found _id : %s", parsedFilter["_id"].(primitive.ObjectID).Hex() ) } fmt.Printf("No document updated for filter %v\n", parsedFilter) } Done: return itemsMap, err } func listExistingIndexes(coll *mongo.Collection){ //var indexView *mongo.IndexView indexView := coll.Indexes() // Specify the MaxTime option to limit the amount of time the operation can run on the server opts := options.ListIndexes().SetMaxTime(2 * time.Second) cursor, err := indexView.List(context.TODO(), opts) if err != nil { log.Fatal(err) } // Get a slice of all indexes returned and print them out. var results []bson.M if err = cursor.All(context.TODO(), &results); err != nil { log.Fatal(err) } fmt.Println(results) } func initSysIndexes(coll *mongo.Collection, iModel []mongo.IndexModel) { indexName, err := coll.Indexes().CreateMany( context.Background(), iModel, ) fmt.Println("indexName: ",indexName, " err: ",err) } var sysProjIndexModels = []mongo.IndexModel{ { Keys: bson.D{{"name", 1},{"owner", 1}}, }, { Keys: bson.D{{"name", 1}}, Options: options.Index().SetUnique(true), }, } var userIndexModels = []mongo.IndexModel{ { Keys: bson.D{{"name", 1}}, Options: options.Index().SetUnique(true), }, } var authIndexModels = []mongo.IndexModel{ { Keys: bson.D{{"uuid", 1}}, // small probability for non-unique entry, just to be on safe side Options: options.Index().SetUnique(true), }, } // private // if err returned from mongo write operation contains duplicate entry // e.g. breaking unique index // returns true/false and the "{...}" part of original error message if true func containsWriteErrDupEntry(err error) (bool, string) {
containsDup := false var errMsg string if v, ok := err.(mongo.WriteException); ok { var msgs []string for idx, werr:=range v.WriteErrors { //log stuff before anything gets altered fmt.Println("err.WriteErrors[",idx,"].Index=",werr.Index) fmt.Println("err.WriteErrors[",idx,"].Code=",werr.Code) fmt.Println("err.WriteErrors[",idx,"].Message=",werr.Message) // err code 11000 or 11001 in MongoDB indicates duplicate key if werr.Code == 11000 || werr.Code == 11001 { containsDup = true // get the dup key msg pat := regexp.MustCompile(`({)(.*?)(})`) msgs = append(msgs,pat.FindString(werr.Message)) } } fmt.Println("-- ",msgs) //errMsg = c.Lines2JSONString(&msgs)
identifier_body
mongodb.go
updatedAt" : now, }) } initSysIndexes(mh.MongoProjectsDB,sysProjIndexModels) initSysIndexes(mh.MongoProjectsDB,userIndexModels) initSysIndexes(mh.MongoClient.Database(c.UsersDBName).Collection(c.UsersDBAuthCollection),authIndexModels) } func (mh *mongoDBHandler) initSystemUsers() ([]map[string]interface{}, error) { var err error var sysAdmin m.UserDBModel var now time.Time var hashedPassword []byte // check if default system admin exists itemsMap, err := mh.GetDocs(c.UsersDBName, c.UsersDBUsersCollection, bson.M{"login":"sysadmin"}) if err !=nil { if _, ok := err.(c.NotFoundError); !ok { goto Done } } if len(itemsMap) == 0 { //insert default admin user hashedPassword, err = bcrypt.GenerateFromPassword([]byte("sysadmin"), 12) if err != nil {goto Done} now = time.Now() sysAdmin = m.UserDBModel{ Login: "sysadmin", Hash: string(hashedPassword), CreatedAt: now, UpdatedAt: now, } if itemsMap, err = mh.InsertOne(c.UsersDBName, c.UsersDBUsersCollection, sysAdmin); err != nil { goto Done } } Done: return itemsMap, err } func ConvertStringIDToObjID(stringID string) (primitive.ObjectID, error) { oid, err := primitive.ObjectIDFromHex(stringID) if err != nil { err = c.InvalidIdError{stringID} } //Maybe wrap original error? return oid, err } // returns slice of acquired docs or error func (mh *mongoDBHandler) GetDocs(dbname string, collectionname string, filter interface{}) ([]map[string]interface{}, error) { fmt.Println("=> GetDocs, filter: %v", filter) var err error var itemsMap []map[string]interface{} //unify filter before passing to actual query var parsedFilter = map[string]interface{}{} parsedFilter = c.ConvertInterfaceToMapStringInterface(filter) // check if "_id" is part of filter, convert accordingly if _, ok := parsedFilter["_id"]; ok { parsedFilter["_id"], err = ConvertStringIDToObjID(parsedFilter["_id"].(string)) if err != nil { /* log */ goto Done } } itemsMap, err = mh.getDocs(dbname,collectionname,parsedFilter) if err != nil { //log } Done: return itemsMap, err } func (mh *mongoDBHandler) InsertOne(dbname string, collectionname string, doc interface{}) ([]map[string]interface{}, error) { var itemsMap []map[string]interface{} var err error db := mh.MongoClient.Database(dbname) collection := db.Collection(collectionname) var insDoc = map[string]interface{}{} insDoc = c.ConvertInterfaceToMapStringInterface(doc) // check if "_id" is part of request, convert accordingly if _, ok := insDoc["_id"]; ok { insDoc["_id"], err = ConvertStringIDToObjID(insDoc["_id"].(string)) if err != nil { /* log */ return itemsMap, err } } var res *mongo.InsertOneResult if err == nil { now := time.Now() insDoc["createdAt"] = now insDoc["updatedAt"] = now // insDoc["owner"] = now res, err = collection.InsertOne(context.TODO(), insDoc) if err == nil { fmt.Printf("inserted document with ID %v\n", res.InsertedID.(primitive.ObjectID).Hex()) itemsMap, err = mh.getDocs(dbname,collectionname,bson.M{"_id":res.InsertedID}) fmt.Println("Inserted doc: ",itemsMap) } else { //v, _ := err.(type) hasDupEntry, msgToPass := containsWriteErrDupEntry(err) if hasDupEntry { err = c.ItemAlreadyExistsError{msgToPass} } } } fmt.Printf("InsertOne: ItemsMap: %s\n InsertOne: err: %s\n",itemsMap,err) return itemsMap, err } func (mh *mongoDBHandler) getDocs(dbname string, collectionname string, filter interface{}) ([]map[string]interface{}, error) { fmt.Println("== getDocs") db := mh.MongoClient.Database(dbname) collection := db.Collection(collectionname) var result bson.M var results []bson.M var itemsMap []map[string]interface{} fmt.Println("=== filter: ",filter) cursor, err := collection.Find(context.TODO(),filter) if err != nil { goto Done } if err = cursor.All(context.TODO(), &results); err != nil { goto Done } if len(results) <= 0 { fmt.Println("No doc found") } else { fmt.Println("Doc(s) found:") for _, result = range results { var itemMap map[string]interface{} b, _ := bson.Marshal(result) bson.Unmarshal(b, &itemMap) itemMap["_id"] = itemMap["_id"].(primitive.ObjectID).Hex() fmt.Printf("itemMap after id: %v\n",itemMap) itemsMap = append(itemsMap, itemMap) } } Done: fmt.Printf("itemsMap: %v\n",itemsMap) for k, v := range itemsMap { fmt.Println("itemsMap[",k,"]=",v) } fmt.Println("== /getDocs") return itemsMap, err } func (mh *mongoDBHandler) DeleteDoc(dbname string, collectionname string, filter interface{}) ([]map[string]interface{}, error) { var itemsMap []map[string]interface{} var err error db := mh.MongoClient.Database(dbname) collection := db.Collection(collectionname) var res *mongo.DeleteResult //unify filter before passing to actual query var parsedFilter = map[string]interface{}{} parsedFilter=c.ConvertInterfaceToMapStringInterface(filter) // check if "_id" is part of filter, convert accordingly if _, ok := parsedFilter["_id"]; ok { parsedFilter["_id"], err = ConvertStringIDToObjID(parsedFilter["_id"].(string)) if err != nil { goto Done } } // grab doc to be deleted, so it can be provided in response for reference itemsMap, err = mh.getDocs(dbname,collectionname,parsedFilter) if err != nil { goto Done } fmt.Println("== DeleteDoc, doc to be deleted: ", itemsMap) res, err = collection.DeleteOne(context.TODO(), parsedFilter) fmt.Printf("== DeleteDoc, deleted %v documents\n", res.DeletedCount) if res.DeletedCount == 0 { err = c.NotFoundError{ fmt.Sprintf( "not found: %s", parsedFilter) } } Done: return itemsMap, err } func (mh *mongoDBHandler) UpdateDoc(dbname string, collectionname string, filter interface{}, doc interface{}) ([]map[string]interface{}, error) { var itemsMap []map[string]interface{} var err error db := mh.MongoClient.Database(dbname) collection := db.Collection(collectionname) var res *mongo.UpdateResult //unify filter before passing to actual query var parsedFilter = map[string]interface{}{} parsedFilter=c.ConvertInterfaceToMapStringInterface(filter) // check if "_id" is part of filter, convert accordingly if _, ok := parsedFilter["_id"]; ok { parsedFilter["_id"], err = ConvertStringIDToObjID(parsedFilter["_id"].(string)) if err != nil { return nil, err } } // parse doc into proper MongoDB update specification // basically: { "$set" : {doc}} var updateDoc = map[string]interface{}{} updateDoc["$set"]=doc updateDoc["$set"].(map[string]interface{})["updatedAt"]=time.Now() res, err = collection.UpdateOne(context.TODO(), parsedFilter, updateDoc) if err != nil { goto Done } if res.MatchedCount != 0 { itemsMap, err = mh.getDocs(dbname,collectionname,parsedFilter) fmt.Printf("Updated existing document %v\n for filter %v\n", itemsMap, parsedFilter) } else { err = c.NotFoundError{ fmt.Sprintf( "not found _id : %s", parsedFilter["_id"].(primitive.ObjectID).Hex() ) } fmt.Printf("No document updated for filter %v\n", parsedFilter) } Done: return itemsMap, err } func listExistingIndexes(coll *mongo.Collection){ //var indexView *mongo.IndexView indexView := coll.Indexes() // Specify the MaxTime option to limit the amount of time the operation can run on the server opts := options.ListIndexes().SetMaxTime(2 * time.Second) cursor, err := indexView.List(context.TODO(), opts) if err != nil { log.Fatal(err) } // Get a slice of all indexes returned and print them out. var results []bson.M if err = cursor.All(context.TODO(), &results); err != nil { log.Fatal(err) } fmt.Println(results)
}
random_line_split
mongodb.go
:= mh.GetDocs(c.SysDBName, appConfig.DBConfig.MongoConfig.ProjectsColl, bson.M{"name":"trion"}) if err != nil { if _, ok := err.(c.NotFoundError); !ok
fmt.Println("To był not found, len(itemsMap)",len(itemsMap)) } now := time.Now() if len(itemsMap) == 0 { fmt.Println("Inserting Trion...") mh.InsertOne(c.SysDBName, appConfig.DBConfig.MongoConfig.ProjectsColl, bson.M{"name" : "trion", "type" : "system", "schema_rev": "1", "owner":users[0]["_id"].(string), "createdAt" : now, "updatedAt" : now, }) } initSysIndexes(mh.MongoProjectsDB,sysProjIndexModels) initSysIndexes(mh.MongoProjectsDB,userIndexModels) initSysIndexes(mh.MongoClient.Database(c.UsersDBName).Collection(c.UsersDBAuthCollection),authIndexModels) } func (mh *mongoDBHandler) initSystemUsers() ([]map[string]interface{}, error) { var err error var sysAdmin m.UserDBModel var now time.Time var hashedPassword []byte // check if default system admin exists itemsMap, err := mh.GetDocs(c.UsersDBName, c.UsersDBUsersCollection, bson.M{"login":"sysadmin"}) if err !=nil { if _, ok := err.(c.NotFoundError); !ok { goto Done } } if len(itemsMap) == 0 { //insert default admin user hashedPassword, err = bcrypt.GenerateFromPassword([]byte("sysadmin"), 12) if err != nil {goto Done} now = time.Now() sysAdmin = m.UserDBModel{ Login: "sysadmin", Hash: string(hashedPassword), CreatedAt: now, UpdatedAt: now, } if itemsMap, err = mh.InsertOne(c.UsersDBName, c.UsersDBUsersCollection, sysAdmin); err != nil { goto Done } } Done: return itemsMap, err } func ConvertStringIDToObjID(stringID string) (primitive.ObjectID, error) { oid, err := primitive.ObjectIDFromHex(stringID) if err != nil { err = c.InvalidIdError{stringID} } //Maybe wrap original error? return oid, err } // returns slice of acquired docs or error func (mh *mongoDBHandler) GetDocs(dbname string, collectionname string, filter interface{}) ([]map[string]interface{}, error) { fmt.Println("=> GetDocs, filter: %v", filter) var err error var itemsMap []map[string]interface{} //unify filter before passing to actual query var parsedFilter = map[string]interface{}{} parsedFilter = c.ConvertInterfaceToMapStringInterface(filter) // check if "_id" is part of filter, convert accordingly if _, ok := parsedFilter["_id"]; ok { parsedFilter["_id"], err = ConvertStringIDToObjID(parsedFilter["_id"].(string)) if err != nil { /* log */ goto Done } } itemsMap, err = mh.getDocs(dbname,collectionname,parsedFilter) if err != nil { //log } Done: return itemsMap, err } func (mh *mongoDBHandler) InsertOne(dbname string, collectionname string, doc interface{}) ([]map[string]interface{}, error) { var itemsMap []map[string]interface{} var err error db := mh.MongoClient.Database(dbname) collection := db.Collection(collectionname) var insDoc = map[string]interface{}{} insDoc = c.ConvertInterfaceToMapStringInterface(doc) // check if "_id" is part of request, convert accordingly if _, ok := insDoc["_id"]; ok { insDoc["_id"], err = ConvertStringIDToObjID(insDoc["_id"].(string)) if err != nil { /* log */ return itemsMap, err } } var res *mongo.InsertOneResult if err == nil { now := time.Now() insDoc["createdAt"] = now insDoc["updatedAt"] = now // insDoc["owner"] = now res, err = collection.InsertOne(context.TODO(), insDoc) if err == nil { fmt.Printf("inserted document with ID %v\n", res.InsertedID.(primitive.ObjectID).Hex()) itemsMap, err = mh.getDocs(dbname,collectionname,bson.M{"_id":res.InsertedID}) fmt.Println("Inserted doc: ",itemsMap) } else { //v, _ := err.(type) hasDupEntry, msgToPass := containsWriteErrDupEntry(err) if hasDupEntry { err = c.ItemAlreadyExistsError{msgToPass} } } } fmt.Printf("InsertOne: ItemsMap: %s\n InsertOne: err: %s\n",itemsMap,err) return itemsMap, err } func (mh *mongoDBHandler) getDocs(dbname string, collectionname string, filter interface{}) ([]map[string]interface{}, error) { fmt.Println("== getDocs") db := mh.MongoClient.Database(dbname) collection := db.Collection(collectionname) var result bson.M var results []bson.M var itemsMap []map[string]interface{} fmt.Println("=== filter: ",filter) cursor, err := collection.Find(context.TODO(),filter) if err != nil { goto Done } if err = cursor.All(context.TODO(), &results); err != nil { goto Done } if len(results) <= 0 { fmt.Println("No doc found") } else { fmt.Println("Doc(s) found:") for _, result = range results { var itemMap map[string]interface{} b, _ := bson.Marshal(result) bson.Unmarshal(b, &itemMap) itemMap["_id"] = itemMap["_id"].(primitive.ObjectID).Hex() fmt.Printf("itemMap after id: %v\n",itemMap) itemsMap = append(itemsMap, itemMap) } } Done: fmt.Printf("itemsMap: %v\n",itemsMap) for k, v := range itemsMap { fmt.Println("itemsMap[",k,"]=",v) } fmt.Println("== /getDocs") return itemsMap, err } func (mh *mongoDBHandler) DeleteDoc(dbname string, collectionname string, filter interface{}) ([]map[string]interface{}, error) { var itemsMap []map[string]interface{} var err error db := mh.MongoClient.Database(dbname) collection := db.Collection(collectionname) var res *mongo.DeleteResult //unify filter before passing to actual query var parsedFilter = map[string]interface{}{} parsedFilter=c.ConvertInterfaceToMapStringInterface(filter) // check if "_id" is part of filter, convert accordingly if _, ok := parsedFilter["_id"]; ok { parsedFilter["_id"], err = ConvertStringIDToObjID(parsedFilter["_id"].(string)) if err != nil { goto Done } } // grab doc to be deleted, so it can be provided in response for reference itemsMap, err = mh.getDocs(dbname,collectionname,parsedFilter) if err != nil { goto Done } fmt.Println("== DeleteDoc, doc to be deleted: ", itemsMap) res, err = collection.DeleteOne(context.TODO(), parsedFilter) fmt.Printf("== DeleteDoc, deleted %v documents\n", res.DeletedCount) if res.DeletedCount == 0 { err = c.NotFoundError{ fmt.Sprintf( "not found: %s", parsedFilter) } } Done: return itemsMap, err } func (mh *mongoDBHandler) UpdateDoc(dbname string, collectionname string, filter interface{}, doc interface{}) ([]map[string]interface{}, error) { var itemsMap []map[string]interface{} var err error db := mh.MongoClient.Database(dbname) collection := db.Collection(collectionname) var res *mongo.UpdateResult //unify filter before passing to actual query var parsedFilter = map[string]interface{}{} parsedFilter=c.ConvertInterfaceToMapStringInterface(filter) // check if "_id" is part of filter, convert accordingly if _, ok := parsedFilter["_id"]; ok { parsedFilter["_id"], err = ConvertStringIDToObjID(parsedFilter["_id"].(string)) if err != nil { return nil, err } } // parse doc into proper MongoDB update specification // basically: { "$set" : {doc}} var updateDoc = map[string]interface{}{} updateDoc["$set"]=doc updateDoc["$set"].(map[string]interface{})["updatedAt"]=time.Now() res, err = collection.UpdateOne(context.TODO(), parsedFilter, updateDoc) if err != nil { goto Done } if res.MatchedCount != 0 { itemsMap, err = mh.getDocs(dbname,collectionname,parsedFilter) fmt.Printf("Updated existing document %v\n for filter %v\n", itemsMap, parsedFilter) } else { err = c.NotFoundError{ fmt.Sprintf( "not found _id :
{ log.Fatal(err) }
conditional_block
mongodb.go
:= mh.GetDocs(c.SysDBName, appConfig.DBConfig.MongoConfig.ProjectsColl, bson.M{"name":"trion"}) if err != nil { if _, ok := err.(c.NotFoundError); !ok { log.Fatal(err) } fmt.Println("To był not found, len(itemsMap)",len(itemsMap)) } now := time.Now() if len(itemsMap) == 0 { fmt.Println("Inserting Trion...") mh.InsertOne(c.SysDBName, appConfig.DBConfig.MongoConfig.ProjectsColl, bson.M{"name" : "trion", "type" : "system", "schema_rev": "1", "owner":users[0]["_id"].(string), "createdAt" : now, "updatedAt" : now, }) } initSysIndexes(mh.MongoProjectsDB,sysProjIndexModels) initSysIndexes(mh.MongoProjectsDB,userIndexModels) initSysIndexes(mh.MongoClient.Database(c.UsersDBName).Collection(c.UsersDBAuthCollection),authIndexModels) } func (mh *mongoDBHandler) i
) ([]map[string]interface{}, error) { var err error var sysAdmin m.UserDBModel var now time.Time var hashedPassword []byte // check if default system admin exists itemsMap, err := mh.GetDocs(c.UsersDBName, c.UsersDBUsersCollection, bson.M{"login":"sysadmin"}) if err !=nil { if _, ok := err.(c.NotFoundError); !ok { goto Done } } if len(itemsMap) == 0 { //insert default admin user hashedPassword, err = bcrypt.GenerateFromPassword([]byte("sysadmin"), 12) if err != nil {goto Done} now = time.Now() sysAdmin = m.UserDBModel{ Login: "sysadmin", Hash: string(hashedPassword), CreatedAt: now, UpdatedAt: now, } if itemsMap, err = mh.InsertOne(c.UsersDBName, c.UsersDBUsersCollection, sysAdmin); err != nil { goto Done } } Done: return itemsMap, err } func ConvertStringIDToObjID(stringID string) (primitive.ObjectID, error) { oid, err := primitive.ObjectIDFromHex(stringID) if err != nil { err = c.InvalidIdError{stringID} } //Maybe wrap original error? return oid, err } // returns slice of acquired docs or error func (mh *mongoDBHandler) GetDocs(dbname string, collectionname string, filter interface{}) ([]map[string]interface{}, error) { fmt.Println("=> GetDocs, filter: %v", filter) var err error var itemsMap []map[string]interface{} //unify filter before passing to actual query var parsedFilter = map[string]interface{}{} parsedFilter = c.ConvertInterfaceToMapStringInterface(filter) // check if "_id" is part of filter, convert accordingly if _, ok := parsedFilter["_id"]; ok { parsedFilter["_id"], err = ConvertStringIDToObjID(parsedFilter["_id"].(string)) if err != nil { /* log */ goto Done } } itemsMap, err = mh.getDocs(dbname,collectionname,parsedFilter) if err != nil { //log } Done: return itemsMap, err } func (mh *mongoDBHandler) InsertOne(dbname string, collectionname string, doc interface{}) ([]map[string]interface{}, error) { var itemsMap []map[string]interface{} var err error db := mh.MongoClient.Database(dbname) collection := db.Collection(collectionname) var insDoc = map[string]interface{}{} insDoc = c.ConvertInterfaceToMapStringInterface(doc) // check if "_id" is part of request, convert accordingly if _, ok := insDoc["_id"]; ok { insDoc["_id"], err = ConvertStringIDToObjID(insDoc["_id"].(string)) if err != nil { /* log */ return itemsMap, err } } var res *mongo.InsertOneResult if err == nil { now := time.Now() insDoc["createdAt"] = now insDoc["updatedAt"] = now // insDoc["owner"] = now res, err = collection.InsertOne(context.TODO(), insDoc) if err == nil { fmt.Printf("inserted document with ID %v\n", res.InsertedID.(primitive.ObjectID).Hex()) itemsMap, err = mh.getDocs(dbname,collectionname,bson.M{"_id":res.InsertedID}) fmt.Println("Inserted doc: ",itemsMap) } else { //v, _ := err.(type) hasDupEntry, msgToPass := containsWriteErrDupEntry(err) if hasDupEntry { err = c.ItemAlreadyExistsError{msgToPass} } } } fmt.Printf("InsertOne: ItemsMap: %s\n InsertOne: err: %s\n",itemsMap,err) return itemsMap, err } func (mh *mongoDBHandler) getDocs(dbname string, collectionname string, filter interface{}) ([]map[string]interface{}, error) { fmt.Println("== getDocs") db := mh.MongoClient.Database(dbname) collection := db.Collection(collectionname) var result bson.M var results []bson.M var itemsMap []map[string]interface{} fmt.Println("=== filter: ",filter) cursor, err := collection.Find(context.TODO(),filter) if err != nil { goto Done } if err = cursor.All(context.TODO(), &results); err != nil { goto Done } if len(results) <= 0 { fmt.Println("No doc found") } else { fmt.Println("Doc(s) found:") for _, result = range results { var itemMap map[string]interface{} b, _ := bson.Marshal(result) bson.Unmarshal(b, &itemMap) itemMap["_id"] = itemMap["_id"].(primitive.ObjectID).Hex() fmt.Printf("itemMap after id: %v\n",itemMap) itemsMap = append(itemsMap, itemMap) } } Done: fmt.Printf("itemsMap: %v\n",itemsMap) for k, v := range itemsMap { fmt.Println("itemsMap[",k,"]=",v) } fmt.Println("== /getDocs") return itemsMap, err } func (mh *mongoDBHandler) DeleteDoc(dbname string, collectionname string, filter interface{}) ([]map[string]interface{}, error) { var itemsMap []map[string]interface{} var err error db := mh.MongoClient.Database(dbname) collection := db.Collection(collectionname) var res *mongo.DeleteResult //unify filter before passing to actual query var parsedFilter = map[string]interface{}{} parsedFilter=c.ConvertInterfaceToMapStringInterface(filter) // check if "_id" is part of filter, convert accordingly if _, ok := parsedFilter["_id"]; ok { parsedFilter["_id"], err = ConvertStringIDToObjID(parsedFilter["_id"].(string)) if err != nil { goto Done } } // grab doc to be deleted, so it can be provided in response for reference itemsMap, err = mh.getDocs(dbname,collectionname,parsedFilter) if err != nil { goto Done } fmt.Println("== DeleteDoc, doc to be deleted: ", itemsMap) res, err = collection.DeleteOne(context.TODO(), parsedFilter) fmt.Printf("== DeleteDoc, deleted %v documents\n", res.DeletedCount) if res.DeletedCount == 0 { err = c.NotFoundError{ fmt.Sprintf( "not found: %s", parsedFilter) } } Done: return itemsMap, err } func (mh *mongoDBHandler) UpdateDoc(dbname string, collectionname string, filter interface{}, doc interface{}) ([]map[string]interface{}, error) { var itemsMap []map[string]interface{} var err error db := mh.MongoClient.Database(dbname) collection := db.Collection(collectionname) var res *mongo.UpdateResult //unify filter before passing to actual query var parsedFilter = map[string]interface{}{} parsedFilter=c.ConvertInterfaceToMapStringInterface(filter) // check if "_id" is part of filter, convert accordingly if _, ok := parsedFilter["_id"]; ok { parsedFilter["_id"], err = ConvertStringIDToObjID(parsedFilter["_id"].(string)) if err != nil { return nil, err } } // parse doc into proper MongoDB update specification // basically: { "$set" : {doc}} var updateDoc = map[string]interface{}{} updateDoc["$set"]=doc updateDoc["$set"].(map[string]interface{})["updatedAt"]=time.Now() res, err = collection.UpdateOne(context.TODO(), parsedFilter, updateDoc) if err != nil { goto Done } if res.MatchedCount != 0 { itemsMap, err = mh.getDocs(dbname,collectionname,parsedFilter) fmt.Printf("Updated existing document %v\n for filter %v\n", itemsMap, parsedFilter) } else { err = c.NotFoundError{ fmt.Sprintf( "not found _id :
nitSystemUsers(
identifier_name
startnode.go
ODEID string) { var ( CannotRead = make(chan bool, 1) GetName = make(chan bool, 1) stdin io.Writer stdout io.Reader ) for { AdminData, err := utils.ExtractPayload(*connToAdmin, AgentStatus.AESKey, NODEID, false) if err != nil { AdminOffline(reConn, monitor, listenPort, passive) go SendInfo(NODEID) //重连后发送自身信息 go SendNote(NODEID) //重连后发送admin设置的备忘 continue } if AdminData.NodeId == NODEID { switch AdminData.Type { case "DATA": switch AdminData.Command { case "SOCKSDATA": SocksDataChanMap.RLock() if _, ok := SocksDataChanMap.Payload[AdminData.Clientid]; ok { SocksDataChanMap.Payload[AdminData.Clientid] <- AdminData.Info SocksDataChanMap.RUnlock() } else { SocksDataChanMap.RUnlock() SocksDataChanMap.Lock() SocksDataChanMap.Payload[AdminData.Clientid] = make(chan string, 1) go HanleClientSocksConn(SocksDataChanMap.Payload[AdminData.Clientid], SocksInfo.SocksUsername, SocksInfo.SocksPass, AdminData.Clientid, NODEID) SocksDataChanMap.Payload[AdminData.Clientid] <- AdminData.Info SocksDataChanMap.Unlock() } case "FILEDATA": //接收文件内容 slicenum, _ := strconv.Atoi(AdminData.FileSliceNum) FileDataMap.Lock() FileDataMap.Payload[slicenum] = AdminData.Info FileDataMap.Unlock() case "FORWARD": TryForward(AdminData.Info, AdminData.Clientid) case "FORWARDDATA": ForwardConnMap.RLock() if _, ok := ForwardConnMap.Payload[AdminData.Clientid]; ok { PortFowardMap.Lock() if _, ok := PortFowardMap.Payload[AdminData.Clientid]; ok { PortFowardMap.Payload[AdminData.Clientid] <- AdminData.Info } else { PortFowardMap.Payload[AdminData.Clientid] = make(chan string, 1) go HandleForward(PortFowardMap.Payload[AdminData.Clientid], AdminData.Clientid) PortFowardMap.Payload[AdminData.Clientid] <- AdminData.Info } PortFowardMap.Unlock() } ForwardConnMap.RUnlock() case "FORWARDFIN": ForwardConnMap.Lock() if _, ok := ForwardConnMap.Payload[AdminData.Clientid]; ok { ForwardConnMap.Payload[AdminData.Clientid].Close() delete(ForwardConnMap.Payload, AdminData.Clientid) } ForwardConnMap.Unlock() PortFowardMap.Lock() if _, ok := PortFowardMap.Payload[AdminData.Clientid]; ok { if !utils.IsClosed(PortFowardMap.Payload[AdminData.Clientid]) { if !utils.IsClosed(PortFowardMap.Payload[AdminData.Clientid]) { close(PortFowardMap.Payload[AdminData.Clientid]) } delete(PortFowardMap.Payload, AdminData.Clientid) } } PortFowardMap.Unlock() case "REFLECTDATARESP": ReflectConnMap.Lock() ReflectConnMap.Payload[AdminData.Clientid].Write([]byte(AdminData.Info)) ReflectConnMap.Unlock() case "REFLECTTIMEOUT": fallthrough case "REFLECTOFFLINE": ReflectConnMap.Lock() if _, ok := ReflectConnMap.Payload[AdminData.Clientid]; ok { ReflectConnMap.Payload[AdminData.Clientid].Close() delete(ReflectConnMap.Payload, AdminData.Clientid) } ReflectConnMap.Unlock() case "FINOK": SocksDataChanMap.Lock() //性能损失? if _, ok := SocksDataChanMap.Payload[AdminData.Clientid]; ok { if !utils.IsClosed(SocksDataChanMap.Payload[AdminData.Clientid]) { close(SocksDataChanMap.Payload[AdminData.Clientid]) } delete(SocksDataChanMap.Payload, AdminData.Clientid) } SocksDataChanMap.Unlock() case "FIN": CurrentConn.Lock() if _, ok := CurrentConn.Payload[AdminData.Clientid]; ok { CurrentConn.Payload[AdminData.Clientid].Close() delete(CurrentConn.Payload, AdminData.Clientid) } CurrentConn.Unlock() SocksDataChanMap.Lock() if _, ok := SocksDataChanMap.Payload[AdminData.Clientid]; ok { if !utils.IsClosed(SocksDataChanMap.Payload[AdminData.Clientid]) { close(SocksDataChanMap.Payload[AdminData.Clientid]) } delete(SocksDataChanMap.Payload, AdminData.Clientid) } SocksDataChanMap.Unlock() case "HEARTBEAT": hbdatapack, _ := utils.ConstructPayload(utils.AdminId, "", "COMMAND", "KEEPALIVE", " ", " ", 0, NODEID, AgentStatus.AESKey, false) ProxyChan.ProxyChanToUpperNode <- hbdatapack default: continue } case "COMMAND": switch AdminData.Command { case "SHELL": switch AdminData.Info { case "": stdout, stdin = CreatInteractiveShell() go func() { StartShell("", stdin, stdout, NODEID) }() case "exit\n": fallthrough default: go func() { StartShell(AdminData.Info, stdin, stdout, NODEID) }() } case "SOCKS": socksinfo := strings.Split(AdminData.Info, ":::") SocksInfo.SocksUsername = socksinfo[1] SocksInfo.SocksPass = socksinfo[2] StartSocks() case "SOCKSOFF": case "SSH": err := StartSSH(AdminData.Info, NODEID) if err == nil { go ReadCommand() } else { break } case "SSHCOMMAND": go WriteCommand(AdminData.Info) case "SSHTUNNEL": err := SSHTunnelNextNode(AdminData.Info, NODEID) if err != nil { fmt.Println("[*]", err) break } case "CONNECT": var status bool = false command := strings.Split(AdminData.Info, ":::") addr := command[0] choice := command[1] if choice == "1" { //连接的节点是否是在reuseport? status = node.ConnectNextNodeReuse(addr, NODEID, AgentStatus.AESKey) } else { status = node.ConnectNextNode(addr, NODEID, AgentStatus.AESKey) } if !status { message, _ := utils.ConstructPayload(utils.AdminId, "", "COMMAND", "NODECONNECTFAIL", " ", "", 0, NODEID, AgentStatus.AESKey, false) ProxyChan.ProxyChanToUpperNode <- message } case "FILENAME": var err error UploadFile, err := os.Create(AdminData.Info) if err != nil { respComm, _ := utils.ConstructPayload(utils.AdminId, "", "COMMAND", "CREATEFAIL", " ", " ", 0, NODEID, AgentStatus.AESKey, false) ProxyChan.ProxyChanToUpperNode <- respComm } else { respComm, _ := utils.ConstructPayload(utils.AdminId, "", "COMMAND", "NAMECONFIRM", " ", " ", 0, NODEID, AgentStatus.AESKey, false) ProxyChan.ProxyChanToUpperNode <- respComm go share.ReceiveFile("", connToAdmin, FileDataMap, CannotRead, UploadFile, AgentStatus.AESKey, false, NODEID) } case "FILESIZE": filesize, _ := strconv.ParseInt(AdminData.Info, 10, 64) share.File.FileSize = filesize respComm, _ := utils.ConstructPayload(utils.AdminId, "", "COMMAND", "FILESIZECONFIRM", " ", " ", 0, NODEID, AgentStatus.AESKey, false) ProxyChan.ProxyChanToUpperNode <- respComm share.File.ReceiveFileSize <- true case "FILESLICENUM": share.File.TotalSilceNum, _ = strconv.Atoi(AdminData.Info) respComm, _ := utils.ConstructPayload(utils.AdminId, "", "COMMAND", "FILESLICENUMCONFIRM", " ", " ", 0, NODEID, AgentStatus.AESKey, false) ProxyChan.ProxyChanToUpperNode <- respComm share.File.ReceiveFileSliceNum <- true case "FILESLICENUMCONFIRM": share.File.TotalConfirm <- true case "FILESIZECONFIRM": share.File.TotalConfirm <- true case "DOWNLOAD
} } //管理admin端下发的数据 func HandleConnFromAdmin(connToAdmin *net.Conn, monitor, listenPort, reConn string, passive bool, N
identifier_body
startnode.go
licenum] = AdminData.Info FileDataMap.Unlock() case "FORWARD": TryForward(AdminData.Info, AdminData.Clientid) case "FORWARDDATA": ForwardConnMap.RLock() if _, ok := ForwardConnMap.Payload[AdminData.Clientid]; ok { PortFowardMap.Lock() if _, ok := PortFowardMap.Payload[AdminData.Clientid]; ok { PortFowardMap.Payload[AdminData.Clientid] <- AdminData.Info } else { PortFowardMap.Payload[AdminData.Clientid] = make(chan string, 1) go HandleForward(PortFowardMap.Payload[AdminData.Clientid], AdminData.Clientid) PortFowardMap.Payload[AdminData.Clientid] <- AdminData.Info } PortFowardMap.Unlock() } ForwardConnMap.RUnlock() case "FORWARDFIN": ForwardConnMap.Lock() if _, ok := ForwardConnMap.Payload[AdminData.Clientid]; ok { ForwardConnMap.Payload[AdminData.Clientid].Close() delete(ForwardConnMap.Payload, AdminData.Clientid) } ForwardConnMap.Unlock() PortFowardMap.Lock() if _, ok := PortFowardMap.Payload[AdminData.Clientid]; ok { if !utils.IsClosed(PortFowardMap.Payload[AdminData.Clientid]) { if !utils.IsClosed(PortFowardMap.Payload[AdminData.Clientid]) { close(PortFowardMap.Payload[AdminData.Clientid]) } delete(PortFowardMap.Payload, AdminData.Clientid) } } PortFowardMap.Unlock() case "REFLECTDATARESP": ReflectConnMap.Lock() ReflectConnMap.Payload[AdminData.Clientid].Write([]byte(AdminData.Info)) ReflectConnMap.Unlock() case "REFLECTTIMEOUT": fallthrough case "REFLECTOFFLINE": ReflectConnMap.Lock() if _, ok := ReflectConnMap.Payload[AdminData.Clientid]; ok { ReflectConnMap.Payload[AdminData.Clientid].Close() delete(ReflectConnMap.Payload, AdminData.Clientid) } ReflectConnMap.Unlock() case "FINOK": SocksDataChanMap.Lock() //性能损失? if _, ok := SocksDataChanMap.Payload[AdminData.Clientid]; ok { if !utils.IsClosed(SocksDataChanMap.Payload[AdminData.Clientid]) { close(SocksDataChanMap.Payload[AdminData.Clientid]) } delete(SocksDataChanMap.Payload, AdminData.Clientid) } SocksDataChanMap.Unlock() case "FIN": CurrentConn.Lock() if _, ok := CurrentConn.Payload[AdminData.Clientid]; ok { CurrentConn.Payload[AdminData.Clientid].Close() delete(CurrentConn.Payload, AdminData.Clientid) } CurrentConn.Unlock() SocksDataChanMap.Lock() if _, ok := SocksDataChanMap.Payload[AdminData.Clientid]; ok { if !utils.IsClosed(SocksDataChanMap.Payload[AdminData.Clientid]) { close(SocksDataChanMap.Payload[AdminData.Clientid]) } delete(SocksDataChanMap.Payload, AdminData.Clientid) } SocksDataChanMap.Unlock() case "HEARTBEAT": hbdatapack, _ := utils.ConstructPayload(utils.AdminId, "", "COMMAND", "KEEPALIVE", " ", " ", 0, NODEID, AgentStatus.AESKey, false) ProxyChan.ProxyChanToUpperNode <- hbdatapack default: continue } case "COMMAND": switch AdminData.Command { case "SHELL": switch AdminData.Info { case "": stdout, stdin = CreatInteractiveShell() go func() { StartShell("", stdin, stdout, NODEID) }() case "exit\n": fallthrough default: go func() { StartShell(AdminData.Info, stdin, stdout, NODEID) }() } case "SOCKS": socksinfo := strings.Split(AdminData.Info, ":::") SocksInfo.SocksUsername = socksinfo[1] SocksInfo.SocksPass = socksinfo[2] StartSocks() case "SOCKSOFF": case "SSH": err := StartSSH(AdminData.Info, NODEID) if err == nil { go ReadCommand() } else { break } case "SSHCOMMAND": go WriteCommand(AdminData.Info) case "SSHTUNNEL": err := SSHTunnelNextNode(AdminData.Info, NODEID) if err != nil { fmt.Println("[*]", err) break } case "CONNECT": var status bool = false command := strings.Split(AdminData.Info, ":::") addr := command[0] choice := command[1] if choice == "1" { //连接的节点是否是在reuseport? status = node.ConnectNextNodeReuse(addr, NODEID, AgentStatus.AESKey) } else { status = node.ConnectNextNode(addr, NODEID, AgentStatus.AESKey) } if !status { message, _ := utils.ConstructPayload(utils.AdminId, "", "COMMAND", "NODECONNECTFAIL", " ", "", 0, NODEID, AgentStatus.AESKey, false) ProxyChan.ProxyChanToUpperNode <- message } case "FILENAME": var err error UploadFile, err := os.Create(AdminData.Info) if err != nil { respComm, _ := utils.ConstructPayload(utils.AdminId, "", "COMMAND", "CREATEFAIL", " ", " ", 0, NODEID, AgentStatus.AESKey, false) ProxyChan.ProxyChanToUpperNode <- respComm } else { respComm, _ := utils.ConstructPayload(utils.AdminId, "", "COMMAND", "NAMECONFIRM", " ", " ", 0, NODEID, AgentStatus.AESKey, false) ProxyChan.ProxyChanToUpperNode <- respComm go share.ReceiveFile("", connToAdmin, FileDataMap, CannotRead, UploadFile, AgentStatus.AESKey, false, NODEID) } case "FILESIZE": filesize, _ := strconv.ParseInt(AdminData.Info, 10, 64) share.File.FileSize = filesize respComm, _ := utils.ConstructPayload(utils.AdminId, "", "COMMAND", "FILESIZECONFIRM", " ", " ", 0, NODEID, AgentStatus.AESKey, false) ProxyChan.ProxyChanToUpperNode <- respComm share.File.ReceiveFileSize <- true case "FILESLICENUM": share.File.TotalSilceNum, _ = strconv.Atoi(AdminData.Info) respComm, _ := utils.ConstructPayload(utils.AdminId, "", "COMMAND", "FILESLICENUMCONFIRM", " ", " ", 0, NODEID, AgentStatus.AESKey, false) ProxyChan.ProxyChanToUpperNode <- respComm share.File.ReceiveFileSliceNum <- true case "FILESLICENUMCONFIRM": share.File.TotalConfirm <- true case "FILESIZECONFIRM": share.File.TotalConfirm <- true case "DOWNLOADFILE": go share.UploadFile("", AdminData.Info, connToAdmin, utils.AdminId, GetName, AgentStatus.AESKey, NODEID, false) case "NAMECONFIRM": GetName <- true case "CREATEFAIL": GetName <- false case "CANNOTREAD": CannotRead <- true share.File.ReceiveFileSliceNum <- false os.Remove(AdminData.Info) //删除空文件 case "FORWARDTEST": go TestForward(AdminData.Info) case "REFLECTTEST": go TestReflect(AdminData.Info) case "REFLECTNUM": ReflectStatus.ReflectNum <- AdminData.Clientid case "STOPREFLECT": ReflectConnMap.Lock() for key, conn := range ReflectConnMap.Payload { conn.Close() delete(ForwardConnMap.Payload, key) } ReflectConnMap.Unlock() for _, listener := range CurrentPortReflectListener { listener.Close() } case "LISTEN": err := TestListen(AdminData.Info) if err != nil { respComm, _ := utils.ConstructPayload(utils.AdminId, "", "COMMAND", "LISTENRESP", " ", "FAILED", 0, NODEID, AgentStatus.AESKey, false) ProxyChan.ProxyChanToUpperNode <- respComm } else { respComm, _ := utils.ConstructPayload(utils.AdminId, "", "COMMAND", "LISTENRESP", " ", "SUCCESS", 0, NODEID, AgentStatus.AESKey, false) ProxyChan.ProxyChanToUpperNode <- respComm go node.StartNodeListen(AdminData.Info, NODEID, AgentStatus.AESKey) } case "YOURINFO": //接收note AgentStatus.NodeNote = AdminData.Info case "KEEPALIVE": default: continue } }
random_line_split
startnode.go
HandleConnFromAdmin(connToAdmin *net.Conn, monitor, listenPort, reConn string, passive bool, NODEID string) { var ( CannotRead = make(chan bool, 1) GetName = make(chan bool, 1) stdin io.Writer stdout io.Reader ) for { AdminData, err := utils.ExtractPayload(*connToAdmin, AgentStatus.AESKey, NODEID, false) if err != nil { AdminO
tor, listenPort, passive) go SendInfo(NODEID) //重连后发送自身信息 go SendNote(NODEID) //重连后发送admin设置的备忘 continue } if AdminData.NodeId == NODEID { switch AdminData.Type { case "DATA": switch AdminData.Command { case "SOCKSDATA": SocksDataChanMap.RLock() if _, ok := SocksDataChanMap.Payload[AdminData.Clientid]; ok { SocksDataChanMap.Payload[AdminData.Clientid] <- AdminData.Info SocksDataChanMap.RUnlock() } else { SocksDataChanMap.RUnlock() SocksDataChanMap.Lock() SocksDataChanMap.Payload[AdminData.Clientid] = make(chan string, 1) go HanleClientSocksConn(SocksDataChanMap.Payload[AdminData.Clientid], SocksInfo.SocksUsername, SocksInfo.SocksPass, AdminData.Clientid, NODEID) SocksDataChanMap.Payload[AdminData.Clientid] <- AdminData.Info SocksDataChanMap.Unlock() } case "FILEDATA": //接收文件内容 slicenum, _ := strconv.Atoi(AdminData.FileSliceNum) FileDataMap.Lock() FileDataMap.Payload[slicenum] = AdminData.Info FileDataMap.Unlock() case "FORWARD": TryForward(AdminData.Info, AdminData.Clientid) case "FORWARDDATA": ForwardConnMap.RLock() if _, ok := ForwardConnMap.Payload[AdminData.Clientid]; ok { PortFowardMap.Lock() if _, ok := PortFowardMap.Payload[AdminData.Clientid]; ok { PortFowardMap.Payload[AdminData.Clientid] <- AdminData.Info } else { PortFowardMap.Payload[AdminData.Clientid] = make(chan string, 1) go HandleForward(PortFowardMap.Payload[AdminData.Clientid], AdminData.Clientid) PortFowardMap.Payload[AdminData.Clientid] <- AdminData.Info } PortFowardMap.Unlock() } ForwardConnMap.RUnlock() case "FORWARDFIN": ForwardConnMap.Lock() if _, ok := ForwardConnMap.Payload[AdminData.Clientid]; ok { ForwardConnMap.Payload[AdminData.Clientid].Close() delete(ForwardConnMap.Payload, AdminData.Clientid) } ForwardConnMap.Unlock() PortFowardMap.Lock() if _, ok := PortFowardMap.Payload[AdminData.Clientid]; ok { if !utils.IsClosed(PortFowardMap.Payload[AdminData.Clientid]) { if !utils.IsClosed(PortFowardMap.Payload[AdminData.Clientid]) { close(PortFowardMap.Payload[AdminData.Clientid]) } delete(PortFowardMap.Payload, AdminData.Clientid) } } PortFowardMap.Unlock() case "REFLECTDATARESP": ReflectConnMap.Lock() ReflectConnMap.Payload[AdminData.Clientid].Write([]byte(AdminData.Info)) ReflectConnMap.Unlock() case "REFLECTTIMEOUT": fallthrough case "REFLECTOFFLINE": ReflectConnMap.Lock() if _, ok := ReflectConnMap.Payload[AdminData.Clientid]; ok { ReflectConnMap.Payload[AdminData.Clientid].Close() delete(ReflectConnMap.Payload, AdminData.Clientid) } ReflectConnMap.Unlock() case "FINOK": SocksDataChanMap.Lock() //性能损失? if _, ok := SocksDataChanMap.Payload[AdminData.Clientid]; ok { if !utils.IsClosed(SocksDataChanMap.Payload[AdminData.Clientid]) { close(SocksDataChanMap.Payload[AdminData.Clientid]) } delete(SocksDataChanMap.Payload, AdminData.Clientid) } SocksDataChanMap.Unlock() case "FIN": CurrentConn.Lock() if _, ok := CurrentConn.Payload[AdminData.Clientid]; ok { CurrentConn.Payload[AdminData.Clientid].Close() delete(CurrentConn.Payload, AdminData.Clientid) } CurrentConn.Unlock() SocksDataChanMap.Lock() if _, ok := SocksDataChanMap.Payload[AdminData.Clientid]; ok { if !utils.IsClosed(SocksDataChanMap.Payload[AdminData.Clientid]) { close(SocksDataChanMap.Payload[AdminData.Clientid]) } delete(SocksDataChanMap.Payload, AdminData.Clientid) } SocksDataChanMap.Unlock() case "HEARTBEAT": hbdatapack, _ := utils.ConstructPayload(utils.AdminId, "", "COMMAND", "KEEPALIVE", " ", " ", 0, NODEID, AgentStatus.AESKey, false) ProxyChan.ProxyChanToUpperNode <- hbdatapack default: continue } case "COMMAND": switch AdminData.Command { case "SHELL": switch AdminData.Info { case "": stdout, stdin = CreatInteractiveShell() go func() { StartShell("", stdin, stdout, NODEID) }() case "exit\n": fallthrough default: go func() { StartShell(AdminData.Info, stdin, stdout, NODEID) }() } case "SOCKS": socksinfo := strings.Split(AdminData.Info, ":::") SocksInfo.SocksUsername = socksinfo[1] SocksInfo.SocksPass = socksinfo[2] StartSocks() case "SOCKSOFF": case "SSH": err := StartSSH(AdminData.Info, NODEID) if err == nil { go ReadCommand() } else { break } case "SSHCOMMAND": go WriteCommand(AdminData.Info) case "SSHTUNNEL": err := SSHTunnelNextNode(AdminData.Info, NODEID) if err != nil { fmt.Println("[*]", err) break } case "CONNECT": var status bool = false command := strings.Split(AdminData.Info, ":::") addr := command[0] choice := command[1] if choice == "1" { //连接的节点是否是在reuseport? status = node.ConnectNextNodeReuse(addr, NODEID, AgentStatus.AESKey) } else { status = node.ConnectNextNode(addr, NODEID, AgentStatus.AESKey) } if !status { message, _ := utils.ConstructPayload(utils.AdminId, "", "COMMAND", "NODECONNECTFAIL", " ", "", 0, NODEID, AgentStatus.AESKey, false) ProxyChan.ProxyChanToUpperNode <- message } case "FILENAME": var err error UploadFile, err := os.Create(AdminData.Info) if err != nil { respComm, _ := utils.ConstructPayload(utils.AdminId, "", "COMMAND", "CREATEFAIL", " ", " ", 0, NODEID, AgentStatus.AESKey, false) ProxyChan.ProxyChanToUpperNode <- respComm } else { respComm, _ := utils.ConstructPayload(utils.AdminId, "", "COMMAND", "NAMECONFIRM", " ", " ", 0, NODEID, AgentStatus.AESKey, false) ProxyChan.ProxyChanToUpperNode <- respComm go share.ReceiveFile("", connToAdmin, FileDataMap, CannotRead, UploadFile, AgentStatus.AESKey, false, NODEID) } case "FILESIZE": filesize, _ := strconv.ParseInt(AdminData.Info, 10, 64) share.File.FileSize = filesize respComm, _ := utils.ConstructPayload(utils.AdminId, "", "COMMAND", "FILESIZECONFIRM", " ", " ", 0, NODEID, AgentStatus.AESKey, false) ProxyChan.ProxyChanToUpperNode <- respComm share.File.ReceiveFileSize <- true case "FILESLICENUM": share.File.TotalSilceNum, _ = strconv.Atoi(AdminData.Info) respComm, _ := utils.ConstructPayload(utils.AdminId, "", "COMMAND", "FILESLICENUMCONFIRM", " ", " ", 0, NODEID, AgentStatus.AESKey, false) ProxyChan.ProxyChanToUpperNode <- respComm share.File.ReceiveFileSliceNum <- true case "FILESLICENUMCONFIRM": share.File.TotalConfirm <- true case "FILESIZECONFIRM": share.File.TotalConfirm <- true case "DOWNLOADFILE": go share.UploadFile("",
ffline(reConn, moni
identifier_name
startnode.go
func HandleConnFromAdmin(connToAdmin *net.Conn, monitor, listenPort, reConn string, passive bool, NODEID string) { var ( CannotRead = make(chan bool, 1) GetName = make(chan bool, 1) stdin io.Writer stdout io.Reader ) for { AdminData, err := utils.ExtractPayload(*connToAdmin, AgentStatus.AESKey, NODEID, false) if err != nil { AdminOffline(reConn, monitor, listenPort, passive) go SendInfo(NODEID) //重连后发送自身信息 go SendNote(NODEID) //重连后发送admin设置的备忘 continue } if AdminData.NodeId == NODEID { switch AdminData.Type { case "DATA": switch AdminData.Command { case "SOCKSDATA": SocksDataChanMap.RLock() if _, ok := SocksDataChanMap.Payload[AdminData.Clientid]; ok { SocksDataChanMap.Payload[AdminData.Clientid] <- AdminData.Info SocksDataChanMap.RUnlock() } else { SocksDataChanMap.RUnlock() SocksDataChanMap.Lock() SocksDataChanMap.Payload[AdminData.Clientid] = make(chan string, 1) go HanleClientSocksConn(SocksDataChanMap.Payload[AdminData.Clientid], SocksInfo.SocksUsername, SocksInfo.SocksPass, AdminData.Clientid, NODEID) SocksDataChanMap.Payload[AdminData.Clientid] <- AdminData.Info SocksDataChanMap.Unlock() } case "FILEDATA": //接收文件内容 slicenum, _ := strconv.Atoi(AdminData.FileSliceNum) FileDataMap.Lock() FileDataMap.Payload[slicenum] = AdminData.Info FileDataMap.Unlock() case "FORWARD": TryForward(AdminData.Info, AdminData.Clientid) case "FORWARDDATA": ForwardConnMap.RLock() if _, ok := ForwardConnMap.Payload[AdminData.Clientid]; ok { PortFowardMap.Lock() if _, ok := PortFowardMap.Payload[AdminData.Clientid]; ok { PortFowardMap.Payload[AdminData.Clientid] <- AdminData.Info } else { PortFowardMap.Payload[AdminData.Clientid] = make(chan string, 1) go HandleForward(PortFowardMap.Payload[AdminData.Clientid], AdminData.Clientid) PortFowardMap.Payload[AdminData.Clientid] <- AdminData.Info } PortFowardMap.Unlock() } ForwardConnMap.RUnlock() case "FORWARDFIN": ForwardConnMap.Lock() if _, ok := ForwardConnMap.Payload[AdminData.Clientid]; ok { ForwardConnMap.Payload[AdminData.Clientid].Close() delete(ForwardConnMap.Payload, AdminData.Clientid) } ForwardConnMap.Unlock() PortFowardMap.Lock() if _, ok := PortFowardMap.Payload[AdminData.Clientid]; ok { if !utils.IsClosed(PortFowardMap.Payload[AdminData.Clientid]) { if !utils.IsClosed(PortFowardMap.Payload[AdminData.Clientid]) { close(PortFowardMap.Payload[AdminData.Clientid]) } delete(PortFowardMap.Payload, AdminData.Clientid) } } PortFowardMap.Unlock() case "REFLECTDATARESP": ReflectConnMap.Lock() ReflectConnMap.Payload[AdminData.Clientid].Write([]byte(AdminData.Info)) ReflectConnMap.Unlock() case "REFLECTTIMEOUT": fallthrough case "REFLECTOFFLINE": ReflectConnMap.Lock() if _, ok := ReflectConnMap.Payload[AdminData.Clientid]; ok { ReflectConnMap.Payload[AdminData.Clientid].Close() delete(ReflectConnMap.Payload, AdminData.Clientid) } ReflectConnMap.Unlock() case "FINOK": SocksDataChanMap.Lock() //性能损失? if _, ok := SocksDataChanMap.Payload[AdminData.Clientid]; ok { if !utils.IsClosed(SocksDataChanMap.Payload[AdminData.Clientid]) { close(SocksDataChanMap.Payload[AdminData.Clientid]) } delete(SocksDataChanMap.Payload, AdminData.Clientid) } SocksDataChanMap.Unlock() case "FIN": CurrentConn.Lock() if _, ok := CurrentConn.Payload[AdminData.Clientid]; ok { CurrentConn.Payload[AdminData.Clientid].Close() delete(CurrentConn.Payload, AdminData.Clientid) } CurrentConn.Unlock() SocksDataChanMap.Lock() if _, ok := SocksDataChanMap.Payload[AdminData.Clientid]; ok { if !utils.IsClosed(SocksDataChanMap.Payload[AdminData.Clientid]) { close(SocksDataChanMap.Payload[AdminData.Clientid]) } delete(SocksDataChanMap.Payload, AdminData.Clientid) } SocksDataChanMap.Unlock() case "HEARTBEAT": hbdatapack, _ := utils.ConstructPayload(utils.AdminId, "", "COMMAND", "KEEPALIVE", " ", " ", 0, NODEID, AgentStatus.AESKey, false) ProxyChan.ProxyChanToUpperNode <- hbdatapack default: continue } case "COMMAND": switch AdminData.Command { case "SHELL": switch AdminData.Info { case "": stdout, stdin = CreatInteractiveShell() go func() { StartShell("", stdin, stdout, NODEID) }() case "exit\n": fallthrough default: go func() { StartShell(AdminData.Info, stdin, stdout, NODEID) }() } case "SOCKS": socksinfo := strings.Split(AdminData.Info, ":::") SocksInfo.SocksUsername = socksinfo[1] SocksInfo.SocksPass = socksinfo[2] StartSocks() case "SOCKSOFF": case "SSH": err := StartSSH(AdminData.Info, NODEID) if err == nil { go ReadCommand() } else { break } case "SSHCOMMAND": go WriteCommand(AdminData.Info) case "SSHTUNNEL": err := SSHTunnelNextNode(AdminData.Info, NODEID) if err != nil { fmt.Println("[*]", err) break } case "CONNECT": var status bool = false command := strings.Split(AdminData.Info, ":::") addr := command[0] choice := command[1] if choice == "1" { //连接的节点是否是在reuseport? status = node.ConnectNextNodeReuse(addr, NODEID, AgentStatus.AESKey) } else { status = node.ConnectNextNode(addr, NODEID, AgentStatus.AESKey) } if !status { message, _ := utils.ConstructPayload(utils.AdminId, "", "COMMAND", "NODECONNECTFAIL", " ", "", 0, NODEID, AgentStatus.AESKey, false) ProxyChan.ProxyChanToUpperNode <- message } case "FILENAME": var err error UploadFile, err := os.Create(AdminData.Info) if err != nil { respComm, _ := utils.ConstructPayload(utils.AdminId, "", "COMMAND", "CREATEFAIL", " ", " ", 0, NODEID, AgentStatus.AESKey, false) ProxyChan.ProxyChanToUpperNode <- respComm } else { respComm, _ := utils.ConstructPayload(utils.AdminId, "", "COMMAND", "NAMECONFIRM", " ", " ", 0, NODEID, AgentStatus.AESKey, false) ProxyChan.ProxyChanToUpperNode <- respComm go share.ReceiveFile("", connToAdmin, FileDat
respComm, _ := utils.ConstructPayload(utils.AdminId, "", "COMMAND", "FILESIZECONFIRM", " ", " ", 0, NODEID, AgentStatus.AESKey, false) ProxyChan.ProxyChanToUpperNode <- respComm share.File.ReceiveFileSize <- true case "FILESLICENUM": share.File.TotalSilceNum, _ = strconv.Atoi(AdminData.Info) respComm, _ := utils.ConstructPayload(utils.AdminId, "", "COMMAND", "FILESLICENUMCONFIRM", " ", " ", 0, NODEID, AgentStatus.AESKey, false) ProxyChan.ProxyChanToUpperNode <- respComm share.File.ReceiveFileSliceNum <- true case "FILESLICENUMCONFIRM": share.File.TotalConfirm <- true case "FILESIZECONFIRM": share.File.TotalConfirm <- true case "DOWNLOADFILE": go share.UploadFile("", Admin
aMap, CannotRead, UploadFile, AgentStatus.AESKey, false, NODEID) } case "FILESIZE": filesize, _ := strconv.ParseInt(AdminData.Info, 10, 64) share.File.FileSize = filesize
conditional_block
console.go
Bright Yellow bg[104] = color{85, 85, 255} // Bright Blue bg[105] = color{255, 85, 255} // Bright Magenta bg[106] = color{85, 255, 255} // Bright Cyan bg[107] = color{255, 255, 255} // Bright White c, ok := bg[i] return c, ok } func New() *Console { c := &Console{} c.bgColor, _ = colorParserBg(40) c.fgColor, _ = colorParserFg(37) c.width = 80 * 9 c.height = 25 * 16 c.scale = 1.5 c.title = "term" c.cursorSetBlink = true c.img = image.NewRGBA(image.Rect(0, 0, c.width, c.height)) c.tmpScreen = ebiten.NewImage(c.width, c.height) c.font.bitmap = bitmap c.font.height = 16 c.font.width = 9 c.clear() return c } func (c *Console) Run() (err error) { // SetRunnableOnUnfocused ebiten.SetRunnableOnUnfocused(true) ebiten.SetWindowSize(c.width, c.height) ebiten.SetWindowTitle(c.title) err = ebiten.RunGame(c) return err } func (c *Console) input() { var r rune for c := 'A'; c <= 'Z'; c++ { if ebiten.IsKeyPressed(ebiten.Key(c) - 'A' + ebiten.KeyA) { if ebiten.IsKeyPressed(ebiten.KeyShift) { r = c } r = (c + 32) // convert to lowercase fmt.Println(string(r)) } } } func (c *Console) Update() error { c.input() c.drawText() return nil } func (c *Console) Layout(outsideWidth, outsideHeight int) (int, int) { return c.width, c.height } func (c *Console) Draw(screen *ebiten.Image) { c.tmpScreen.ReplacePixels(c.img.Pix) screen.DrawImage(c.tmpScreen, nil) } func (c *Console) clear() { c.cursor = 0 for i := 0; i < len(c.videoTextMemory); i++ { c.videoTextMemory[i].charID = ' ' c.videoTextMemory[i].bgColor = c.bgColor c.videoTextMemory[i].fgColor = c.fgColor c.videoTextMemory[i].blink = false } } func (c *Console) drawText() { i := 0 rows := 25 columns := 80 for row := 0; row < rows; row++ { for col := 0; col < columns; col++ { v := c.videoTextMemory[i] if i == c.cursor { c.drawCursor(v.charID, v.fgColor, v.bgColor, col, row) } else { c.drawChar(v.charID, v.fgColor, v.bgColor, col, row) } i++ } } } func (c *Console) moveUp() { columns := 80 copy(c.videoTextMemory[0:], c.videoTextMemory[columns:]) for i := len(c.videoTextMemory) - columns; i < len(c.videoTextMemory); i++ { c.videoTextMemory[i].charID = ' ' c.videoTextMemory[i].bgColor = c.bgColor c.videoTextMemory[i].fgColor = c.fgColor c.videoTextMemory[i].blink = false } } func (c *Console) put(charID int) { c.videoTextMemory[c.cursor].fgColor = c.fgColor c.videoTextMemory[c.cursor].bgColor = c.bgColor c.videoTextMemory[c.cursor].charID = charID c.cursor++ c.cursorLimit() } func (c *Console) cursorLimit() { if c.cursor < 0 { c.cursor = 0 return } columns := 80 rows := 25 for c.cursor >= rows*columns { c.cursor -= columns c.moveUp() } } func (c *Console) Print(msg string) { columns := 80 parseMode := false csi := false s := "" for i := 0; i < len(msg); i++ { v := msg[i] switch { case v == 7: // bell // not implemented case v == 8: // Backspace c.cursor-- c.cursorLimit() case v == 9: // tab \t lin := int(c.cursor / columns) col := int(c.cursor % columns) ncol := int(col/4)*4 + 4 // tab size 4 and remove mod c.cursor = lin*columns + ncol c.cursorLimit() case v == 10: // Line Feed, \n c.cursor += columns c.cursorLimit() case v == 11: // Vertical tab // not implemented case v == 12: // Formfeed // not implemented case v == 13: // Carriage return \r c.cursor = int(c.cursor/columns) * columns c.cursorLimit() case v == 27: parseMode = true case v == '7' && parseMode: // DEC primitive save cursor position c.auxCursorPos = c.cursor // Save cursor position parseMode = false csi = false case v == '8' && parseMode: // DEC primitive restore cursor position c.cursor = c.auxCursorPos // Restore cursor position parseMode = false csi = false case v == '[' && parseMode: // Control Sequence Introducer csi = true s = "" case v == 'c' && csi: // Reset display to initial state c.clear() c.bgColor, _ = colorParserBg(40) c.fgColor, _ = colorParserFg(37) //bold = false parseMode = false csi = false continue case v == 'm' && csi: sv := strings.Split(s, ";") //bold := false for _, j := range sv { if j == "" { continue } else if j == "0" { c.bgColor, _ = colorParserBg(40) c.fgColor, _ = colorParserFg(37) //bold = false continue } else if j == "1" { //bool = true continue } else if j == "39" { // Default foreground color c.fgColor, _ = colorParserFg(37) continue } else if j == "49" { // Default background color c.bgColor, _ = colorParserBg(37) continue } else { i, err := strconv.Atoi(j) if err != nil { fmt.Println(err, "code:", s) continue } fgColor, ok := colorParserFg(i) if ok { c.fgColor = fgColor continue } bgColor, ok := colorParserBg(i) if ok { c.bgColor = bgColor continue } fmt.Println("ANSI code not implemented:", i) } } parseMode = false csi = false case v == 'd' && csi: i := 1 if s != "" { var err error i, err = strconv.Atoi(s) if err != nil { fmt.Println(err) } } cpos := i * columns if cpos < 2000 { c.cursor = cpos } parseMode = false csi = false case v == 's' && csi: c.auxCursorPos = c.cursor // Save cursor position parseMode = false csi = false case v == 'u' && csi: c.cursor = c.auxCursorPos // Restore cursor position parseMode = false csi = false case v == 'A' && csi: // Cursor up i := 1 if s != "" { var err error i, err = strconv.Atoi(s) if err != nil { fmt.Println(err) } } c.cursor -= i * columns c.cursorLimit() parseMode = false csi = false case v == 'B' && csi: // Cursor down i := 1 if s != "" { var err error i, err = strconv.Atoi(s) if err != nil
{ fmt.Println(err) }
conditional_block
console.go
255, 255} // Bright White c, ok := fg[i] return c, ok } func colorParserBg(i int) (color, bool) { bg := make(map[int]color, 16) bg[40] = color{0, 0, 0} // Black bg[41] = color{170, 0, 0} // Red bg[42] = color{0, 170, 0} // Green bg[43] = color{170, 85, 0} // Yellow bg[44] = color{0, 0, 170} // Blue bg[45] = color{170, 0, 170} // Magenta bg[46] = color{0, 170, 170} // Cyan bg[47] = color{170, 170, 170} // White bg[100] = color{85, 85, 85} // Bright Black (Gray) bg[101] = color{255, 85, 85} // Bright Red bg[102] = color{85, 255, 85} // Bright Green bg[103] = color{255, 255, 85} // Bright Yellow bg[104] = color{85, 85, 255} // Bright Blue bg[105] = color{255, 85, 255} // Bright Magenta bg[106] = color{85, 255, 255} // Bright Cyan bg[107] = color{255, 255, 255} // Bright White c, ok := bg[i] return c, ok }
c.fgColor, _ = colorParserFg(37) c.width = 80 * 9 c.height = 25 * 16 c.scale = 1.5 c.title = "term" c.cursorSetBlink = true c.img = image.NewRGBA(image.Rect(0, 0, c.width, c.height)) c.tmpScreen = ebiten.NewImage(c.width, c.height) c.font.bitmap = bitmap c.font.height = 16 c.font.width = 9 c.clear() return c } func (c *Console) Run() (err error) { // SetRunnableOnUnfocused ebiten.SetRunnableOnUnfocused(true) ebiten.SetWindowSize(c.width, c.height) ebiten.SetWindowTitle(c.title) err = ebiten.RunGame(c) return err } func (c *Console) input() { var r rune for c := 'A'; c <= 'Z'; c++ { if ebiten.IsKeyPressed(ebiten.Key(c) - 'A' + ebiten.KeyA) { if ebiten.IsKeyPressed(ebiten.KeyShift) { r = c } r = (c + 32) // convert to lowercase fmt.Println(string(r)) } } } func (c *Console) Update() error { c.input() c.drawText() return nil } func (c *Console) Layout(outsideWidth, outsideHeight int) (int, int) { return c.width, c.height } func (c *Console) Draw(screen *ebiten.Image) { c.tmpScreen.ReplacePixels(c.img.Pix) screen.DrawImage(c.tmpScreen, nil) } func (c *Console) clear() { c.cursor = 0 for i := 0; i < len(c.videoTextMemory); i++ { c.videoTextMemory[i].charID = ' ' c.videoTextMemory[i].bgColor = c.bgColor c.videoTextMemory[i].fgColor = c.fgColor c.videoTextMemory[i].blink = false } } func (c *Console) drawText() { i := 0 rows := 25 columns := 80 for row := 0; row < rows; row++ { for col := 0; col < columns; col++ { v := c.videoTextMemory[i] if i == c.cursor { c.drawCursor(v.charID, v.fgColor, v.bgColor, col, row) } else { c.drawChar(v.charID, v.fgColor, v.bgColor, col, row) } i++ } } } func (c *Console) moveUp() { columns := 80 copy(c.videoTextMemory[0:], c.videoTextMemory[columns:]) for i := len(c.videoTextMemory) - columns; i < len(c.videoTextMemory); i++ { c.videoTextMemory[i].charID = ' ' c.videoTextMemory[i].bgColor = c.bgColor c.videoTextMemory[i].fgColor = c.fgColor c.videoTextMemory[i].blink = false } } func (c *Console) put(charID int) { c.videoTextMemory[c.cursor].fgColor = c.fgColor c.videoTextMemory[c.cursor].bgColor = c.bgColor c.videoTextMemory[c.cursor].charID = charID c.cursor++ c.cursorLimit() } func (c *Console) cursorLimit() { if c.cursor < 0 { c.cursor = 0 return } columns := 80 rows := 25 for c.cursor >= rows*columns { c.cursor -= columns c.moveUp() } } func (c *Console) Print(msg string) { columns := 80 parseMode := false csi := false s := "" for i := 0; i < len(msg); i++ { v := msg[i] switch { case v == 7: // bell // not implemented case v == 8: // Backspace c.cursor-- c.cursorLimit() case v == 9: // tab \t lin := int(c.cursor / columns) col := int(c.cursor % columns) ncol := int(col/4)*4 + 4 // tab size 4 and remove mod c.cursor = lin*columns + ncol c.cursorLimit() case v == 10: // Line Feed, \n c.cursor += columns c.cursorLimit() case v == 11: // Vertical tab // not implemented case v == 12: // Formfeed // not implemented case v == 13: // Carriage return \r c.cursor = int(c.cursor/columns) * columns c.cursorLimit() case v == 27: parseMode = true case v == '7' && parseMode: // DEC primitive save cursor position c.auxCursorPos = c.cursor // Save cursor position parseMode = false csi = false case v == '8' && parseMode: // DEC primitive restore cursor position c.cursor = c.auxCursorPos // Restore cursor position parseMode = false csi = false case v == '[' && parseMode: // Control Sequence Introducer csi = true s = "" case v == 'c' && csi: // Reset display to initial state c.clear() c.bgColor, _ = colorParserBg(40) c.fgColor, _ = colorParserFg(37) //bold = false parseMode = false csi = false continue case v == 'm' && csi: sv := strings.Split(s, ";") //bold := false for _, j := range sv { if j == "" { continue } else if j == "0" { c.bgColor, _ = colorParserBg(40) c.fgColor, _ = colorParserFg(37) //bold = false continue } else if j == "1" { //bool = true continue } else if j == "39" { // Default foreground color c.fgColor, _ = colorParserFg(37) continue } else if j == "49" { // Default background color c.bgColor, _ = colorParserBg(37) continue } else { i, err := strconv.Atoi(j) if err != nil { fmt.Println(err, "code:", s) continue } fgColor, ok := colorParserFg(i) if ok { c.fgColor = fgColor continue } bgColor, ok := colorParserBg(i) if ok { c.bgColor = bgColor continue
func New() *Console { c := &Console{} c.bgColor, _ = colorParserBg(40)
random_line_split
console.go
255, 255} // Bright White c, ok := fg[i] return c, ok } func colorParserBg(i int) (color, bool) { bg := make(map[int]color, 16) bg[40] = color{0, 0, 0} // Black bg[41] = color{170, 0, 0} // Red bg[42] = color{0, 170, 0} // Green bg[43] = color{170, 85, 0} // Yellow bg[44] = color{0, 0, 170} // Blue bg[45] = color{170, 0, 170} // Magenta bg[46] = color{0, 170, 170} // Cyan bg[47] = color{170, 170, 170} // White bg[100] = color{85, 85, 85} // Bright Black (Gray) bg[101] = color{255, 85, 85} // Bright Red bg[102] = color{85, 255, 85} // Bright Green bg[103] = color{255, 255, 85} // Bright Yellow bg[104] = color{85, 85, 255} // Bright Blue bg[105] = color{255, 85, 255} // Bright Magenta bg[106] = color{85, 255, 255} // Bright Cyan bg[107] = color{255, 255, 255} // Bright White c, ok := bg[i] return c, ok } func New() *Console { c := &Console{} c.bgColor, _ = colorParserBg(40) c.fgColor, _ = colorParserFg(37) c.width = 80 * 9 c.height = 25 * 16 c.scale = 1.5 c.title = "term" c.cursorSetBlink = true c.img = image.NewRGBA(image.Rect(0, 0, c.width, c.height)) c.tmpScreen = ebiten.NewImage(c.width, c.height) c.font.bitmap = bitmap c.font.height = 16 c.font.width = 9 c.clear() return c } func (c *Console) Run() (err error) { // SetRunnableOnUnfocused ebiten.SetRunnableOnUnfocused(true) ebiten.SetWindowSize(c.width, c.height) ebiten.SetWindowTitle(c.title) err = ebiten.RunGame(c) return err } func (c *Console) input() { var r rune for c := 'A'; c <= 'Z'; c++ { if ebiten.IsKeyPressed(ebiten.Key(c) - 'A' + ebiten.KeyA) { if ebiten.IsKeyPressed(ebiten.KeyShift) { r = c } r = (c + 32) // convert to lowercase fmt.Println(string(r)) } } } func (c *Console) Update() error { c.input() c.drawText() return nil } func (c *Console) Layout(outsideWidth, outsideHeight int) (int, int) { return c.width, c.height } func (c *Console) Draw(screen *ebiten.Image) { c.tmpScreen.ReplacePixels(c.img.Pix) screen.DrawImage(c.tmpScreen, nil) } func (c *Console) clear() { c.cursor = 0 for i := 0; i < len(c.videoTextMemory); i++ { c.videoTextMemory[i].charID = ' ' c.videoTextMemory[i].bgColor = c.bgColor c.videoTextMemory[i].fgColor = c.fgColor c.videoTextMemory[i].blink = false } } func (c *Console) drawText() { i := 0 rows := 25 columns := 80 for row := 0; row < rows; row++ { for col := 0; col < columns; col++ { v := c.videoTextMemory[i] if i == c.cursor { c.drawCursor(v.charID, v.fgColor, v.bgColor, col, row) } else { c.drawChar(v.charID, v.fgColor, v.bgColor, col, row) } i++ } } } func (c *Console) moveUp() { columns := 80 copy(c.videoTextMemory[0:], c.videoTextMemory[columns:]) for i := len(c.videoTextMemory) - columns; i < len(c.videoTextMemory); i++ { c.videoTextMemory[i].charID = ' ' c.videoTextMemory[i].bgColor = c.bgColor c.videoTextMemory[i].fgColor = c.fgColor c.videoTextMemory[i].blink = false } } func (c *Console)
(charID int) { c.videoTextMemory[c.cursor].fgColor = c.fgColor c.videoTextMemory[c.cursor].bgColor = c.bgColor c.videoTextMemory[c.cursor].charID = charID c.cursor++ c.cursorLimit() } func (c *Console) cursorLimit() { if c.cursor < 0 { c.cursor = 0 return } columns := 80 rows := 25 for c.cursor >= rows*columns { c.cursor -= columns c.moveUp() } } func (c *Console) Print(msg string) { columns := 80 parseMode := false csi := false s := "" for i := 0; i < len(msg); i++ { v := msg[i] switch { case v == 7: // bell // not implemented case v == 8: // Backspace c.cursor-- c.cursorLimit() case v == 9: // tab \t lin := int(c.cursor / columns) col := int(c.cursor % columns) ncol := int(col/4)*4 + 4 // tab size 4 and remove mod c.cursor = lin*columns + ncol c.cursorLimit() case v == 10: // Line Feed, \n c.cursor += columns c.cursorLimit() case v == 11: // Vertical tab // not implemented case v == 12: // Formfeed // not implemented case v == 13: // Carriage return \r c.cursor = int(c.cursor/columns) * columns c.cursorLimit() case v == 27: parseMode = true case v == '7' && parseMode: // DEC primitive save cursor position c.auxCursorPos = c.cursor // Save cursor position parseMode = false csi = false case v == '8' && parseMode: // DEC primitive restore cursor position c.cursor = c.auxCursorPos // Restore cursor position parseMode = false csi = false case v == '[' && parseMode: // Control Sequence Introducer csi = true s = "" case v == 'c' && csi: // Reset display to initial state c.clear() c.bgColor, _ = colorParserBg(40) c.fgColor, _ = colorParserFg(37) //bold = false parseMode = false csi = false continue case v == 'm' && csi: sv := strings.Split(s, ";") //bold := false for _, j := range sv { if j == "" { continue } else if j == "0" { c.bgColor, _ = colorParserBg(40) c.fgColor, _ = colorParserFg(37) //bold = false continue } else if j == "1" { //bool = true continue } else if j == "39" { // Default foreground color c.fgColor, _ = colorParserFg(37) continue } else if j == "49" { // Default background color c.bgColor, _ = colorParserBg(37) continue } else { i, err := strconv.Atoi(j) if err != nil { fmt.Println(err, "code:", s) continue } fgColor, ok := colorParserFg(i) if ok { c.fgColor = fgColor continue } bgColor, ok := colorParserBg(i) if ok { c.bgColor = bgColor continue
put
identifier_name
console.go
++ c.cursorLimit() } func (c *Console) cursorLimit() { if c.cursor < 0 { c.cursor = 0 return } columns := 80 rows := 25 for c.cursor >= rows*columns { c.cursor -= columns c.moveUp() } } func (c *Console) Print(msg string) { columns := 80 parseMode := false csi := false s := "" for i := 0; i < len(msg); i++ { v := msg[i] switch { case v == 7: // bell // not implemented case v == 8: // Backspace c.cursor-- c.cursorLimit() case v == 9: // tab \t lin := int(c.cursor / columns) col := int(c.cursor % columns) ncol := int(col/4)*4 + 4 // tab size 4 and remove mod c.cursor = lin*columns + ncol c.cursorLimit() case v == 10: // Line Feed, \n c.cursor += columns c.cursorLimit() case v == 11: // Vertical tab // not implemented case v == 12: // Formfeed // not implemented case v == 13: // Carriage return \r c.cursor = int(c.cursor/columns) * columns c.cursorLimit() case v == 27: parseMode = true case v == '7' && parseMode: // DEC primitive save cursor position c.auxCursorPos = c.cursor // Save cursor position parseMode = false csi = false case v == '8' && parseMode: // DEC primitive restore cursor position c.cursor = c.auxCursorPos // Restore cursor position parseMode = false csi = false case v == '[' && parseMode: // Control Sequence Introducer csi = true s = "" case v == 'c' && csi: // Reset display to initial state c.clear() c.bgColor, _ = colorParserBg(40) c.fgColor, _ = colorParserFg(37) //bold = false parseMode = false csi = false continue case v == 'm' && csi: sv := strings.Split(s, ";") //bold := false for _, j := range sv { if j == "" { continue } else if j == "0" { c.bgColor, _ = colorParserBg(40) c.fgColor, _ = colorParserFg(37) //bold = false continue } else if j == "1" { //bool = true continue } else if j == "39" { // Default foreground color c.fgColor, _ = colorParserFg(37) continue } else if j == "49" { // Default background color c.bgColor, _ = colorParserBg(37) continue } else { i, err := strconv.Atoi(j) if err != nil { fmt.Println(err, "code:", s) continue } fgColor, ok := colorParserFg(i) if ok { c.fgColor = fgColor continue } bgColor, ok := colorParserBg(i) if ok { c.bgColor = bgColor continue } fmt.Println("ANSI code not implemented:", i) } } parseMode = false csi = false case v == 'd' && csi: i := 1 if s != "" { var err error i, err = strconv.Atoi(s) if err != nil { fmt.Println(err) } } cpos := i * columns if cpos < 2000 { c.cursor = cpos } parseMode = false csi = false case v == 's' && csi: c.auxCursorPos = c.cursor // Save cursor position parseMode = false csi = false case v == 'u' && csi: c.cursor = c.auxCursorPos // Restore cursor position parseMode = false csi = false case v == 'A' && csi: // Cursor up i := 1 if s != "" { var err error i, err = strconv.Atoi(s) if err != nil { fmt.Println(err) } } c.cursor -= i * columns c.cursorLimit() parseMode = false csi = false case v == 'B' && csi: // Cursor down i := 1 if s != "" { var err error i, err = strconv.Atoi(s) if err != nil { fmt.Println(err) } } c.cursor += i * columns c.cursorLimit() parseMode = false csi = false case v == 'C' && csi: // Cursor forward i := 1 if s != "" { var err error i, err = strconv.Atoi(s) if err != nil { fmt.Println(err) } } c.cursor += i c.cursorLimit() parseMode = false csi = false case v == 'D' && csi: // Cursor back i := 1 if s != "" { var err error i, err = strconv.Atoi(s) if err != nil { fmt.Println(err) } } c.cursor -= i c.cursorLimit() parseMode = false csi = false case v == 'G' && csi: i := 1 if s != "" { var err error i, err = strconv.Atoi(s) if err != nil { fmt.Println(err) } } lin := int(c.cursor / columns) cpos := lin*columns + i if cpos < 2000 { c.cursor = cpos } parseMode = false csi = false case v == 'f' && csi: // the same as H fallthrough case v == 'H' && csi: // set horizontal and vertical position if s == "" { c.cursor = 0 } else { sv := strings.Split(s, ";") if len(sv) == 2 { lin, _ := strconv.Atoi(sv[0]) col, _ := strconv.Atoi(sv[1]) cpos := lin*columns + col if cpos <= 2000 { // 25*80 c.cursor = cpos } } } parseMode = false csi = false case v == 'X' && csi: // Erase n characters from the current position cpos := c.cursor i := 1 if s != "" { i, _ = strconv.Atoi(s) } for x := 1; x <= i; x++ { if cpos+x < 2000 { c.videoTextMemory[cpos+x].charID = ' ' } } c.cursor = cpos parseMode = false csi = false case v == 'J' && csi: if len(s) > 0 { if s[0] == '2' { c.clear() } } parseMode = false csi = false case v >= 'a' && v <= 'z' && v <= 'A' && v <= 'Z' && parseMode: parseMode = false csi = false case csi || parseMode: s += string(v) default: c.put(int(msg[i])) } } } func (c *Console) set(x, y int, color color) { p := 4*y*c.width + 4*x c.img.Pix[p] = color.R c.img.Pix[p+1] = color.G c.img.Pix[p+2] = color.B c.img.Pix[p+3] = 0xff } func (c *Console) drawCursor(index int, fgColor, bgColor color, x, y int)
{ if c.cursorSetBlink { if c.cursorBlinkTimer < 15 { fgColor, bgColor = bgColor, fgColor } c.drawChar(index, fgColor, bgColor, x, y) c.cursorBlinkTimer++ if c.cursorBlinkTimer > 30 { c.cursorBlinkTimer = 0 } return } c.drawChar(index, bgColor, fgColor, x, y) }
identifier_body
ledger_cleanup_service.rs
Handle}; use std::time::Duration; // - To try and keep the RocksDB size under 400GB: // Seeing about 1600b/shred, using 2000b/shred for margin, so 200m shreds can be stored in 400gb. // at 5k shreds/slot at 50k tps, this is 500k slots (~5 hours). // At idle, 60 shreds/slot this is about 4m slots (18 days) // This is chosen to allow enough time for // - A validator to download a snapshot from a peer and boot from it // - To make sure that if a validator needs to reboot from its own snapshot, it has enough slots locally // to catch back up to where it was when it stopped pub const DEFAULT_MAX_LEDGER_SHREDS: u64 = 200_000_000; // Allow down to 50m, or 3.5 days at idle, 1hr at 50k load, around ~100GB pub const DEFAULT_MIN_MAX_LEDGER_SHREDS: u64 = 50_000_000; // Check for removing slots at this interval so we don't purge too often // and starve other blockstore users. pub const DEFAULT_PURGE_SLOT_INTERVAL: u64 = 512; // Remove a limited number of slots at a time, so the operation // does not take too long and block other blockstore users. pub const DEFAULT_PURGE_BATCH_SIZE: u64 = 256; pub struct LedgerCleanupService { t_cleanup: JoinHandle<()>, } impl LedgerCleanupService { pub fn new( new_root_receiver: Receiver<Slot>, blockstore: Arc<Blockstore>, max_ledger_shreds: u64, exit: &Arc<AtomicBool>, ) -> Self { info!( "LedgerCleanupService active. Max Ledger Slots {}", max_ledger_shreds ); let exit = exit.clone(); let mut last_purge_slot = 0; let t_cleanup = Builder::new() .name("solana-ledger-cleanup".to_string()) .spawn(move || loop { if exit.load(Ordering::Relaxed) { break; } if let Err(e) = Self::cleanup_ledger( &new_root_receiver, &blockstore, max_ledger_shreds, &mut last_purge_slot, DEFAULT_PURGE_SLOT_INTERVAL, ) { match e { RecvTimeoutError::Disconnected => break, RecvTimeoutError::Timeout => (), } } }) .unwrap(); Self { t_cleanup } } fn find_slots_to_clean( blockstore: &Arc<Blockstore>, root: Slot, max_ledger_shreds: u64, ) -> (u64, Slot, Slot)
max_ledger_shreds, shreds.len(), total_shreds, iterate_time ); if (total_shreds as u64) < max_ledger_shreds { return (0, 0, 0); } let mut cur_shreds = 0; let mut lowest_slot_to_clean = shreds[0].0; for (slot, num_shreds) in shreds.iter().rev() { cur_shreds += *num_shreds as u64; if cur_shreds > max_ledger_shreds { lowest_slot_to_clean = *slot; break; } } (cur_shreds, lowest_slot_to_clean, first_slot) } pub fn cleanup_ledger( new_root_receiver: &Receiver<Slot>, blockstore: &Arc<Blockstore>, max_ledger_shreds: u64, last_purge_slot: &mut u64, purge_interval: u64, ) -> Result<(), RecvTimeoutError> { let mut root = new_root_receiver.recv_timeout(Duration::from_secs(1))?; // Get the newest root while let Ok(new_root) = new_root_receiver.try_recv() { root = new_root; } if root - *last_purge_slot > purge_interval { let disk_utilization_pre = blockstore.storage_size(); info!( "purge: new root: {} last_purge: {} purge_interval: {} disk: {:?}", root, last_purge_slot, purge_interval, disk_utilization_pre ); *last_purge_slot = root; let (num_shreds_to_clean, lowest_slot_to_clean, mut first_slot) = Self::find_slots_to_clean(blockstore, root, max_ledger_shreds); if num_shreds_to_clean > 0 { debug!( "cleaning up to: {} shreds: {} first: {}", lowest_slot_to_clean, num_shreds_to_clean, first_slot ); loop { let current_lowest = std::cmp::min(lowest_slot_to_clean, first_slot + DEFAULT_PURGE_BATCH_SIZE); let mut slot_update_time = Measure::start("slot_update"); *blockstore.lowest_cleanup_slot.write().unwrap() = current_lowest; slot_update_time.stop(); let mut clean_time = Measure::start("ledger_clean"); blockstore.purge_slots(first_slot, Some(current_lowest)); clean_time.stop(); debug!( "ledger purge {} -> {}: {} {}", first_slot, current_lowest, slot_update_time, clean_time ); first_slot += DEFAULT_PURGE_BATCH_SIZE; if current_lowest == lowest_slot_to_clean { break; } thread::sleep(Duration::from_millis(500)); } } let disk_utilization_post = blockstore.storage_size(); Self::report_disk_metrics(disk_utilization_pre, disk_utilization_post); } Ok(()) } fn report_disk_metrics(pre: BlockstoreResult<u64>, post: BlockstoreResult<u64>) { if let (Ok(pre), Ok(post)) = (pre, post) { datapoint_debug!( "ledger_disk_utilization", ("disk_utilization_pre", pre as i64, i64), ("disk_utilization_post", post as i64, i64), ("disk_utilization_delta", (pre as i64 - post as i64), i64) ); } } pub fn join(self) -> thread::Result<()> { self.t_cleanup.join() } } #[cfg(test)] mod tests { use super::*; use solana_ledger::blockstore::make_many_slot_entries; use solana_ledger::get_tmp_ledger_path; use std::sync::mpsc::channel; #[test] fn test_cleanup() { solana_logger::setup(); let blockstore_path = get_tmp_ledger_path!(); let blockstore = Blockstore::open(&blockstore_path).unwrap(); let (shreds, _) = make_many_slot_entries(0, 50, 5); blockstore.insert_shreds(shreds, None, false).unwrap(); let blockstore = Arc::new(blockstore); let (sender, receiver) = channel(); //send a signal to kill all but 5 shreds, which will be in the newest slots let mut last_purge_slot = 0; sender.send(50).unwrap(); LedgerCleanupService::cleanup_ledger(&receiver, &blockstore, 5, &mut last_purge_slot, 10) .unwrap(); //check that 0-40 don't exist blockstore .slot_meta_iterator(0) .unwrap() .for_each(|(slot, _)| assert!(slot > 40)); drop(blockstore); Blockstore::destroy(&blockstore_path).expect("Expected successful database destruction"); } #[test] fn test_cleanup_speed() { solana_logger::setup(); let blockstore_path = get_tmp_ledger_path!(); let mut blockstore = Blockstore::open(&blockstore_path).unwrap(); blockstore.set_no_compaction(true); let blockstore = Arc::new(blockstore); let (sender, receiver) = channel(); let mut first_insert = Measure::start("first_insert"); let initial_slots = 50; let initial_entries = 5; let (shreds, _) = make_many_slot_entries(0, initial_slots, initial_entries); blockstore.insert_shreds(shreds, None, false).unwrap(); first
{ let mut shreds = Vec::new(); let mut iterate_time = Measure::start("iterate_time"); let mut total_shreds = 0; let mut first_slot = 0; for (i, (slot, meta)) in blockstore.slot_meta_iterator(0).unwrap().enumerate() { if i == 0 { first_slot = slot; debug!("purge: searching from slot: {}", slot); } // Not exact since non-full slots will have holes total_shreds += meta.received; shreds.push((slot, meta.received)); if slot > root { break; } } iterate_time.stop(); info!( "checking for ledger purge: max_shreds: {} slots: {} total_shreds: {} {}",
identifier_body
ledger_cleanup_service.rs
Handle}; use std::time::Duration; // - To try and keep the RocksDB size under 400GB: // Seeing about 1600b/shred, using 2000b/shred for margin, so 200m shreds can be stored in 400gb. // at 5k shreds/slot at 50k tps, this is 500k slots (~5 hours). // At idle, 60 shreds/slot this is about 4m slots (18 days) // This is chosen to allow enough time for // - A validator to download a snapshot from a peer and boot from it // - To make sure that if a validator needs to reboot from its own snapshot, it has enough slots locally // to catch back up to where it was when it stopped pub const DEFAULT_MAX_LEDGER_SHREDS: u64 = 200_000_000; // Allow down to 50m, or 3.5 days at idle, 1hr at 50k load, around ~100GB pub const DEFAULT_MIN_MAX_LEDGER_SHREDS: u64 = 50_000_000; // Check for removing slots at this interval so we don't purge too often // and starve other blockstore users. pub const DEFAULT_PURGE_SLOT_INTERVAL: u64 = 512; // Remove a limited number of slots at a time, so the operation // does not take too long and block other blockstore users. pub const DEFAULT_PURGE_BATCH_SIZE: u64 = 256; pub struct LedgerCleanupService { t_cleanup: JoinHandle<()>, } impl LedgerCleanupService { pub fn new( new_root_receiver: Receiver<Slot>, blockstore: Arc<Blockstore>, max_ledger_shreds: u64, exit: &Arc<AtomicBool>, ) -> Self { info!( "LedgerCleanupService active. Max Ledger Slots {}", max_ledger_shreds ); let exit = exit.clone(); let mut last_purge_slot = 0; let t_cleanup = Builder::new() .name("solana-ledger-cleanup".to_string()) .spawn(move || loop { if exit.load(Ordering::Relaxed) { break; } if let Err(e) = Self::cleanup_ledger( &new_root_receiver, &blockstore, max_ledger_shreds, &mut last_purge_slot, DEFAULT_PURGE_SLOT_INTERVAL, ) { match e { RecvTimeoutError::Disconnected => break, RecvTimeoutError::Timeout => (), } } }) .unwrap(); Self { t_cleanup } } fn find_slots_to_clean( blockstore: &Arc<Blockstore>, root: Slot, max_ledger_shreds: u64, ) -> (u64, Slot, Slot) { let mut shreds = Vec::new(); let mut iterate_time = Measure::start("iterate_time"); let mut total_shreds = 0; let mut first_slot = 0; for (i, (slot, meta)) in blockstore.slot_meta_iterator(0).unwrap().enumerate() { if i == 0 { first_slot = slot; debug!("purge: searching from slot: {}", slot); } // Not exact since non-full slots will have holes total_shreds += meta.received; shreds.push((slot, meta.received)); if slot > root { break; } } iterate_time.stop(); info!( "checking for ledger purge: max_shreds: {} slots: {} total_shreds: {} {}", max_ledger_shreds, shreds.len(), total_shreds, iterate_time ); if (total_shreds as u64) < max_ledger_shreds { return (0, 0, 0); } let mut cur_shreds = 0; let mut lowest_slot_to_clean = shreds[0].0; for (slot, num_shreds) in shreds.iter().rev() { cur_shreds += *num_shreds as u64; if cur_shreds > max_ledger_shreds { lowest_slot_to_clean = *slot; break; } } (cur_shreds, lowest_slot_to_clean, first_slot) } pub fn cleanup_ledger( new_root_receiver: &Receiver<Slot>, blockstore: &Arc<Blockstore>, max_ledger_shreds: u64, last_purge_slot: &mut u64, purge_interval: u64, ) -> Result<(), RecvTimeoutError> { let mut root = new_root_receiver.recv_timeout(Duration::from_secs(1))?; // Get the newest root while let Ok(new_root) = new_root_receiver.try_recv() { root = new_root; } if root - *last_purge_slot > purge_interval { let disk_utilization_pre = blockstore.storage_size(); info!( "purge: new root: {} last_purge: {} purge_interval: {} disk: {:?}", root, last_purge_slot, purge_interval, disk_utilization_pre ); *last_purge_slot = root; let (num_shreds_to_clean, lowest_slot_to_clean, mut first_slot) = Self::find_slots_to_clean(blockstore, root, max_ledger_shreds); if num_shreds_to_clean > 0 { debug!( "cleaning up to: {} shreds: {} first: {}", lowest_slot_to_clean, num_shreds_to_clean, first_slot ); loop { let current_lowest = std::cmp::min(lowest_slot_to_clean, first_slot + DEFAULT_PURGE_BATCH_SIZE); let mut slot_update_time = Measure::start("slot_update"); *blockstore.lowest_cleanup_slot.write().unwrap() = current_lowest; slot_update_time.stop(); let mut clean_time = Measure::start("ledger_clean"); blockstore.purge_slots(first_slot, Some(current_lowest)); clean_time.stop(); debug!( "ledger purge {} -> {}: {} {}", first_slot, current_lowest, slot_update_time, clean_time ); first_slot += DEFAULT_PURGE_BATCH_SIZE; if current_lowest == lowest_slot_to_clean { break; } thread::sleep(Duration::from_millis(500)); } } let disk_utilization_post = blockstore.storage_size(); Self::report_disk_metrics(disk_utilization_pre, disk_utilization_post); } Ok(()) } fn report_disk_metrics(pre: BlockstoreResult<u64>, post: BlockstoreResult<u64>) { if let (Ok(pre), Ok(post)) = (pre, post) { datapoint_debug!( "ledger_disk_utilization", ("disk_utilization_pre", pre as i64, i64), ("disk_utilization_post", post as i64, i64), ("disk_utilization_delta", (pre as i64 - post as i64), i64) ); } } pub fn join(self) -> thread::Result<()> { self.t_cleanup.join() } } #[cfg(test)] mod tests { use super::*; use solana_ledger::blockstore::make_many_slot_entries; use solana_ledger::get_tmp_ledger_path; use std::sync::mpsc::channel; #[test] fn
() { solana_logger::setup(); let blockstore_path = get_tmp_ledger_path!(); let blockstore = Blockstore::open(&blockstore_path).unwrap(); let (shreds, _) = make_many_slot_entries(0, 50, 5); blockstore.insert_shreds(shreds, None, false).unwrap(); let blockstore = Arc::new(blockstore); let (sender, receiver) = channel(); //send a signal to kill all but 5 shreds, which will be in the newest slots let mut last_purge_slot = 0; sender.send(50).unwrap(); LedgerCleanupService::cleanup_ledger(&receiver, &blockstore, 5, &mut last_purge_slot, 10) .unwrap(); //check that 0-40 don't exist blockstore .slot_meta_iterator(0) .unwrap() .for_each(|(slot, _)| assert!(slot > 40)); drop(blockstore); Blockstore::destroy(&blockstore_path).expect("Expected successful database destruction"); } #[test] fn test_cleanup_speed() { solana_logger::setup(); let blockstore_path = get_tmp_ledger_path!(); let mut blockstore = Blockstore::open(&blockstore_path).unwrap(); blockstore.set_no_compaction(true); let blockstore = Arc::new(blockstore); let (sender, receiver) = channel(); let mut first_insert = Measure::start("first_insert"); let initial_slots = 50; let initial_entries = 5; let (shreds, _) = make_many_slot_entries(0, initial_slots, initial_entries); blockstore.insert_shreds(shreds, None, false).unwrap();
test_cleanup
identifier_name
ledger_cleanup_service.rs
Handle}; use std::time::Duration; // - To try and keep the RocksDB size under 400GB: // Seeing about 1600b/shred, using 2000b/shred for margin, so 200m shreds can be stored in 400gb. // at 5k shreds/slot at 50k tps, this is 500k slots (~5 hours). // At idle, 60 shreds/slot this is about 4m slots (18 days) // This is chosen to allow enough time for // - A validator to download a snapshot from a peer and boot from it // - To make sure that if a validator needs to reboot from its own snapshot, it has enough slots locally // to catch back up to where it was when it stopped pub const DEFAULT_MAX_LEDGER_SHREDS: u64 = 200_000_000; // Allow down to 50m, or 3.5 days at idle, 1hr at 50k load, around ~100GB pub const DEFAULT_MIN_MAX_LEDGER_SHREDS: u64 = 50_000_000; // Check for removing slots at this interval so we don't purge too often // and starve other blockstore users. pub const DEFAULT_PURGE_SLOT_INTERVAL: u64 = 512; // Remove a limited number of slots at a time, so the operation // does not take too long and block other blockstore users. pub const DEFAULT_PURGE_BATCH_SIZE: u64 = 256; pub struct LedgerCleanupService { t_cleanup: JoinHandle<()>, } impl LedgerCleanupService { pub fn new( new_root_receiver: Receiver<Slot>, blockstore: Arc<Blockstore>, max_ledger_shreds: u64, exit: &Arc<AtomicBool>, ) -> Self { info!( "LedgerCleanupService active. Max Ledger Slots {}", max_ledger_shreds ); let exit = exit.clone(); let mut last_purge_slot = 0; let t_cleanup = Builder::new() .name("solana-ledger-cleanup".to_string()) .spawn(move || loop { if exit.load(Ordering::Relaxed) { break; } if let Err(e) = Self::cleanup_ledger( &new_root_receiver, &blockstore, max_ledger_shreds, &mut last_purge_slot, DEFAULT_PURGE_SLOT_INTERVAL, ) { match e { RecvTimeoutError::Disconnected => break, RecvTimeoutError::Timeout => (), } } }) .unwrap(); Self { t_cleanup } } fn find_slots_to_clean( blockstore: &Arc<Blockstore>, root: Slot, max_ledger_shreds: u64, ) -> (u64, Slot, Slot) { let mut shreds = Vec::new(); let mut iterate_time = Measure::start("iterate_time"); let mut total_shreds = 0; let mut first_slot = 0; for (i, (slot, meta)) in blockstore.slot_meta_iterator(0).unwrap().enumerate() { if i == 0 { first_slot = slot; debug!("purge: searching from slot: {}", slot); } // Not exact since non-full slots will have holes total_shreds += meta.received; shreds.push((slot, meta.received)); if slot > root { break; } } iterate_time.stop(); info!( "checking for ledger purge: max_shreds: {} slots: {} total_shreds: {} {}", max_ledger_shreds, shreds.len(), total_shreds, iterate_time ); if (total_shreds as u64) < max_ledger_shreds { return (0, 0, 0); } let mut cur_shreds = 0; let mut lowest_slot_to_clean = shreds[0].0; for (slot, num_shreds) in shreds.iter().rev() { cur_shreds += *num_shreds as u64; if cur_shreds > max_ledger_shreds { lowest_slot_to_clean = *slot; break; } } (cur_shreds, lowest_slot_to_clean, first_slot) } pub fn cleanup_ledger( new_root_receiver: &Receiver<Slot>, blockstore: &Arc<Blockstore>, max_ledger_shreds: u64, last_purge_slot: &mut u64, purge_interval: u64, ) -> Result<(), RecvTimeoutError> { let mut root = new_root_receiver.recv_timeout(Duration::from_secs(1))?; // Get the newest root while let Ok(new_root) = new_root_receiver.try_recv() { root = new_root; } if root - *last_purge_slot > purge_interval { let disk_utilization_pre = blockstore.storage_size(); info!( "purge: new root: {} last_purge: {} purge_interval: {} disk: {:?}", root, last_purge_slot, purge_interval, disk_utilization_pre ); *last_purge_slot = root; let (num_shreds_to_clean, lowest_slot_to_clean, mut first_slot) = Self::find_slots_to_clean(blockstore, root, max_ledger_shreds); if num_shreds_to_clean > 0 { debug!( "cleaning up to: {} shreds: {} first: {}", lowest_slot_to_clean, num_shreds_to_clean, first_slot ); loop { let current_lowest = std::cmp::min(lowest_slot_to_clean, first_slot + DEFAULT_PURGE_BATCH_SIZE); let mut slot_update_time = Measure::start("slot_update"); *blockstore.lowest_cleanup_slot.write().unwrap() = current_lowest; slot_update_time.stop(); let mut clean_time = Measure::start("ledger_clean"); blockstore.purge_slots(first_slot, Some(current_lowest)); clean_time.stop(); debug!( "ledger purge {} -> {}: {} {}", first_slot, current_lowest, slot_update_time, clean_time ); first_slot += DEFAULT_PURGE_BATCH_SIZE; if current_lowest == lowest_slot_to_clean { break; } thread::sleep(Duration::from_millis(500)); } } let disk_utilization_post = blockstore.storage_size(); Self::report_disk_metrics(disk_utilization_pre, disk_utilization_post); } Ok(()) } fn report_disk_metrics(pre: BlockstoreResult<u64>, post: BlockstoreResult<u64>) { if let (Ok(pre), Ok(post)) = (pre, post) { datapoint_debug!( "ledger_disk_utilization", ("disk_utilization_pre", pre as i64, i64), ("disk_utilization_post", post as i64, i64), ("disk_utilization_delta", (pre as i64 - post as i64), i64) ); } } pub fn join(self) -> thread::Result<()> { self.t_cleanup.join() } } #[cfg(test)] mod tests { use super::*; use solana_ledger::blockstore::make_many_slot_entries; use solana_ledger::get_tmp_ledger_path; use std::sync::mpsc::channel; #[test] fn test_cleanup() { solana_logger::setup(); let blockstore_path = get_tmp_ledger_path!(); let blockstore = Blockstore::open(&blockstore_path).unwrap(); let (shreds, _) = make_many_slot_entries(0, 50, 5); blockstore.insert_shreds(shreds, None, false).unwrap();
//send a signal to kill all but 5 shreds, which will be in the newest slots let mut last_purge_slot = 0; sender.send(50).unwrap(); LedgerCleanupService::cleanup_ledger(&receiver, &blockstore, 5, &mut last_purge_slot, 10) .unwrap(); //check that 0-40 don't exist blockstore .slot_meta_iterator(0) .unwrap() .for_each(|(slot, _)| assert!(slot > 40)); drop(blockstore); Blockstore::destroy(&blockstore_path).expect("Expected successful database destruction"); } #[test] fn test_cleanup_speed() { solana_logger::setup(); let blockstore_path = get_tmp_ledger_path!(); let mut blockstore = Blockstore::open(&blockstore_path).unwrap(); blockstore.set_no_compaction(true); let blockstore = Arc::new(blockstore); let (sender, receiver) = channel(); let mut first_insert = Measure::start("first_insert"); let initial_slots = 50; let initial_entries = 5; let (shreds, _) = make_many_slot_entries(0, initial_slots, initial_entries); blockstore.insert_shreds(shreds, None, false).unwrap();
let blockstore = Arc::new(blockstore); let (sender, receiver) = channel();
random_line_split