file_name
large_stringlengths
4
140
prefix
large_stringlengths
0
12.1k
suffix
large_stringlengths
0
12k
middle
large_stringlengths
0
7.51k
fim_type
large_stringclasses
4 values
SpatialGP.py
(lldda1, lldda2, params=None): import sigvisa.utils.geog as geog import numpy as np ll = geog.dist_km(tuple(lldda1[0:2]), tuple(lldda2[0:2])) depth = ( lldda1[2] - lldda2[2] ) * params[0] r = np.sqrt(ll ** 2 + depth ** 2) return r def lon_lat_depth_distfn_deriv(i, lldda1, lldda2, params=None): import sigvisa.utils.geog as geog import numpy as np assert (i == 0) ll = geog.dist_km(tuple(lldda1[0:2]), tuple(lldda2[0:2])) depth = ( lldda1[2] - lldda2[2] ) * params[0] r = np.sqrt(ll ** 2 + depth ** 2) return ( params[0] * ( lldda1[2] - lldda2[2] )**2 ) / r if r != 0 else 0.0 def logdist_diff_distfn(lldda1, lldda2, params=None): import numpy as np dist = np.log(lldda1[3] + 1) - np.log(lldda2[3] + 1) return dist def azi_diff_distfn(lldda1, lldda2, params=None): import sigvisa.utils.geog as geog import numpy as np azi = np.abs ( geog.degdiff(lldda1[4], lldda2[4]) ) return azi def logdepth_diff_distfn(lldda1, lldda2, params=None): import numpy as np depth = np.log(lldda1[2] + 1) - np.log(lldda2[2] + 1) return depth X_LON, X_LAT, X_DEPTH, X_DIST, X_AZI = range(5) def spatial_kernel_from_str(kernel_str, target=None, params=None): params = params if params is not None else start_params[kernel_str][target] priors = [None,] * len(params) # TODO: use real priors if kernel_str == "dad_log": k = kernels.setup_kernel(name='distfn', params = params, extra=[dist_azi_depth_distfn_log, dist_azi_depth_distfn_deriv_log], ) elif kernel_str == "lld": noise_kernel = kernels.DiagonalKernel(params=params[0:1], priors = priors[0:1]) local_kernel = kernels.DistFNKernel(params=params[1:4], priors=priors[1:4], distfn = lon_lat_depth_distfn, deriv=lon_lat_depth_distfn_deriv) k = noise_kernel + local_kernel elif kernel_str == "composite": # assume we are passed the following params/priors: # 0 : sigma2_n -- noise variance # 1 : sigma2_f_dist -- function variance wrt dist_diff # 2 : w_dist -- length scale for dist_diff # 3 : sigma2_f_azi -- function variance wrt azi_diff # 4 : w_azi -- length scale for azi_diff # 5 : sigma2_f_depth -- function variance wrt depth_diff # 6 : w_depth -- length scale for depth_diff # 7 : sigma2_f_local -- function variance wrt local_dist # 8 : w_local -- length scale for local_dist noise_kernel = kernels.DiagonalKernel(params=params[0:1], priors = priors[0:1]) distdiff_kernel = kernels.DistFNKernel(params=params[1:3], priors=priors[1:3], distfn = logdist_diff_distfn, deriv=None) azidiff_kernel = kernels.DistFNKernel(params=params[3:5], priors=priors[3:5], distfn = azi_diff_distfn, deriv=None) depthdiff_kernel = kernels.DistFNKernel(params=params[5:7], priors=priors[5:7], distfn = logdepth_diff_distfn, deriv=None) local_kernel = kernels.DistFNKernel(params=params[7:10], priors=priors[7:10], distfn = lon_lat_depth_distfn, deriv=lon_lat_depth_distfn_deriv) k = noise_kernel + distdiff_kernel + azidiff_kernel + depthdiff_kernel + local_kernel return k """ def spatial_kernel_from_str(target=None, params=None): params = params if params is not None else start_params_lld[target] return params """ class SpatialGP(GaussianProcess, ParamModel): def init_hyperparams(self, hyperparams): (noise_var, signal_var, ll_scale, d_scale) = hyperparams self.noise_var = noise_var self.dfn_params = np.array((ll_scale, d_scale), dtype=np.float) self.wfn_params = np.array((signal_var,), copy=True, dtype=np.float) def build_kernel_matrix(self, X, hyperparams): self.init_hyperparams(hyperparams) vt = VectorTree(X[0:1,:], 1, "lld", self.dfn_params) K = vt.kernel_matrix(X, X, "se", self.wfn_params, False) + self.noise_var * np.eye(len(X), dtype=np.float64) K += np.eye(K.shape[0], dtype=np.float64) * 1e-8 # try to avoid losing # positive-definiteness # to numeric issues return K def invert_kernel_matrix(self, K): L = None alpha = None try: L = scipy.linalg.cholesky(K, lower=True) alpha = scipy.linalg.cho_solve((L, True), self.y) Kinv = scipy.linalg.inv(K) except np.linalg.linalg.LinAlgError: #u,v = np.linalg.eig(K) #print K, u #import pdb; pdb.set_trace() raise except ValueError: raise return alpha, L, Kinv def build_parametric_model(self, alpha, Kinv_sp, H, b, B): # notation follows section 2.7 of Rasmussen and Williams Binv = scipy.linalg.inv(B) tmp = np.dot(H, alpha) + np.dot(Binv, b) # H * K^-1 * y + B^-1 * b HKinv = H * Kinv_sp M_inv = Binv + np.dot(HKinv, H.T) # here M = (inv(B) + # H*K^-1*H.T)^-1 is the # posterior covariance # matrix on the params. c = scipy.linalg.cholesky(M_inv, lower=True) # c = sqrt(inv(B) + H*K^-1*H.T) beta_bar = scipy.linalg.cho_solve((c, True), tmp) invc = scipy.linalg.inv(c) return c, beta_bar, invc, HKinv def sparsify(self, M): return scipy.sparse.csr_matrix(M * (np.abs(M) > self.sparse_threshold)) def sort_events(self, X, y): combined = np.hstack([X, np.reshape(y, (-1, 1))]) combined_sorted = np.array(sorted(combined, key = lambda x: x[0]), dtype=float) X_sorted = np.array(combined_sorted[:, :-1], copy=True, dtype=float) y_sorted = combined_sorted[:, -1].flatten() return X_sorted, y_sorted def __init__(self, X=None, y=None, fname=None, basisfns=None, hyperparams=None, param_mean=None, param_cov=None, compute_ll=False, compute_grad=False, sparse_threshold=1e-20, sta = None, sort_events=True): try: ParamModel.__init__(self, sta=sta) except KeyError: pass if fname is not None: self.load_trained_model(fname) else: if sort_events: X, y = self.sort_events(X, y) # arrange events by # lon/lat, as a # heuristic to expose # block structure in the # kernel matrix self.hyperparams = np.array(hyperparams) self.sparse_threshold = sparse_threshold self.X = X self.n = X.shape[0] self.basisfns = basisfns mu, self.y, H = self.setup_mean("parametric", X, y) # train model #t0 = time.time() K = self.build_kernel_matrix(self.X, hyperparams) #t1 = time.time() self.alpha, L, Kinv = self.invert_kernel_matrix(K) Kinv_tri = 2 * np.tril(Kinv, k=0) - np.diag(np.diag(Kinv)) #t2 = time.time() self.Kinv_sp = self.sparsify(Kinv) self.Kinv_sp_tri = self.sparsify(Kinv_tri) #t3
lon_lat_depth_distfn
identifier_name
SpatialGP.py
0.0 def logdist_diff_distfn(lldda1, lldda2, params=None): import numpy as np dist = np.log(lldda1[3] + 1) - np.log(lldda2[3] + 1) return dist def azi_diff_distfn(lldda1, lldda2, params=None): import sigvisa.utils.geog as geog import numpy as np azi = np.abs ( geog.degdiff(lldda1[4], lldda2[4]) ) return azi def logdepth_diff_distfn(lldda1, lldda2, params=None): import numpy as np depth = np.log(lldda1[2] + 1) - np.log(lldda2[2] + 1) return depth X_LON, X_LAT, X_DEPTH, X_DIST, X_AZI = range(5) def spatial_kernel_from_str(kernel_str, target=None, params=None): params = params if params is not None else start_params[kernel_str][target] priors = [None,] * len(params) # TODO: use real priors if kernel_str == "dad_log": k = kernels.setup_kernel(name='distfn', params = params, extra=[dist_azi_depth_distfn_log, dist_azi_depth_distfn_deriv_log], ) elif kernel_str == "lld": noise_kernel = kernels.DiagonalKernel(params=params[0:1], priors = priors[0:1]) local_kernel = kernels.DistFNKernel(params=params[1:4], priors=priors[1:4], distfn = lon_lat_depth_distfn, deriv=lon_lat_depth_distfn_deriv) k = noise_kernel + local_kernel elif kernel_str == "composite": # assume we are passed the following params/priors: # 0 : sigma2_n -- noise variance # 1 : sigma2_f_dist -- function variance wrt dist_diff # 2 : w_dist -- length scale for dist_diff # 3 : sigma2_f_azi -- function variance wrt azi_diff # 4 : w_azi -- length scale for azi_diff # 5 : sigma2_f_depth -- function variance wrt depth_diff # 6 : w_depth -- length scale for depth_diff # 7 : sigma2_f_local -- function variance wrt local_dist # 8 : w_local -- length scale for local_dist
return k """ def spatial_kernel_from_str(target=None, params=None): params = params if params is not None else start_params_lld[target] return params """ class SpatialGP(GaussianProcess, ParamModel): def init_hyperparams(self, hyperparams): (noise_var, signal_var, ll_scale, d_scale) = hyperparams self.noise_var = noise_var self.dfn_params = np.array((ll_scale, d_scale), dtype=np.float) self.wfn_params = np.array((signal_var,), copy=True, dtype=np.float) def build_kernel_matrix(self, X, hyperparams): self.init_hyperparams(hyperparams) vt = VectorTree(X[0:1,:], 1, "lld", self.dfn_params) K = vt.kernel_matrix(X, X, "se", self.wfn_params, False) + self.noise_var * np.eye(len(X), dtype=np.float64) K += np.eye(K.shape[0], dtype=np.float64) * 1e-8 # try to avoid losing # positive-definiteness # to numeric issues return K def invert_kernel_matrix(self, K): L = None alpha = None try: L = scipy.linalg.cholesky(K, lower=True) alpha = scipy.linalg.cho_solve((L, True), self.y) Kinv = scipy.linalg.inv(K) except np.linalg.linalg.LinAlgError: #u,v = np.linalg.eig(K) #print K, u #import pdb; pdb.set_trace() raise except ValueError: raise return alpha, L, Kinv def build_parametric_model(self, alpha, Kinv_sp, H, b, B): # notation follows section 2.7 of Rasmussen and Williams Binv = scipy.linalg.inv(B) tmp = np.dot(H, alpha) + np.dot(Binv, b) # H * K^-1 * y + B^-1 * b HKinv = H * Kinv_sp M_inv = Binv + np.dot(HKinv, H.T) # here M = (inv(B) + # H*K^-1*H.T)^-1 is the # posterior covariance # matrix on the params. c = scipy.linalg.cholesky(M_inv, lower=True) # c = sqrt(inv(B) + H*K^-1*H.T) beta_bar = scipy.linalg.cho_solve((c, True), tmp) invc = scipy.linalg.inv(c) return c, beta_bar, invc, HKinv def sparsify(self, M): return scipy.sparse.csr_matrix(M * (np.abs(M) > self.sparse_threshold)) def sort_events(self, X, y): combined = np.hstack([X, np.reshape(y, (-1, 1))]) combined_sorted = np.array(sorted(combined, key = lambda x: x[0]), dtype=float) X_sorted = np.array(combined_sorted[:, :-1], copy=True, dtype=float) y_sorted = combined_sorted[:, -1].flatten() return X_sorted, y_sorted def __init__(self, X=None, y=None, fname=None, basisfns=None, hyperparams=None, param_mean=None, param_cov=None, compute_ll=False, compute_grad=False, sparse_threshold=1e-20, sta = None, sort_events=True): try: ParamModel.__init__(self, sta=sta) except KeyError: pass if fname is not None: self.load_trained_model(fname) else: if sort_events: X, y = self.sort_events(X, y) # arrange events by # lon/lat, as a # heuristic to expose # block structure in the # kernel matrix self.hyperparams = np.array(hyperparams) self.sparse_threshold = sparse_threshold self.X = X self.n = X.shape[0] self.basisfns = basisfns mu, self.y, H = self.setup_mean("parametric", X, y) # train model #t0 = time.time() K = self.build_kernel_matrix(self.X, hyperparams) #t1 = time.time() self.alpha, L, Kinv = self.invert_kernel_matrix(K) Kinv_tri = 2 * np.tril(Kinv, k=0) - np.diag(np.diag(Kinv)) #t2 = time.time() self.Kinv_sp = self.sparsify(Kinv) self.Kinv_sp_tri = self.sparsify(Kinv_tri) #t3 = time.time() self.c,self.beta_bar, self.invc, self.HKinv = self.build_parametric_model(self.alpha, self.Kinv_sp, H, b=param_mean, B=param_cov) #t4 = time.time() r = self.y - np.dot(H.T, self.beta_bar) self.alpha_r = scipy.linalg.cho_solve((L, True), r) #t5 = time.time() self.build_point_tree(HKinv = self.HKinv, Kinv = Kinv_tri, Kinv_sp=self.Kinv_sp_tri, alpha_r = self.alpha_r) #t6 = time.time() # precompute training set log likelihood, so we don't need # to keep L around. z = np.dot(H.T, param_mean) - self.y B = param_cov if compute_ll: self._compute_marginal_likelihood(L=L, z=z, B=B, H=H, K=K, Kinv_sp=self.Kinv_sp_tri) else: self.ll = -np.inf #t7 = time.time() if compute_grad: self.ll_grad = self
noise_kernel = kernels.DiagonalKernel(params=params[0:1], priors = priors[0:1]) distdiff_kernel = kernels.DistFNKernel(params=params[1:3], priors=priors[1:3], distfn = logdist_diff_distfn, deriv=None) azidiff_kernel = kernels.DistFNKernel(params=params[3:5], priors=priors[3:5], distfn = azi_diff_distfn, deriv=None) depthdiff_kernel = kernels.DistFNKernel(params=params[5:7], priors=priors[5:7], distfn = logdepth_diff_distfn, deriv=None) local_kernel = kernels.DistFNKernel(params=params[7:10], priors=priors[7:10], distfn = lon_lat_depth_distfn, deriv=lon_lat_depth_distfn_deriv) k = noise_kernel + distdiff_kernel + azidiff_kernel + depthdiff_kernel + local_kernel
conditional_block
kubeconfig.go
("incorrect value of state parameter = %s", req.decodedState.Nonce) } oidcTokens, err := oidcIssuer.Exchange(ctx, req.code) if err != nil { return nil, kcerrors.NewBadRequest("error while exchaning oidc code for token = %v", err) } if len(oidcTokens.RefreshToken) == 0 { return nil, kcerrors.NewBadRequest("the refresh token is missing but required, try setting/unsetting \"oidc-offline-access-as-scope\" command line flag") } claims, err := oidcVerifier.Verify(ctx, oidcTokens.IDToken) if err != nil { return nil, kcerrors.New(http.StatusUnauthorized, err.Error()) } if len(claims.Email) == 0 { return nil, kcerrors.NewBadRequest("the token doesn't contain the mandatory \"email\" claim") } adminKubeConfig, err := clusterProvider.GetAdminKubeconfigForCustomerCluster(cluster) if err != nil { return nil, common.KubernetesErrorToHTTPError(err) } // create a kubeconfig that contains OIDC tokens oidcKubeCfg := clientcmdapi.NewConfig() { // grab admin kubeconfig to read the cluster info var clusterFromAdminKubeCfg *clientcmdapi.Cluster for clusterName, cluster := range adminKubeConfig.Clusters { if clusterName == req.ClusterID { clusterFromAdminKubeCfg = cluster } } if clusterFromAdminKubeCfg == nil { return nil, kcerrors.New(http.StatusInternalServerError, fmt.Sprintf("unable to construct kubeconfig because couldn't find %s cluster enty in existing kubecfg", req.ClusterID)) } // create cluster entry clientCmdCluster := clientcmdapi.NewCluster() clientCmdCluster.Server = clusterFromAdminKubeCfg.Server clientCmdCluster.CertificateAuthorityData = clusterFromAdminKubeCfg.CertificateAuthorityData oidcKubeCfg.Clusters[req.ClusterID] = clientCmdCluster // create auth entry clientCmdAuth := clientcmdapi.NewAuthInfo() clientCmdAuthProvider := &clientcmdapi.AuthProviderConfig{Config: map[string]string{}} clientCmdAuthProvider.Name = "oidc" clientCmdAuthProvider.Config["id-token"] = oidcTokens.IDToken clientCmdAuthProvider.Config["refresh-token"] = oidcTokens.RefreshToken clientCmdAuthProvider.Config["idp-issuer-url"] = oidcCfg.URL clientCmdAuthProvider.Config["client-id"] = oidcCfg.ClientID clientCmdAuthProvider.Config["client-secret"] = oidcCfg.ClientSecret clientCmdAuth.AuthProvider = clientCmdAuthProvider oidcKubeCfg.AuthInfos[claims.Email] = clientCmdAuth // create default ctx clientCmdCtx := clientcmdapi.NewContext() clientCmdCtx.Cluster = req.ClusterID clientCmdCtx.AuthInfo = claims.Email oidcKubeCfg.Contexts["default"] = clientCmdCtx oidcKubeCfg.CurrentContext = "default" } // prepare final rsp that holds kubeconfig rsp := createOIDCKubeconfigRsp{} rsp.phase = kubeconfigGenerated rsp.oidcKubeConfig = oidcKubeCfg rsp.secureCookieMode = oidcCfg.CookieSecureMode return rsp, nil } // PHASE initial handles request from the end-user that wants to authenticate // and kicksoff the process of kubeconfig generation if req.phase != initialPhase { return nil, kcerrors.NewBadRequest(fmt.Sprintf("bad request unexpected phase = %d, expected phase = %d, did you forget to set the phase while decoding the request ?", req.phase, initialPhase)) } rsp := createOIDCKubeconfigRsp{} scopes := []string{"openid", "email"} if oidcCfg.OfflineAccessAsScope { scopes = append(scopes, "offline_access") } // pass nonce nonce := rand.String(rand.IntnRange(10, 15)) rsp.nonce = nonce rsp.secureCookieMode = oidcCfg.CookieSecureMode oidcState := OIDCState{ Nonce: nonce, ClusterID: req.ClusterID, ProjectID: req.ProjectID, UserID: req.UserID, Datacenter: req.Datacenter, } rawState, err := json.Marshal(oidcState) if err != nil { return nil, err } encodedState := base64.StdEncoding.EncodeToString(rawState) urlSafeState := url.QueryEscape(encodedState) rsp.authCodeURL = oidcIssuer.AuthCodeURL(urlSafeState, oidcCfg.OfflineAccessAsScope, scopes...) return rsp, nil } } type encodeKubeConifgResponse struct { clientCfg *clientcmdapi.Config filePrefix string } func EncodeKubeconfig(c context.Context, w http.ResponseWriter, response interface{}) (err error) { rsp := response.(*encodeKubeConifgResponse) cfg := rsp.clientCfg filename := "kubeconfig" if len(rsp.filePrefix) > 0 { filename = fmt.Sprintf("%s-%s", filename, rsp.filePrefix) } if len(cfg.Contexts) > 0 { filename = fmt.Sprintf("%s-%s", filename, cfg.Contexts[cfg.CurrentContext].Cluster) } w.Header().Set("Content-Type", "application/yaml") w.Header().Set("Content-disposition", fmt.Sprintf("attachment; filename=%s", filename)) w.Header().Add("Cache-Control", "no-cache") b, err := clientcmd.Write(*cfg) if err != nil { return err } _, err = w.Write(b) return err } type createOIDCKubeconfigRsp struct { // authCodeURL holds a URL to OpenID provider's consent page that asks for permissions for the required scopes explicitly. authCodeURL string // phase tells encoding function how to handle response phase int // oidcKubeConfig holds not serialized kubeconfig oidcKubeConfig *clientcmdapi.Config // nonce holds an arbitrary number storied in cookie to prevent Cross-site Request Forgery attack. nonce string // cookie received only with HTTPS, never with HTTP. secureCookieMode bool } func EncodeOIDCKubeconfig(c context.Context, w http.ResponseWriter, response interface{}) (err error) { rsp := response.(createOIDCKubeconfigRsp) // handles kubeconfig Generated PHASE // it means that kubeconfig was generated and we need to properly encode it. if rsp.phase == kubeconfigGenerated { // clear cookie by setting MaxAge<0 err = setCookie(w, "", rsp.secureCookieMode, -1) if err != nil { return fmt.Errorf("the cookie can't be removed, err = %v", err) } return EncodeKubeconfig(c, w, &encodeKubeConifgResponse{clientCfg: rsp.oidcKubeConfig}) } // handles initialPhase // redirects request to OpenID provider's consent page // and set cookie with nonce err = setCookie(w, rsp.nonce, rsp.secureCookieMode, cookieMaxAge) if err != nil { return fmt.Errorf("the cookie can't be created, err = %v", err) } w.Header().Add("Location", rsp.authCodeURL) w.Header().Add("Cache-Control", "no-cache") w.WriteHeader(http.StatusSeeOther) return nil } func DecodeGetAdminKubeconfig(c context.Context, r *http.Request) (interface{}, error) { req, err := common.DecodeGetClusterReq(c, r) if err != nil { return nil, err } return req, nil } const ( initialPhase = iota exchangeCodePhase = iota kubeconfigGenerated = iota ) // OIDCState holds data that are send and retrieved from OIDC provider type OIDCState struct { // nonce a random string that binds requests / responses of API server and OIDC provider // see https://tools.ietf.org/html/rfc6749#section-10.12 Nonce string `json:"nonce"` ClusterID string `json:"cluster_id"` ProjectID string `json:"project_id"` // UserID holds the ID of the user on behalf of which the request is being handled. UserID string `json:"user_id"` Datacenter string `json:"datacenter"` } // CreateOIDCKubeconfigReq represent a request for creating kubeconfig for a cluster with OIDC credentials // swagger:parameters createOIDCKubeconfig type CreateOIDCKubeconfigReq struct { // in: query ClusterID string `json:"cluster_id,omitempty"` ProjectID string `json:"project_id,omitempty"` UserID string `json:"user_id,omitempty"` Datacenter string `json:"datacenter,omitempty"` // not exported so that they don't leak to swagger spec. code string encodedState string decodedState OIDCState phase int cookieNonceValue string } func
DecodeCreateOIDCKubeconfig
identifier_name
kubeconfig.go
cluster.Spec.OIDC.ExtraScopes } clientCmdAuth.AuthProvider = clientCmdAuthProvider adminClientCfg.AuthInfos = map[string]*clientcmdapi.AuthInfo{} adminClientCfg.AuthInfos["default"] = clientCmdAuth return &encodeKubeConifgResponse{clientCfg: adminClientCfg, filePrefix: "oidc"}, nil } } func CreateOIDCKubeconfigEndpoint(projectProvider provider.ProjectProvider, oidcIssuerVerifier auth.OIDCIssuerVerifier, oidcCfg common.OIDCConfiguration) endpoint.Endpoint { return func(ctx context.Context, request interface{}) (interface{}, error) { oidcIssuer := oidcIssuerVerifier.(auth.OIDCIssuer) oidcVerifier := oidcIssuerVerifier.(auth.TokenVerifier) req := request.(CreateOIDCKubeconfigReq) clusterProvider := ctx.Value(middleware.ClusterProviderContextKey).(provider.ClusterProvider) userInfo := ctx.Value(middleware.UserInfoContextKey).(*provider.UserInfo) if secureCookie == nil { secureCookie = securecookie.New([]byte(oidcCfg.CookieHashKey), nil) } _, err := projectProvider.Get(userInfo, req.ProjectID, &provider.ProjectGetOptions{}) if err != nil { return nil, common.KubernetesErrorToHTTPError(err) } cluster, err := clusterProvider.Get(userInfo, req.ClusterID, &provider.ClusterGetOptions{}) if err != nil { return nil, common.KubernetesErrorToHTTPError(err) } // PHASE exchangeCode handles callback response from OIDC provider // and generates kubeconfig if req.phase == exchangeCodePhase { // validate the state if req.decodedState.Nonce != req.cookieNonceValue { return nil, kcerrors.NewBadRequest("incorrect value of state parameter = %s", req.decodedState.Nonce) } oidcTokens, err := oidcIssuer.Exchange(ctx, req.code) if err != nil { return nil, kcerrors.NewBadRequest("error while exchaning oidc code for token = %v", err) } if len(oidcTokens.RefreshToken) == 0 { return nil, kcerrors.NewBadRequest("the refresh token is missing but required, try setting/unsetting \"oidc-offline-access-as-scope\" command line flag") } claims, err := oidcVerifier.Verify(ctx, oidcTokens.IDToken) if err != nil { return nil, kcerrors.New(http.StatusUnauthorized, err.Error()) } if len(claims.Email) == 0 { return nil, kcerrors.NewBadRequest("the token doesn't contain the mandatory \"email\" claim") } adminKubeConfig, err := clusterProvider.GetAdminKubeconfigForCustomerCluster(cluster) if err != nil { return nil, common.KubernetesErrorToHTTPError(err) } // create a kubeconfig that contains OIDC tokens oidcKubeCfg := clientcmdapi.NewConfig() { // grab admin kubeconfig to read the cluster info var clusterFromAdminKubeCfg *clientcmdapi.Cluster for clusterName, cluster := range adminKubeConfig.Clusters { if clusterName == req.ClusterID { clusterFromAdminKubeCfg = cluster } } if clusterFromAdminKubeCfg == nil { return nil, kcerrors.New(http.StatusInternalServerError, fmt.Sprintf("unable to construct kubeconfig because couldn't find %s cluster enty in existing kubecfg", req.ClusterID)) } // create cluster entry clientCmdCluster := clientcmdapi.NewCluster() clientCmdCluster.Server = clusterFromAdminKubeCfg.Server clientCmdCluster.CertificateAuthorityData = clusterFromAdminKubeCfg.CertificateAuthorityData oidcKubeCfg.Clusters[req.ClusterID] = clientCmdCluster // create auth entry clientCmdAuth := clientcmdapi.NewAuthInfo() clientCmdAuthProvider := &clientcmdapi.AuthProviderConfig{Config: map[string]string{}} clientCmdAuthProvider.Name = "oidc" clientCmdAuthProvider.Config["id-token"] = oidcTokens.IDToken clientCmdAuthProvider.Config["refresh-token"] = oidcTokens.RefreshToken clientCmdAuthProvider.Config["idp-issuer-url"] = oidcCfg.URL clientCmdAuthProvider.Config["client-id"] = oidcCfg.ClientID clientCmdAuthProvider.Config["client-secret"] = oidcCfg.ClientSecret clientCmdAuth.AuthProvider = clientCmdAuthProvider oidcKubeCfg.AuthInfos[claims.Email] = clientCmdAuth // create default ctx clientCmdCtx := clientcmdapi.NewContext() clientCmdCtx.Cluster = req.ClusterID clientCmdCtx.AuthInfo = claims.Email oidcKubeCfg.Contexts["default"] = clientCmdCtx oidcKubeCfg.CurrentContext = "default" } // prepare final rsp that holds kubeconfig rsp := createOIDCKubeconfigRsp{} rsp.phase = kubeconfigGenerated rsp.oidcKubeConfig = oidcKubeCfg rsp.secureCookieMode = oidcCfg.CookieSecureMode return rsp, nil } // PHASE initial handles request from the end-user that wants to authenticate // and kicksoff the process of kubeconfig generation if req.phase != initialPhase { return nil, kcerrors.NewBadRequest(fmt.Sprintf("bad request unexpected phase = %d, expected phase = %d, did you forget to set the phase while decoding the request ?", req.phase, initialPhase)) } rsp := createOIDCKubeconfigRsp{} scopes := []string{"openid", "email"} if oidcCfg.OfflineAccessAsScope { scopes = append(scopes, "offline_access") } // pass nonce nonce := rand.String(rand.IntnRange(10, 15)) rsp.nonce = nonce rsp.secureCookieMode = oidcCfg.CookieSecureMode oidcState := OIDCState{ Nonce: nonce, ClusterID: req.ClusterID, ProjectID: req.ProjectID, UserID: req.UserID, Datacenter: req.Datacenter, } rawState, err := json.Marshal(oidcState) if err != nil { return nil, err } encodedState := base64.StdEncoding.EncodeToString(rawState) urlSafeState := url.QueryEscape(encodedState) rsp.authCodeURL = oidcIssuer.AuthCodeURL(urlSafeState, oidcCfg.OfflineAccessAsScope, scopes...) return rsp, nil } } type encodeKubeConifgResponse struct { clientCfg *clientcmdapi.Config filePrefix string } func EncodeKubeconfig(c context.Context, w http.ResponseWriter, response interface{}) (err error) { rsp := response.(*encodeKubeConifgResponse) cfg := rsp.clientCfg filename := "kubeconfig" if len(rsp.filePrefix) > 0 { filename = fmt.Sprintf("%s-%s", filename, rsp.filePrefix) } if len(cfg.Contexts) > 0 { filename = fmt.Sprintf("%s-%s", filename, cfg.Contexts[cfg.CurrentContext].Cluster) } w.Header().Set("Content-Type", "application/yaml") w.Header().Set("Content-disposition", fmt.Sprintf("attachment; filename=%s", filename)) w.Header().Add("Cache-Control", "no-cache") b, err := clientcmd.Write(*cfg) if err != nil { return err } _, err = w.Write(b) return err } type createOIDCKubeconfigRsp struct { // authCodeURL holds a URL to OpenID provider's consent page that asks for permissions for the required scopes explicitly. authCodeURL string // phase tells encoding function how to handle response phase int // oidcKubeConfig holds not serialized kubeconfig oidcKubeConfig *clientcmdapi.Config // nonce holds an arbitrary number storied in cookie to prevent Cross-site Request Forgery attack. nonce string // cookie received only with HTTPS, never with HTTP. secureCookieMode bool } func EncodeOIDCKubeconfig(c context.Context, w http.ResponseWriter, response interface{}) (err error)
{ rsp := response.(createOIDCKubeconfigRsp) // handles kubeconfig Generated PHASE // it means that kubeconfig was generated and we need to properly encode it. if rsp.phase == kubeconfigGenerated { // clear cookie by setting MaxAge<0 err = setCookie(w, "", rsp.secureCookieMode, -1) if err != nil { return fmt.Errorf("the cookie can't be removed, err = %v", err) } return EncodeKubeconfig(c, w, &encodeKubeConifgResponse{clientCfg: rsp.oidcKubeConfig}) } // handles initialPhase // redirects request to OpenID provider's consent page // and set cookie with nonce err = setCookie(w, rsp.nonce, rsp.secureCookieMode, cookieMaxAge) if err != nil { return fmt.Errorf("the cookie can't be created, err = %v", err)
identifier_body
kubeconfig.go
CreateOIDCKubeconfigEndpoint(projectProvider provider.ProjectProvider, oidcIssuerVerifier auth.OIDCIssuerVerifier, oidcCfg common.OIDCConfiguration) endpoint.Endpoint { return func(ctx context.Context, request interface{}) (interface{}, error) { oidcIssuer := oidcIssuerVerifier.(auth.OIDCIssuer) oidcVerifier := oidcIssuerVerifier.(auth.TokenVerifier) req := request.(CreateOIDCKubeconfigReq) clusterProvider := ctx.Value(middleware.ClusterProviderContextKey).(provider.ClusterProvider) userInfo := ctx.Value(middleware.UserInfoContextKey).(*provider.UserInfo) if secureCookie == nil { secureCookie = securecookie.New([]byte(oidcCfg.CookieHashKey), nil) } _, err := projectProvider.Get(userInfo, req.ProjectID, &provider.ProjectGetOptions{}) if err != nil { return nil, common.KubernetesErrorToHTTPError(err) } cluster, err := clusterProvider.Get(userInfo, req.ClusterID, &provider.ClusterGetOptions{}) if err != nil { return nil, common.KubernetesErrorToHTTPError(err) } // PHASE exchangeCode handles callback response from OIDC provider // and generates kubeconfig if req.phase == exchangeCodePhase { // validate the state if req.decodedState.Nonce != req.cookieNonceValue { return nil, kcerrors.NewBadRequest("incorrect value of state parameter = %s", req.decodedState.Nonce) } oidcTokens, err := oidcIssuer.Exchange(ctx, req.code) if err != nil { return nil, kcerrors.NewBadRequest("error while exchaning oidc code for token = %v", err) } if len(oidcTokens.RefreshToken) == 0 { return nil, kcerrors.NewBadRequest("the refresh token is missing but required, try setting/unsetting \"oidc-offline-access-as-scope\" command line flag") } claims, err := oidcVerifier.Verify(ctx, oidcTokens.IDToken) if err != nil { return nil, kcerrors.New(http.StatusUnauthorized, err.Error()) } if len(claims.Email) == 0 { return nil, kcerrors.NewBadRequest("the token doesn't contain the mandatory \"email\" claim") } adminKubeConfig, err := clusterProvider.GetAdminKubeconfigForCustomerCluster(cluster) if err != nil { return nil, common.KubernetesErrorToHTTPError(err) } // create a kubeconfig that contains OIDC tokens oidcKubeCfg := clientcmdapi.NewConfig() { // grab admin kubeconfig to read the cluster info var clusterFromAdminKubeCfg *clientcmdapi.Cluster for clusterName, cluster := range adminKubeConfig.Clusters { if clusterName == req.ClusterID { clusterFromAdminKubeCfg = cluster } } if clusterFromAdminKubeCfg == nil { return nil, kcerrors.New(http.StatusInternalServerError, fmt.Sprintf("unable to construct kubeconfig because couldn't find %s cluster enty in existing kubecfg", req.ClusterID)) } // create cluster entry clientCmdCluster := clientcmdapi.NewCluster() clientCmdCluster.Server = clusterFromAdminKubeCfg.Server clientCmdCluster.CertificateAuthorityData = clusterFromAdminKubeCfg.CertificateAuthorityData oidcKubeCfg.Clusters[req.ClusterID] = clientCmdCluster // create auth entry clientCmdAuth := clientcmdapi.NewAuthInfo() clientCmdAuthProvider := &clientcmdapi.AuthProviderConfig{Config: map[string]string{}} clientCmdAuthProvider.Name = "oidc" clientCmdAuthProvider.Config["id-token"] = oidcTokens.IDToken clientCmdAuthProvider.Config["refresh-token"] = oidcTokens.RefreshToken clientCmdAuthProvider.Config["idp-issuer-url"] = oidcCfg.URL clientCmdAuthProvider.Config["client-id"] = oidcCfg.ClientID clientCmdAuthProvider.Config["client-secret"] = oidcCfg.ClientSecret clientCmdAuth.AuthProvider = clientCmdAuthProvider oidcKubeCfg.AuthInfos[claims.Email] = clientCmdAuth // create default ctx clientCmdCtx := clientcmdapi.NewContext() clientCmdCtx.Cluster = req.ClusterID clientCmdCtx.AuthInfo = claims.Email oidcKubeCfg.Contexts["default"] = clientCmdCtx oidcKubeCfg.CurrentContext = "default" } // prepare final rsp that holds kubeconfig rsp := createOIDCKubeconfigRsp{} rsp.phase = kubeconfigGenerated rsp.oidcKubeConfig = oidcKubeCfg rsp.secureCookieMode = oidcCfg.CookieSecureMode return rsp, nil } // PHASE initial handles request from the end-user that wants to authenticate // and kicksoff the process of kubeconfig generation if req.phase != initialPhase { return nil, kcerrors.NewBadRequest(fmt.Sprintf("bad request unexpected phase = %d, expected phase = %d, did you forget to set the phase while decoding the request ?", req.phase, initialPhase)) } rsp := createOIDCKubeconfigRsp{} scopes := []string{"openid", "email"} if oidcCfg.OfflineAccessAsScope { scopes = append(scopes, "offline_access") } // pass nonce nonce := rand.String(rand.IntnRange(10, 15)) rsp.nonce = nonce rsp.secureCookieMode = oidcCfg.CookieSecureMode oidcState := OIDCState{ Nonce: nonce, ClusterID: req.ClusterID, ProjectID: req.ProjectID, UserID: req.UserID, Datacenter: req.Datacenter, } rawState, err := json.Marshal(oidcState) if err != nil { return nil, err } encodedState := base64.StdEncoding.EncodeToString(rawState) urlSafeState := url.QueryEscape(encodedState) rsp.authCodeURL = oidcIssuer.AuthCodeURL(urlSafeState, oidcCfg.OfflineAccessAsScope, scopes...) return rsp, nil } } type encodeKubeConifgResponse struct { clientCfg *clientcmdapi.Config filePrefix string } func EncodeKubeconfig(c context.Context, w http.ResponseWriter, response interface{}) (err error) { rsp := response.(*encodeKubeConifgResponse) cfg := rsp.clientCfg filename := "kubeconfig" if len(rsp.filePrefix) > 0 { filename = fmt.Sprintf("%s-%s", filename, rsp.filePrefix) } if len(cfg.Contexts) > 0 { filename = fmt.Sprintf("%s-%s", filename, cfg.Contexts[cfg.CurrentContext].Cluster) } w.Header().Set("Content-Type", "application/yaml") w.Header().Set("Content-disposition", fmt.Sprintf("attachment; filename=%s", filename)) w.Header().Add("Cache-Control", "no-cache") b, err := clientcmd.Write(*cfg) if err != nil { return err } _, err = w.Write(b) return err } type createOIDCKubeconfigRsp struct { // authCodeURL holds a URL to OpenID provider's consent page that asks for permissions for the required scopes explicitly. authCodeURL string // phase tells encoding function how to handle response phase int // oidcKubeConfig holds not serialized kubeconfig oidcKubeConfig *clientcmdapi.Config // nonce holds an arbitrary number storied in cookie to prevent Cross-site Request Forgery attack. nonce string // cookie received only with HTTPS, never with HTTP. secureCookieMode bool } func EncodeOIDCKubeconfig(c context.Context, w http.ResponseWriter, response interface{}) (err error) { rsp := response.(createOIDCKubeconfigRsp) // handles kubeconfig Generated PHASE // it means that kubeconfig was generated and we need to properly encode it. if rsp.phase == kubeconfigGenerated { // clear cookie by setting MaxAge<0 err = setCookie(w, "", rsp.secureCookieMode, -1) if err != nil { return fmt.Errorf("the cookie can't be removed, err = %v", err) } return EncodeKubeconfig(c, w, &encodeKubeConifgResponse{clientCfg: rsp.oidcKubeConfig}) } // handles initialPhase // redirects request to OpenID provider's consent page // and set cookie with nonce err = setCookie(w, rsp.nonce, rsp.secureCookieMode, cookieMaxAge) if err != nil { return fmt.Errorf("the cookie can't be created, err = %v", err) } w.Header().Add("Location", rsp.authCodeURL) w.Header().Add("Cache-Control", "no-cache") w.WriteHeader(http.StatusSeeOther) return nil } func DecodeGetAdminKubeconfig(c context.Context, r *http.Request) (interface{}, error) { req, err := common.DecodeGetClusterReq(c, r) if err != nil
{ return nil, err }
conditional_block
kubeconfig.go
.KubernetesErrorToHTTPError(err) } cluster, err := clusterProvider.Get(userInfo, req.ClusterID, &provider.ClusterGetOptions{}) if err != nil { return nil, common.KubernetesErrorToHTTPError(err) } adminClientCfg, err := clusterProvider.GetAdminKubeconfigForCustomerCluster(cluster) if err != nil { return nil, common.KubernetesErrorToHTTPError(err) } clientCmdAuth := clientcmdapi.NewAuthInfo() clientCmdAuthProvider := &clientcmdapi.AuthProviderConfig{Config: map[string]string{}} clientCmdAuthProvider.Name = "oidc" clientCmdAuthProvider.Config["idp-issuer-url"] = cluster.Spec.OIDC.IssuerURL clientCmdAuthProvider.Config["client-id"] = cluster.Spec.OIDC.ClientID if cluster.Spec.OIDC.ClientSecret != "" { clientCmdAuthProvider.Config["client-secret"] = cluster.Spec.OIDC.ClientSecret } if cluster.Spec.OIDC.ExtraScopes != "" { clientCmdAuthProvider.Config["extra-scopes"] = cluster.Spec.OIDC.ExtraScopes } clientCmdAuth.AuthProvider = clientCmdAuthProvider adminClientCfg.AuthInfos = map[string]*clientcmdapi.AuthInfo{} adminClientCfg.AuthInfos["default"] = clientCmdAuth return &encodeKubeConifgResponse{clientCfg: adminClientCfg, filePrefix: "oidc"}, nil } } func CreateOIDCKubeconfigEndpoint(projectProvider provider.ProjectProvider, oidcIssuerVerifier auth.OIDCIssuerVerifier, oidcCfg common.OIDCConfiguration) endpoint.Endpoint { return func(ctx context.Context, request interface{}) (interface{}, error) { oidcIssuer := oidcIssuerVerifier.(auth.OIDCIssuer) oidcVerifier := oidcIssuerVerifier.(auth.TokenVerifier) req := request.(CreateOIDCKubeconfigReq) clusterProvider := ctx.Value(middleware.ClusterProviderContextKey).(provider.ClusterProvider) userInfo := ctx.Value(middleware.UserInfoContextKey).(*provider.UserInfo) if secureCookie == nil { secureCookie = securecookie.New([]byte(oidcCfg.CookieHashKey), nil) } _, err := projectProvider.Get(userInfo, req.ProjectID, &provider.ProjectGetOptions{}) if err != nil { return nil, common.KubernetesErrorToHTTPError(err) } cluster, err := clusterProvider.Get(userInfo, req.ClusterID, &provider.ClusterGetOptions{}) if err != nil { return nil, common.KubernetesErrorToHTTPError(err) } // PHASE exchangeCode handles callback response from OIDC provider // and generates kubeconfig if req.phase == exchangeCodePhase { // validate the state if req.decodedState.Nonce != req.cookieNonceValue { return nil, kcerrors.NewBadRequest("incorrect value of state parameter = %s", req.decodedState.Nonce) } oidcTokens, err := oidcIssuer.Exchange(ctx, req.code) if err != nil { return nil, kcerrors.NewBadRequest("error while exchaning oidc code for token = %v", err) } if len(oidcTokens.RefreshToken) == 0 { return nil, kcerrors.NewBadRequest("the refresh token is missing but required, try setting/unsetting \"oidc-offline-access-as-scope\" command line flag") } claims, err := oidcVerifier.Verify(ctx, oidcTokens.IDToken) if err != nil { return nil, kcerrors.New(http.StatusUnauthorized, err.Error()) } if len(claims.Email) == 0 { return nil, kcerrors.NewBadRequest("the token doesn't contain the mandatory \"email\" claim") } adminKubeConfig, err := clusterProvider.GetAdminKubeconfigForCustomerCluster(cluster) if err != nil { return nil, common.KubernetesErrorToHTTPError(err) } // create a kubeconfig that contains OIDC tokens oidcKubeCfg := clientcmdapi.NewConfig() { // grab admin kubeconfig to read the cluster info var clusterFromAdminKubeCfg *clientcmdapi.Cluster for clusterName, cluster := range adminKubeConfig.Clusters { if clusterName == req.ClusterID { clusterFromAdminKubeCfg = cluster } } if clusterFromAdminKubeCfg == nil { return nil, kcerrors.New(http.StatusInternalServerError, fmt.Sprintf("unable to construct kubeconfig because couldn't find %s cluster enty in existing kubecfg", req.ClusterID)) } // create cluster entry clientCmdCluster := clientcmdapi.NewCluster() clientCmdCluster.Server = clusterFromAdminKubeCfg.Server clientCmdCluster.CertificateAuthorityData = clusterFromAdminKubeCfg.CertificateAuthorityData oidcKubeCfg.Clusters[req.ClusterID] = clientCmdCluster // create auth entry clientCmdAuth := clientcmdapi.NewAuthInfo() clientCmdAuthProvider := &clientcmdapi.AuthProviderConfig{Config: map[string]string{}} clientCmdAuthProvider.Name = "oidc" clientCmdAuthProvider.Config["id-token"] = oidcTokens.IDToken clientCmdAuthProvider.Config["refresh-token"] = oidcTokens.RefreshToken clientCmdAuthProvider.Config["idp-issuer-url"] = oidcCfg.URL clientCmdAuthProvider.Config["client-id"] = oidcCfg.ClientID clientCmdAuthProvider.Config["client-secret"] = oidcCfg.ClientSecret clientCmdAuth.AuthProvider = clientCmdAuthProvider oidcKubeCfg.AuthInfos[claims.Email] = clientCmdAuth // create default ctx clientCmdCtx := clientcmdapi.NewContext() clientCmdCtx.Cluster = req.ClusterID clientCmdCtx.AuthInfo = claims.Email oidcKubeCfg.Contexts["default"] = clientCmdCtx oidcKubeCfg.CurrentContext = "default" } // prepare final rsp that holds kubeconfig rsp := createOIDCKubeconfigRsp{} rsp.phase = kubeconfigGenerated rsp.oidcKubeConfig = oidcKubeCfg rsp.secureCookieMode = oidcCfg.CookieSecureMode return rsp, nil } // PHASE initial handles request from the end-user that wants to authenticate // and kicksoff the process of kubeconfig generation if req.phase != initialPhase { return nil, kcerrors.NewBadRequest(fmt.Sprintf("bad request unexpected phase = %d, expected phase = %d, did you forget to set the phase while decoding the request ?", req.phase, initialPhase)) } rsp := createOIDCKubeconfigRsp{} scopes := []string{"openid", "email"} if oidcCfg.OfflineAccessAsScope { scopes = append(scopes, "offline_access") } // pass nonce nonce := rand.String(rand.IntnRange(10, 15)) rsp.nonce = nonce rsp.secureCookieMode = oidcCfg.CookieSecureMode oidcState := OIDCState{ Nonce: nonce, ClusterID: req.ClusterID, ProjectID: req.ProjectID, UserID: req.UserID, Datacenter: req.Datacenter, } rawState, err := json.Marshal(oidcState) if err != nil { return nil, err } encodedState := base64.StdEncoding.EncodeToString(rawState) urlSafeState := url.QueryEscape(encodedState)
return rsp, nil } } type encodeKubeConifgResponse struct { clientCfg *clientcmdapi.Config filePrefix string } func EncodeKubeconfig(c context.Context, w http.ResponseWriter, response interface{}) (err error) { rsp := response.(*encodeKubeConifgResponse) cfg := rsp.clientCfg filename := "kubeconfig" if len(rsp.filePrefix) > 0 { filename = fmt.Sprintf("%s-%s", filename, rsp.filePrefix) } if len(cfg.Contexts) > 0 { filename = fmt.Sprintf("%s-%s", filename, cfg.Contexts[cfg.CurrentContext].Cluster) } w.Header().Set("Content-Type", "application/yaml") w.Header().Set("Content-disposition", fmt.Sprintf("attachment; filename=%s", filename)) w.Header().Add("Cache-Control", "no-cache") b, err := clientcmd.Write(*cfg) if err != nil { return err } _, err = w.Write(b) return err } type createOIDCKubeconfigRsp struct { // authCodeURL holds a URL to OpenID provider's consent page that asks for permissions for the required scopes explicitly. authCodeURL string // phase tells encoding function how to handle response phase int // oidcKubeConfig holds not serialized kubeconfig oidcKubeConfig *clientcmdapi.Config // nonce holds an arbitrary number storied in cookie to prevent Cross-site Request Forgery attack. nonce string // cookie received only with HTTPS, never with HTTP.
rsp.authCodeURL = oidcIssuer.AuthCodeURL(urlSafeState, oidcCfg.OfflineAccessAsScope, scopes...)
random_line_split
corrector.py
("decode_while_training", False, "Do test decode while training.") tf.app.flags.DEFINE_boolean("decode_whole", False, "decode whole code or just a set.") FLAGS = tf.app.flags.FLAGS # We use a number of buckets and pad to the closest one for efficiency. # See seq2seq_model.Seq2SeqModel for details of how they work. # encoding 이 두개니까 버킷을 어떻게 잡을지 생각좀 해봐야 # _buckets = [(5, 10), (10, 15), (20, 25), (40, 50)] # _buckets = [(10, 15), (40, 50)] _buckets = [(50, 50)] def read_data(train_id_set, max_size=None): data_set = [[] for _ in _buckets] counter = 0 saved = 0 for code i
source_ids = [code[line_idx], code[line_idx + 1]] target_ids = [[]] target_ids[0].append(code_loader.EOS_ID) for bucket_id, (source_size, target_size) in enumerate(_buckets): if len(source_ids[0]) < source_size and len(source_ids[1]) < source_size and len(target_ids[0]) < target_size: saved += 1 data_set[bucket_id].append([source_ids, target_ids]) break print(" not saved: %d" % (counter - saved)) print(" read data line total %d" % counter) return data_set def create_model(session, forward_only, vocab_size=FLAGS.vocab_size): " ""Create translation model and initialize or load parameters in session.""" dtype = tf.float16 if FLAGS.use_fp16 else tf.float32 model = code_model.CodeModel( vocab_size, _buckets, FLAGS.size, FLAGS.num_layers, FLAGS.max_gradient_norm, FLAGS.batch_size, FLAGS.learning_rate, FLAGS.learning_rate_decay_factor, forward_only=forward_only, dtype=dtype) ckpt = tf.train.get_checkpoint_state(FLAGS.train_dir) if ckpt and tf.train.checkpoint_exists(ckpt.model_checkpoint_path): print("Reading model parameters from %s" % ckpt.model_checkpoint_path) model.saver.restore(session, ckpt.model_checkpoint_path) else: print("Created model with fresh parameters.") session.run(tf.global_variables_initializer()) return model def get_perplexity(loss): return math.exp(float(loss)) if loss < 300 else float("inf") def save_checkpoint(model, sess, vocab_size): # Save checkpoint and zero timer and loss. with open(FLAGS.train_dir + "/" + os.path.basename(__file__) + ".ckpt.vb", mode="w") as vocab_size_file: pickle.dump(vocab_size, vocab_size_file) checkpoint_path = os.path.join(FLAGS.train_dir, FLAGS.data_path.split(".")[0] + ".ckpt.size." + str(FLAGS.size) + ".nl." + str(FLAGS.num_layers) + ".vs." + str(vocab_size)) model.saver.save(sess, checkpoint_path, global_step=model.global_step) def train(): train_id_data, id_to_vocab, vocab_to_id = code_loader.prepare_data(FLAGS.data_dir, FLAGS.vocab_size, data_path=FLAGS.data_path, cache=FLAGS.cache) vocab_size = FLAGS.vocab_size if FLAGS.vocab_size != 0 else len(id_to_vocab) # Create decode model # print("Creating %d layers of %d units for decode." % (FLAGS.num_layers, FLAGS.size)) # decode_sess = tf.Session() # decode_model = create_model(decode_sess, True, vocab_size=(FLAGS.vocab_size if FLAGS.vocab_size != 0 else len(id_to_vocab))) # Create model. with tf.Session() as sess: print("Creating %d layers of %d units with %d vocab." % (FLAGS.num_layers, FLAGS.size, vocab_size)) model = create_model(sess, False, vocab_size=vocab_size) # Read data into buckets and compute their sizes. print("Reading development and training data (limit: %d)." % FLAGS.max_train_data_size) # dev_set_path = FLAGS.train_dir + "/dev_set." + str(FLAGS.from_vocab_size) + "." + pickle.__name__ train_set_path = FLAGS.train_dir + "/" + FLAGS.data_path.split(".")[0] + ".train_set.ids" + str(FLAGS.vocab_size) + ".ds" + str(FLAGS.max_train_data_size) + "." + pickle.__name__ if not tf.gfile.Exists(train_set_path) or not FLAGS.cache: print("Reading training data (limit: %d)." % FLAGS.max_train_data_size) train_set = read_data(train_id_data, FLAGS.max_train_data_size) with tf.gfile.GFile(train_set_path, "w") as f: pickle.dump(train_set, f) else: print("Loading training data (limit: %d)." % FLAGS.max_train_data_size) with tf.gfile.GFile(train_set_path, mode="r") as f: train_set = pickle.load(f) train_bucket_sizes = [len(train_set[b]) for b in xrange(len(_buckets))] train_total_size = float(sum(train_bucket_sizes)) print("Total train %d" % train_total_size) # A bucket scale is a list of increasing numbers from 0 to 1 that we'll use # to select a bucket. Length of [scale[i], scale[i+1]] is proportional to # the size if i-th training bucket, as used later. train_buckets_scale = [sum(train_bucket_sizes[:i + 1]) / train_total_size for i in xrange(len(train_bucket_sizes))] # This is the training loop. print("Running the training loop") epoch_step_time, epoch_loss, ckpt_step_time, ckpt_loss = 0.0, 0.0, 0.0, 0.0 # current_step = 0 previous_losses = [] std_out = "Error: not enough steps to run with checkpoint_step." test_out = "" steps_per_epoch = round(int(train_total_size) / FLAGS.batch_size) steps_per_checkpoint = FLAGS.steps_per_checkpoint if FLAGS.steps_per_checkpoint != 0 else steps_per_epoch # batch * total data size epoch_step = model.global_step.eval() // steps_per_epoch while True: # Choose a bucket according to data distribution. We pick a random number # in [0, 1] and use the corresponding interval in train_buckets_scale. random_number_01 = np.random.random_sample() bucket_id = min([i for i in xrange(len(train_buckets_scale)) if train_buckets_scale[i] > random_number_01]) # Get a batch and make a step. start_time = time.time() encoder_inputs_front, encoder_inputs_back, decoder_inputs, target_weights = model.get_batch( train_set, bucket_id) _, step_loss, _ = model.step(sess, encoder_inputs_front, encoder_inputs_back, decoder_inputs, target_weights, bucket_id, False) ckpt_step_time += (time.time() - start_time) / steps_per_checkpoint ckpt_loss += step_loss / steps_per_checkpoint epoch_step_time += (time.time() - start_time) / steps_per_epoch epoch_loss += step_loss / steps_per_epoch # current_step += 1 # print condition if model.global_step.eval() % 1 == 0: print(" global step %d learning rate %.4f step-time %.2f loss %.2f perplexity " "%.2f" % (model.global_step.eval(), model.learning_rate.eval(), (time.time() - start_time), step_loss, get_perplexity(step_loss))) # per epoch if model.global_step.eval() % steps_per_epoch == 0: epoch_step += 1 print("epoch %d" % epoch_step) # Print statistics for the previous epoch. std_out = "epoch: global step %d learning rate %.4f step-time %.2f loss %.2f perplexity %.2f" % ( model.global_step.eval(), model.learning_rate.eval(), epoch_step_time, epoch_loss, get_perplexity(epoch_loss)) print(std_out) # Decrease learning rate if no improvement was seen over last 3 epoch times. if len(previous_losses) > 2 and epoch_loss > max(previous_losses[-3:]): sess.run(model.learning_rate_decay_op) previous_losses.append(epoch_loss) # with tf.variable_scope("decoding") as
n train_id_set: for line_idx in xrange(len(code) - 2): if max_size and counter >= max_size: break counter += 1 source_ids = [code[line_idx], code[line_idx + 2]] target_ids = [code[line_idx + 1]] target_ids[0].append(code_loader.EOS_ID) for bucket_id, (source_size, target_size) in enumerate(_buckets): if len(source_ids[0]) < source_size and len(source_ids[1]) < source_size and len(target_ids[0]) < target_size: saved += 1 data_set[bucket_id].append([source_ids, target_ids]) break for line_idx in xrange(len(code) - 1): if max_size and counter >= max_size: break counter += 1
identifier_body
corrector.py
_boolean("decode_while_training", False, "Do test decode while training.") tf.app.flags.DEFINE_boolean("decode_whole", False, "decode whole code or just a set.") FLAGS = tf.app.flags.FLAGS # We use a number of buckets and pad to the closest one for efficiency. # See seq2seq_model.Seq2SeqModel for details of how they work. # encoding 이 두개니까 버킷을 어떻게 잡을지 생각좀 해봐야 # _buckets = [(5, 10), (10, 15), (20, 25), (40, 50)] # _buckets = [(10, 15), (40, 50)] _buckets = [(50, 50)] def read_data(train_id_set, max_size=None): data_set = [[] for _ in _buckets] counter = 0 saved = 0 for code in train_id_set: for line_idx in xrange(len(code) - 2): if max_size and counter >= max_size: break counter += 1 source_ids = [code[line_idx], code[line_idx + 2]] target_ids = [code[line_idx + 1]] target_ids[0].append(code_loader.EOS_ID) for bucket_id, (source_size, target_size) in enumerate(_buckets): if len(source_ids[0]) < source_size and len(source_ids[1]) < source_size and len(target_ids[0]) < target_size: saved += 1 data_set[bucket_id].append([source_ids, target_ids]) break for line_idx in xrange(len(code) - 1): if max_size and counter >= max_size: break counter += 1 source_ids = [code[line_idx], code[line_idx + 1]] target_ids = [[]] target_ids[0].append(code_loader.EOS_ID) for bucket_id, (source_size, target_size) in enumerate(_buckets): if len(source_ids[0]) < source_size and len(source_ids[1]) < source_size and len(target_ids[0]) < target_size: saved += 1 data_set[bucket_id].append([source_ids, target_ids]) break print(" not saved: %d" % (counter - saved)) print(" read data line total %d" % counter) return data_set def create_model(session, forward_only, vocab_size=FLAGS.vocab_size): """Create translation model and initialize or load parameters in session."""
_buckets, FLAGS.size, FLAGS.num_layers, FLAGS.max_gradient_norm, FLAGS.batch_size, FLAGS.learning_rate, FLAGS.learning_rate_decay_factor, forward_only=forward_only, dtype=dtype) ckpt = tf.train.get_checkpoint_state(FLAGS.train_dir) if ckpt and tf.train.checkpoint_exists(ckpt.model_checkpoint_path): print("Reading model parameters from %s" % ckpt.model_checkpoint_path) model.saver.restore(session, ckpt.model_checkpoint_path) else: print("Created model with fresh parameters.") session.run(tf.global_variables_initializer()) return model def get_perplexity(loss): return math.exp(float(loss)) if loss < 300 else float("inf") def save_checkpoint(model, sess, vocab_size): # Save checkpoint and zero timer and loss. with open(FLAGS.train_dir + "/" + os.path.basename(__file__) + ".ckpt.vb", mode="w") as vocab_size_file: pickle.dump(vocab_size, vocab_size_file) checkpoint_path = os.path.join(FLAGS.train_dir, FLAGS.data_path.split(".")[0] + ".ckpt.size." + str(FLAGS.size) + ".nl." + str(FLAGS.num_layers) + ".vs." + str(vocab_size)) model.saver.save(sess, checkpoint_path, global_step=model.global_step) def train(): train_id_data, id_to_vocab, vocab_to_id = code_loader.prepare_data(FLAGS.data_dir, FLAGS.vocab_size, data_path=FLAGS.data_path, cache=FLAGS.cache) vocab_size = FLAGS.vocab_size if FLAGS.vocab_size != 0 else len(id_to_vocab) # Create decode model # print("Creating %d layers of %d units for decode." % (FLAGS.num_layers, FLAGS.size)) # decode_sess = tf.Session() # decode_model = create_model(decode_sess, True, vocab_size=(FLAGS.vocab_size if FLAGS.vocab_size != 0 else len(id_to_vocab))) # Create model. with tf.Session() as sess: print("Creating %d layers of %d units with %d vocab." % (FLAGS.num_layers, FLAGS.size, vocab_size)) model = create_model(sess, False, vocab_size=vocab_size) # Read data into buckets and compute their sizes. print("Reading development and training data (limit: %d)." % FLAGS.max_train_data_size) # dev_set_path = FLAGS.train_dir + "/dev_set." + str(FLAGS.from_vocab_size) + "." + pickle.__name__ train_set_path = FLAGS.train_dir + "/" + FLAGS.data_path.split(".")[0] + ".train_set.ids" + str(FLAGS.vocab_size) + ".ds" + str(FLAGS.max_train_data_size) + "." + pickle.__name__ if not tf.gfile.Exists(train_set_path) or not FLAGS.cache: print("Reading training data (limit: %d)." % FLAGS.max_train_data_size) train_set = read_data(train_id_data, FLAGS.max_train_data_size) with tf.gfile.GFile(train_set_path, "w") as f: pickle.dump(train_set, f) else: print("Loading training data (limit: %d)." % FLAGS.max_train_data_size) with tf.gfile.GFile(train_set_path, mode="r") as f: train_set = pickle.load(f) train_bucket_sizes = [len(train_set[b]) for b in xrange(len(_buckets))] train_total_size = float(sum(train_bucket_sizes)) print("Total train %d" % train_total_size) # A bucket scale is a list of increasing numbers from 0 to 1 that we'll use # to select a bucket. Length of [scale[i], scale[i+1]] is proportional to # the size if i-th training bucket, as used later. train_buckets_scale = [sum(train_bucket_sizes[:i + 1]) / train_total_size for i in xrange(len(train_bucket_sizes))] # This is the training loop. print("Running the training loop") epoch_step_time, epoch_loss, ckpt_step_time, ckpt_loss = 0.0, 0.0, 0.0, 0.0 # current_step = 0 previous_losses = [] std_out = "Error: not enough steps to run with checkpoint_step." test_out = "" steps_per_epoch = round(int(train_total_size) / FLAGS.batch_size) steps_per_checkpoint = FLAGS.steps_per_checkpoint if FLAGS.steps_per_checkpoint != 0 else steps_per_epoch # batch * total data size epoch_step = model.global_step.eval() // steps_per_epoch while True: # Choose a bucket according to data distribution. We pick a random number # in [0, 1] and use the corresponding interval in train_buckets_scale. random_number_01 = np.random.random_sample() bucket_id = min([i for i in xrange(len(train_buckets_scale)) if train_buckets_scale[i] > random_number_01]) # Get a batch and make a step. start_time = time.time() encoder_inputs_front, encoder_inputs_back, decoder_inputs, target_weights = model.get_batch( train_set, bucket_id) _, step_loss, _ = model.step(sess, encoder_inputs_front, encoder_inputs_back, decoder_inputs, target_weights, bucket_id, False) ckpt_step_time += (time.time() - start_time) / steps_per_checkpoint ckpt_loss += step_loss / steps_per_checkpoint epoch_step_time += (time.time() - start_time) / steps_per_epoch epoch_loss += step_loss / steps_per_epoch # current_step += 1 # print condition if model.global_step.eval() % 1 == 0: print(" global step %d learning rate %.4f step-time %.2f loss %.2f perplexity " "%.2f" % (model.global_step.eval(), model.learning_rate.eval(), (time.time() - start_time), step_loss, get_perplexity(step_loss))) # per epoch if model.global_step.eval() % steps_per_epoch == 0: epoch_step += 1 print("epoch %d" % epoch_step) # Print statistics for the previous epoch. std_out = "epoch: global step %d learning rate %.4f step-time %.2f loss %.2f perplexity %.2f" % ( model.global_step.eval(), model.learning_rate.eval(), epoch_step_time, epoch_loss, get_perplexity(epoch_loss)) print(std_out) # Decrease learning rate if no improvement was seen over last 3 epoch times. if len(previous_losses) > 2 and epoch_loss > max(previous_losses[-3:]): sess.run(model.learning_rate_decay_op) previous_losses.append(epoch_loss) # with tf.variable_scope("decoding") as scope
dtype = tf.float16 if FLAGS.use_fp16 else tf.float32 model = code_model.CodeModel( vocab_size,
random_line_split
corrector.py
_boolean("decode_while_training", False, "Do test decode while training.") tf.app.flags.DEFINE_boolean("decode_whole", False, "decode whole code or just a set.") FLAGS = tf.app.flags.FLAGS # We use a number of buckets and pad to the closest one for efficiency. # See seq2seq_model.Seq2SeqModel for details of how they work. # encoding 이 두개니까 버킷을 어떻게 잡을지 생각좀 해봐야 # _buckets = [(5, 10), (10, 15), (20, 25), (40, 50)] # _buckets = [(10, 15), (40, 50)] _buckets = [(50, 50)] def read_data(train_id_set, max_size=None): data_set = [[] for _ in _buckets] counter = 0 saved = 0 for code in train_id_set: for line_idx in xrange(len(code) - 2): if max_size and counter >= max_size: break counter += 1 source_ids = [code[line_idx], code[line_idx + 2]] target_ids = [code[line_idx + 1]] target_ids[0].append(code_loader.EOS_ID) for bucket_id, (source_size, target_size) in enumerate(_buckets): if len(source_ids[0]) < source_size and len(source_ids[1]) < source_size and len(target_ids[0]) < target_size: saved += 1 data_set[bucket_id].append([source_ids, target_ids]) break for line_idx in xrange(len(code) - 1): if max_size and counter >= max_size: break counter += 1 source_ids = [code[line_idx], code[line_idx + 1]] target_ids = [[]] target_ids[0].append(code_loader.EOS_ID) for bucket_id, (source_size, target_size) in enumerate(_buckets): if len(source_ids[0]) < source_size and len(source_ids[1]) < source_size and len(target_ids[0]) < target_size: saved += 1 data_set[bucket_id].append([source_ids, target_ids]) break print(" not saved: %d" % (counter - saved)) print(" read data line total %d" % counter) return data_set def create_model(session, forward_only, vocab_size=FLAGS.vocab_size): """Create translation model and initialize or load parameters in session.""" dtype = tf.float16 if FLAGS.use_fp16 else tf.float32 model = code_model.CodeModel( vocab_size, _buckets, FLAGS.size, FLAGS.num_layers, FLAGS.max_gradient_norm, FLAGS.batch_size, FLAGS.learning_rate, FLAGS.learning_rate_decay_factor, forward_only=forward_only, dtype=dtype) ckpt = tf.train.get_checkpoint_state(FLAGS.train_dir) if ckpt and tf.train.checkpoint_exists(ckpt.model_checkpoint_path): print("Reading model parameters from %s" % ckpt.model_checkpoint_path) model.saver.restore(session, ckpt.model_checkpoint_path) else: print("Created model with fresh parameters.") session.run(tf.global_variables_initializer()) return model def get_perplexity(loss): return math.exp(float(loss)) if loss < 300 else float("inf") def save_checkpoint(model, sess, vocab_size): # Save checkpoint and zero timer
with open(FLAGS.train_dir + "/" + os.path.basename(__file__) + ".ckpt.vb", mode="w") as vocab_size_file: pickle.dump(vocab_size, vocab_size_file) checkpoint_path = os.path.join(FLAGS.train_dir, FLAGS.data_path.split(".")[0] + ".ckpt.size." + str(FLAGS.size) + ".nl." + str(FLAGS.num_layers) + ".vs." + str(vocab_size)) model.saver.save(sess, checkpoint_path, global_step=model.global_step) def train(): train_id_data, id_to_vocab, vocab_to_id = code_loader.prepare_data(FLAGS.data_dir, FLAGS.vocab_size, data_path=FLAGS.data_path, cache=FLAGS.cache) vocab_size = FLAGS.vocab_size if FLAGS.vocab_size != 0 else len(id_to_vocab) # Create decode model # print("Creating %d layers of %d units for decode." % (FLAGS.num_layers, FLAGS.size)) # decode_sess = tf.Session() # decode_model = create_model(decode_sess, True, vocab_size=(FLAGS.vocab_size if FLAGS.vocab_size != 0 else len(id_to_vocab))) # Create model. with tf.Session() as sess: print("Creating %d layers of %d units with %d vocab." % (FLAGS.num_layers, FLAGS.size, vocab_size)) model = create_model(sess, False, vocab_size=vocab_size) # Read data into buckets and compute their sizes. print("Reading development and training data (limit: %d)." % FLAGS.max_train_data_size) # dev_set_path = FLAGS.train_dir + "/dev_set." + str(FLAGS.from_vocab_size) + "." + pickle.__name__ train_set_path = FLAGS.train_dir + "/" + FLAGS.data_path.split(".")[0] + ".train_set.ids" + str(FLAGS.vocab_size) + ".ds" + str(FLAGS.max_train_data_size) + "." + pickle.__name__ if not tf.gfile.Exists(train_set_path) or not FLAGS.cache: print("Reading training data (limit: %d)." % FLAGS.max_train_data_size) train_set = read_data(train_id_data, FLAGS.max_train_data_size) with tf.gfile.GFile(train_set_path, "w") as f: pickle.dump(train_set, f) else: print("Loading training data (limit: %d)." % FLAGS.max_train_data_size) with tf.gfile.GFile(train_set_path, mode="r") as f: train_set = pickle.load(f) train_bucket_sizes = [len(train_set[b]) for b in xrange(len(_buckets))] train_total_size = float(sum(train_bucket_sizes)) print("Total train %d" % train_total_size) # A bucket scale is a list of increasing numbers from 0 to 1 that we'll use # to select a bucket. Length of [scale[i], scale[i+1]] is proportional to # the size if i-th training bucket, as used later. train_buckets_scale = [sum(train_bucket_sizes[:i + 1]) / train_total_size for i in xrange(len(train_bucket_sizes))] # This is the training loop. print("Running the training loop") epoch_step_time, epoch_loss, ckpt_step_time, ckpt_loss = 0.0, 0.0, 0.0, 0.0 # current_step = 0 previous_losses = [] std_out = "Error: not enough steps to run with checkpoint_step." test_out = "" steps_per_epoch = round(int(train_total_size) / FLAGS.batch_size) steps_per_checkpoint = FLAGS.steps_per_checkpoint if FLAGS.steps_per_checkpoint != 0 else steps_per_epoch # batch * total data size epoch_step = model.global_step.eval() // steps_per_epoch while True: # Choose a bucket according to data distribution. We pick a random number # in [0, 1] and use the corresponding interval in train_buckets_scale. random_number_01 = np.random.random_sample() bucket_id = min([i for i in xrange(len(train_buckets_scale)) if train_buckets_scale[i] > random_number_01]) # Get a batch and make a step. start_time = time.time() encoder_inputs_front, encoder_inputs_back, decoder_inputs, target_weights = model.get_batch( train_set, bucket_id) _, step_loss, _ = model.step(sess, encoder_inputs_front, encoder_inputs_back, decoder_inputs, target_weights, bucket_id, False) ckpt_step_time += (time.time() - start_time) / steps_per_checkpoint ckpt_loss += step_loss / steps_per_checkpoint epoch_step_time += (time.time() - start_time) / steps_per_epoch epoch_loss += step_loss / steps_per_epoch # current_step += 1 # print condition if model.global_step.eval() % 1 == 0: print(" global step %d learning rate %.4f step-time %.2f loss %.2f perplexity " "%.2f" % (model.global_step.eval(), model.learning_rate.eval(), (time.time() - start_time), step_loss, get_perplexity(step_loss))) # per epoch if model.global_step.eval() % steps_per_epoch == 0: epoch_step += 1 print("epoch %d" % epoch_step) # Print statistics for the previous epoch. std_out = "epoch: global step %d learning rate %.4f step-time %.2f loss %.2f perplexity %.2f" % ( model.global_step.eval(), model.learning_rate.eval(), epoch_step_time, epoch_loss, get_perplexity(epoch_loss)) print(std_out) # Decrease learning rate if no improvement was seen over last 3 epoch times. if len(previous_losses) > 2 and epoch_loss > max(previous_losses[-3:]): sess.run(model.learning_rate_decay_op) previous_losses.append(epoch_loss) # with tf.variable_scope("decoding") as scope
and loss.
identifier_name
corrector.py
_step) def train(): train_id_data, id_to_vocab, vocab_to_id = code_loader.prepare_data(FLAGS.data_dir, FLAGS.vocab_size, data_path=FLAGS.data_path, cache=FLAGS.cache) vocab_size = FLAGS.vocab_size if FLAGS.vocab_size != 0 else len(id_to_vocab) # Create decode model # print("Creating %d layers of %d units for decode." % (FLAGS.num_layers, FLAGS.size)) # decode_sess = tf.Session() # decode_model = create_model(decode_sess, True, vocab_size=(FLAGS.vocab_size if FLAGS.vocab_size != 0 else len(id_to_vocab))) # Create model. with tf.Session() as sess: print("Creating %d layers of %d units with %d vocab." % (FLAGS.num_layers, FLAGS.size, vocab_size)) model = create_model(sess, False, vocab_size=vocab_size) # Read data into buckets and compute their sizes. print("Reading development and training data (limit: %d)." % FLAGS.max_train_data_size) # dev_set_path = FLAGS.train_dir + "/dev_set." + str(FLAGS.from_vocab_size) + "." + pickle.__name__ train_set_path = FLAGS.train_dir + "/" + FLAGS.data_path.split(".")[0] + ".train_set.ids" + str(FLAGS.vocab_size) + ".ds" + str(FLAGS.max_train_data_size) + "." + pickle.__name__ if not tf.gfile.Exists(train_set_path) or not FLAGS.cache: print("Reading training data (limit: %d)." % FLAGS.max_train_data_size) train_set = read_data(train_id_data, FLAGS.max_train_data_size) with tf.gfile.GFile(train_set_path, "w") as f: pickle.dump(train_set, f) else: print("Loading training data (limit: %d)." % FLAGS.max_train_data_size) with tf.gfile.GFile(train_set_path, mode="r") as f: train_set = pickle.load(f) train_bucket_sizes = [len(train_set[b]) for b in xrange(len(_buckets))] train_total_size = float(sum(train_bucket_sizes)) print("Total train %d" % train_total_size) # A bucket scale is a list of increasing numbers from 0 to 1 that we'll use # to select a bucket. Length of [scale[i], scale[i+1]] is proportional to # the size if i-th training bucket, as used later. train_buckets_scale = [sum(train_bucket_sizes[:i + 1]) / train_total_size for i in xrange(len(train_bucket_sizes))] # This is the training loop. print("Running the training loop") epoch_step_time, epoch_loss, ckpt_step_time, ckpt_loss = 0.0, 0.0, 0.0, 0.0 # current_step = 0 previous_losses = [] std_out = "Error: not enough steps to run with checkpoint_step." test_out = "" steps_per_epoch = round(int(train_total_size) / FLAGS.batch_size) steps_per_checkpoint = FLAGS.steps_per_checkpoint if FLAGS.steps_per_checkpoint != 0 else steps_per_epoch # batch * total data size epoch_step = model.global_step.eval() // steps_per_epoch while True: # Choose a bucket according to data distribution. We pick a random number # in [0, 1] and use the corresponding interval in train_buckets_scale. random_number_01 = np.random.random_sample() bucket_id = min([i for i in xrange(len(train_buckets_scale)) if train_buckets_scale[i] > random_number_01]) # Get a batch and make a step. start_time = time.time() encoder_inputs_front, encoder_inputs_back, decoder_inputs, target_weights = model.get_batch( train_set, bucket_id) _, step_loss, _ = model.step(sess, encoder_inputs_front, encoder_inputs_back, decoder_inputs, target_weights, bucket_id, False) ckpt_step_time += (time.time() - start_time) / steps_per_checkpoint ckpt_loss += step_loss / steps_per_checkpoint epoch_step_time += (time.time() - start_time) / steps_per_epoch epoch_loss += step_loss / steps_per_epoch # current_step += 1 # print condition if model.global_step.eval() % 1 == 0: print(" global step %d learning rate %.4f step-time %.2f loss %.2f perplexity " "%.2f" % (model.global_step.eval(), model.learning_rate.eval(), (time.time() - start_time), step_loss, get_perplexity(step_loss))) # per epoch if model.global_step.eval() % steps_per_epoch == 0: epoch_step += 1 print("epoch %d" % epoch_step) # Print statistics for the previous epoch. std_out = "epoch: global step %d learning rate %.4f step-time %.2f loss %.2f perplexity %.2f" % ( model.global_step.eval(), model.learning_rate.eval(), epoch_step_time, epoch_loss, get_perplexity(epoch_loss)) print(std_out) # Decrease learning rate if no improvement was seen over last 3 epoch times. if len(previous_losses) > 2 and epoch_loss > max(previous_losses[-3:]): sess.run(model.learning_rate_decay_op) previous_losses.append(epoch_loss) # with tf.variable_scope("decoding") as scope: # decode() epoch_step_time, epoch_loss = 0.0, 0.0 # escape if all epoch is done if FLAGS.epoch != 0 and epoch_step >= FLAGS.epoch: # Save checkpoint. save_checkpoint(model, sess, vocab_size) with open(FLAGS.train_dir + "/result.ids" + str(FLAGS.vocab_size) + ".ds" + str(FLAGS.max_train_data_size), "a") as output_file: output_file.write( datetime.now().strftime("%Y-%m-%d-%H:%M:%S") + ": " + std_out + " // Tag: " + FLAGS.out_tag + "\n" + "\ttest_out: " + test_out + "\n" ) break # per checkpoint jobs if model.global_step.eval() % steps_per_checkpoint == 0: # Print statistics for the previous checkpoint. std_out = "checkpoint: global step %d learning rate %.4f step-time %.2f loss %.2f perplexity %.2f" % ( model.global_step.eval(), model.learning_rate.eval(), ckpt_step_time, ckpt_loss, get_perplexity(ckpt_loss)) print(std_out) # Save checkpoint and zero timer and loss. save_checkpoint(model, sess, vocab_size) ckpt_step_time, ckpt_loss = 0.0, 0.0 def decode_with_session_and_model(sess, model): # Load vocabularies. vocab_path = os.path.join(FLAGS.data_dir, FLAGS.data_path.split(".")[0] + ".vocab%d.%s" % (FLAGS.vocab_size, pickle.__name__)) with gfile.GFile(vocab_path, mode="r") as vocab_file: id_to_vocab, vocab_to_id, vocab_freq = pickle.load(vocab_file) source = code_loader._START_LINE + '\n' + 'print(a+b)' # source = "b = 2\n" + code_loader._END_LINE data = [{ "source": source }] data = source_filter.filter_danger(data) source_filter.remove_redundent_newlines_and_set_line_length(data) data = source_filter.set_token(data) source_data = code_loader.data_to_tokens_list(data) model.batch_size = 1 if not FLAGS.decode_whole else len(source_data[0]) * 2 + 1 # We decode one sentence at a time. id_data = [] for src in source_data: id_source = [[code_loader.START_LINE_ID]] for line in src: id_line = [vocab_to_id.get(word[1], code_loader.UNK_ID) for word in line] id_source.append(id_line) id_source.append([code_loader.END_LINE_ID]) id_data.append(id_source) bucket_id = -1 data_set = [[] for _ in _buckets] for code in id_data: if FLAGS.decode_whole: for line_idx in xrange(len(code) - 2): source_ids = [code[line_idx], code[line_idx + 2]] target_ids = [code[line_idx + 1]] target_ids[0].append(code_loader.EOS_ID) for idx, (source_size, target_size) in enumerate(_buckets): if len(source_ids[0]) < source_size and len(source_ids[1]) < source_size: data_set[idx].append([source_ids, target_ids]) bucket_id = idx break for line_idx in xrange(len(code) - 1): source_ids = [code[line_idx], code[line_idx + 1]] target_ids =
[[]] target_ids[0].append(code_loader.EOS_ID) for idx, (source_size, target_size) in enumerate(_buckets): if len(source_ids[0]) < source_size and len(source_ids[1]) < source_size: data_set[idx].append([source_ids, target_ids]) bucket_id = idx break else: source_ids = [code[1], code[2]] target_
conditional_block
trajectory_multiple.py
_shape) # src_pt = np.float32(table) # dst_pt = np.float32([(0,0), (dst_shape[0],0), dst_shape, (0, dst_shape[1])]) # warp_matrix = cv.getPerspectiveTransform(src_pt, dst_pt) # img = cv.warpPerspective(img, warp_matrix, dst_shape) # 4. Correct known coordinates with the warp matrix. def transform_point(pt, matrix): is_list = type(pt) is list if not is_list: pt = [[pt]] else: pt = [[p] for p in pt] array = np.array(pt) transformed = cv.transform(array, matrix) squeezed = np.squeeze(np.squeeze(transformed)) if not is_list: return tuple(squeezed[:2]) return [tuple(x[:2]) for x in squeezed] # table = transform_point(table, warp_matrix) table = [(0, 0), (1710, 0), (1710, 768), (0, 768)] cue_ball = transform_point(cue_ball, warp_matrix) cue_start = transform_point(cue_start, warp_matrix) cue_tip = transform_point(cue_tip, warp_matrix) balls = transform_point(balls, warp_matrix) pockets = table.copy() x_center = int(img.shape[1] / 2) pockets.append((x_center, 0)) pockets.append((x_center, img.shape[0])) def update(): global img, table, cue_tip, cue_start, cue_ball, cue_ball_radius hit_balls = [] shown_image = img.copy() def points_to_angle(point1, point2): return math.atan2(point2[1] - point1[1], point2[0] - point1[0]) # 5. Draw the trajectory # 5.1 Get the cue angle based on 2 known points. # - To get the angle in Radian measure use the atan2(y1-y2, x1-x2) function. cue_angle = points_to_angle(cue_start, cue_tip) # 5.2 Check if the cue angle overlaps with the location of the cue ball. # - Check if the distance of the cue ball to the line is not more than the radius of the cue ball. # - Distance is using: abs(ax - by + c) / sqrt(aa + bb) # - a = y2-y1 # - b = x2-x1 # - c = x2y1 + y2x1 # - x, y = point of cue ball center. # - Extra check if the cue ball is not behind the cue tip. # http://www.pygame.org/wiki/IntersectingLineDetection def line_intersect(line1, line2): def gradient(points): if points[0][0] == points[1][0]: return None return (points[0][1] - points[1][1]) / (points[0][0] - points[1][0]) def y_intersect(p, m): return p[1] - (m * p[0]) m1, m2 = gradient(line1), gradient(line2) if m1 == m2: return None elif m1 is not None and m2 is not None: b1 = y_intersect(line1[0], m1) b2 = y_intersect(line2[0], m2) x = (b2 - b1) / (m1 - m2) pt = x, (m1 * x) + b1 elif m1 is None: b2 = y_intersect(line2[0], m2) pt = line2[0][0], (m2 * line1[0][0]) + b2 else: b1 = y_intersect(line1[0], m1) pt = line2[0][0], (m1 * line2[0][0]) + b1 return tuple(int(x) for x in pt) def line_circle_collision(pt1, pt2, center, circle_radius): global img # Point opposite of circle if (min(pt2[0], img.shape[1]) - pt1[0]) < 0 == (max(pt2[0], 0) - center[0] < 0) or (pt2[1] - pt1[1]) < 0 == ( pt2[1] - center[1]) < 0: return False a = (pt2[1] - pt1[1]) b = (pt2[0] - pt1[0]) c = (pt2[0] * pt1[1]) - (pt2[1] * pt1[0]) x, y = center dist = abs(a * x - b * y + c) / math.sqrt(a * a + b * b) if circle_radius >= dist: return True else: return False # https://stackoverflow.com/questions/29384494/the-intersection-between-a-trajectory-and-the-circles-in-the-same-area def line_circle_intersection(pt1, pt2, center, circle_radius): x1, y1 = [int(x) for x in pt1] x2, y2 = [int(x) for x in pt2] xc, yc = [int(x) for x in center] r = circle_radius dx = x1 - x2 dy = y1 - y2 rx = xc - x1 ry = yc - y1 a = dx * dx + dy * dy b = dx * rx + dy * ry c = rx * rx + ry * ry - r * r # Now solve a*t^2 + 2*b*t + c = 0 d = b * b - a * c if d < 0.: # no real intersection return s = math.sqrt(d) t1 = (- b - s) / a t2 = (- b + s) / a points = [] if 0. <= t1 <= 1.: points.append(tuple([round((1 - t1) * x1 + t1 * x2), round((1 - t1) * y1 + t1 * y2)])) if 0. <= t2 <= 1.: points.append(tuple([round((1 - t2) * x1 + t2 * x2), round((1 - t2) * y1 + t2 * y2)])) return points def invert_angle(angle): return (angle + math.pi) % (2 * math.pi) if line_circle_collision(cue_start, cue_tip, cue_ball, cue_ball_radius): # 5.3 Get the angle of the cue ball trajectory. trj_angle = cue_angle start_point = cue_ball collisions = 1 while collisions <= 5: collisions += 1 # 5.4 Use the angle, center and radius of the cue ball to calculate at which point the line starts. # - The point is: x = (x1 + r + cos(radians)), y = (y1 + r + sin(radians)) end_point = (int(start_point[0] + 2000 * np.cos(trj_angle)), int(start_point[1] + 2000 * np.sin(trj_angle))) # 5.5 Draw the trajectory. # - When the edge of the image is released then continue on a new angle or stop after 5 collision. line = np.array([start_point, end_point]) # Filter out balls that are possible to hit selected_balls = [] for i in range(0, len(balls)): if i not in hit_balls: selected_balls.append(balls[i]) # Sort the balls based on distance def
(pt1, pt2): return math.sqrt(math.pow(pt2[0]-pt1[0], 2)+math.pow(pt2[1]-pt1[1], 2)) def point_by_angle(pt, angle, distance): x = pt[0] + (distance * math.cos(angle)) y = pt[1] + (distance * math.sin(angle)) return tuple([round(x), round(y)]) selected_balls.sort(key=lambda ball: point_distance(start_point, ball)) ball_hit = False for ball in selected_balls: if ball in hit_balls: continue if line_circle_collision(start_point, end_point, ball, cue_ball_radius*2): points = line_circle_intersection(start_point, end_point, ball, cue_ball_radius*2) if len(points) <= 0 or start_point == points[0]: continue end_point = points[0] cv.circle(shown_image, end_point, cue_ball_radius, (0, 255,255), thickness=2) ball_hit = True trj_angle = invert_angle(points_to_angle(ball, end_point)) cv.line(shown_image, end_point, point
point_distance
identifier_name
trajectory_multiple.py
_shape) # src_pt = np.float32(table) # dst_pt = np.float32([(0,0), (dst_shape[0],0), dst_shape, (0, dst_shape[1])]) # warp_matrix = cv.getPerspectiveTransform(src_pt, dst_pt) # img = cv.warpPerspective(img, warp_matrix, dst_shape) # 4. Correct known coordinates with the warp matrix. def transform_point(pt, matrix): is_list = type(pt) is list if not is_list: pt = [[pt]] else: pt = [[p] for p in pt] array = np.array(pt) transformed = cv.transform(array, matrix) squeezed = np.squeeze(np.squeeze(transformed)) if not is_list:
return [tuple(x[:2]) for x in squeezed] # table = transform_point(table, warp_matrix) table = [(0, 0), (1710, 0), (1710, 768), (0, 768)] cue_ball = transform_point(cue_ball, warp_matrix) cue_start = transform_point(cue_start, warp_matrix) cue_tip = transform_point(cue_tip, warp_matrix) balls = transform_point(balls, warp_matrix) pockets = table.copy() x_center = int(img.shape[1] / 2) pockets.append((x_center, 0)) pockets.append((x_center, img.shape[0])) def update(): global img, table, cue_tip, cue_start, cue_ball, cue_ball_radius hit_balls = [] shown_image = img.copy() def points_to_angle(point1, point2): return math.atan2(point2[1] - point1[1], point2[0] - point1[0]) # 5. Draw the trajectory # 5.1 Get the cue angle based on 2 known points. # - To get the angle in Radian measure use the atan2(y1-y2, x1-x2) function. cue_angle = points_to_angle(cue_start, cue_tip) # 5.2 Check if the cue angle overlaps with the location of the cue ball. # - Check if the distance of the cue ball to the line is not more than the radius of the cue ball. # - Distance is using: abs(ax - by + c) / sqrt(aa + bb) # - a = y2-y1 # - b = x2-x1 # - c = x2y1 + y2x1 # - x, y = point of cue ball center. # - Extra check if the cue ball is not behind the cue tip. # http://www.pygame.org/wiki/IntersectingLineDetection def line_intersect(line1, line2): def gradient(points): if points[0][0] == points[1][0]: return None return (points[0][1] - points[1][1]) / (points[0][0] - points[1][0]) def y_intersect(p, m): return p[1] - (m * p[0]) m1, m2 = gradient(line1), gradient(line2) if m1 == m2: return None elif m1 is not None and m2 is not None: b1 = y_intersect(line1[0], m1) b2 = y_intersect(line2[0], m2) x = (b2 - b1) / (m1 - m2) pt = x, (m1 * x) + b1 elif m1 is None: b2 = y_intersect(line2[0], m2) pt = line2[0][0], (m2 * line1[0][0]) + b2 else: b1 = y_intersect(line1[0], m1) pt = line2[0][0], (m1 * line2[0][0]) + b1 return tuple(int(x) for x in pt) def line_circle_collision(pt1, pt2, center, circle_radius): global img # Point opposite of circle if (min(pt2[0], img.shape[1]) - pt1[0]) < 0 == (max(pt2[0], 0) - center[0] < 0) or (pt2[1] - pt1[1]) < 0 == ( pt2[1] - center[1]) < 0: return False a = (pt2[1] - pt1[1]) b = (pt2[0] - pt1[0]) c = (pt2[0] * pt1[1]) - (pt2[1] * pt1[0]) x, y = center dist = abs(a * x - b * y + c) / math.sqrt(a * a + b * b) if circle_radius >= dist: return True else: return False # https://stackoverflow.com/questions/29384494/the-intersection-between-a-trajectory-and-the-circles-in-the-same-area def line_circle_intersection(pt1, pt2, center, circle_radius): x1, y1 = [int(x) for x in pt1] x2, y2 = [int(x) for x in pt2] xc, yc = [int(x) for x in center] r = circle_radius dx = x1 - x2 dy = y1 - y2 rx = xc - x1 ry = yc - y1 a = dx * dx + dy * dy b = dx * rx + dy * ry c = rx * rx + ry * ry - r * r # Now solve a*t^2 + 2*b*t + c = 0 d = b * b - a * c if d < 0.: # no real intersection return s = math.sqrt(d) t1 = (- b - s) / a t2 = (- b + s) / a points = [] if 0. <= t1 <= 1.: points.append(tuple([round((1 - t1) * x1 + t1 * x2), round((1 - t1) * y1 + t1 * y2)])) if 0. <= t2 <= 1.: points.append(tuple([round((1 - t2) * x1 + t2 * x2), round((1 - t2) * y1 + t2 * y2)])) return points def invert_angle(angle): return (angle + math.pi) % (2 * math.pi) if line_circle_collision(cue_start, cue_tip, cue_ball, cue_ball_radius): # 5.3 Get the angle of the cue ball trajectory. trj_angle = cue_angle start_point = cue_ball collisions = 1 while collisions <= 5: collisions += 1 # 5.4 Use the angle, center and radius of the cue ball to calculate at which point the line starts. # - The point is: x = (x1 + r + cos(radians)), y = (y1 + r + sin(radians)) end_point = (int(start_point[0] + 2000 * np.cos(trj_angle)), int(start_point[1] + 2000 * np.sin(trj_angle))) # 5.5 Draw the trajectory. # - When the edge of the image is released then continue on a new angle or stop after 5 collision. line = np.array([start_point, end_point]) # Filter out balls that are possible to hit selected_balls = [] for i in range(0, len(balls)): if i not in hit_balls: selected_balls.append(balls[i]) # Sort the balls based on distance def point_distance(pt1, pt2): return math.sqrt(math.pow(pt2[0]-pt1[0], 2)+math.pow(pt2[1]-pt1[1], 2)) def point_by_angle(pt, angle, distance): x = pt[0] + (distance * math.cos(angle)) y = pt[1] + (distance * math.sin(angle)) return tuple([round(x), round(y)]) selected_balls.sort(key=lambda ball: point_distance(start_point, ball)) ball_hit = False for ball in selected_balls: if ball in hit_balls: continue if line_circle_collision(start_point, end_point, ball, cue_ball_radius*2): points = line_circle_intersection(start_point, end_point, ball, cue_ball_radius*2) if len(points) <= 0 or start_point == points[0]: continue end_point = points[0] cv.circle(shown_image, end_point, cue_ball_radius, (0, 255,255), thickness=2) ball_hit = True trj_angle = invert_angle(points_to_angle(ball, end_point)) cv.line(shown_image, end_point, point
return tuple(squeezed[:2])
conditional_block
trajectory_multiple.py
_shape) # src_pt = np.float32(table) # dst_pt = np.float32([(0,0), (dst_shape[0],0), dst_shape, (0, dst_shape[1])]) # warp_matrix = cv.getPerspectiveTransform(src_pt, dst_pt) # img = cv.warpPerspective(img, warp_matrix, dst_shape) # 4. Correct known coordinates with the warp matrix. def transform_point(pt, matrix): is_list = type(pt) is list if not is_list: pt = [[pt]] else: pt = [[p] for p in pt] array = np.array(pt) transformed = cv.transform(array, matrix) squeezed = np.squeeze(np.squeeze(transformed)) if not is_list: return tuple(squeezed[:2]) return [tuple(x[:2]) for x in squeezed] # table = transform_point(table, warp_matrix) table = [(0, 0), (1710, 0), (1710, 768), (0, 768)] cue_ball = transform_point(cue_ball, warp_matrix) cue_start = transform_point(cue_start, warp_matrix) cue_tip = transform_point(cue_tip, warp_matrix) balls = transform_point(balls, warp_matrix) pockets = table.copy() x_center = int(img.shape[1] / 2) pockets.append((x_center, 0)) pockets.append((x_center, img.shape[0]))
def update(): global img, table, cue_tip, cue_start, cue_ball, cue_ball_radius hit_balls = [] shown_image = img.copy() def points_to_angle(point1, point2): return math.atan2(point2[1] - point1[1], point2[0] - point1[0]) # 5. Draw the trajectory # 5.1 Get the cue angle based on 2 known points. # - To get the angle in Radian measure use the atan2(y1-y2, x1-x2) function. cue_angle = points_to_angle(cue_start, cue_tip) # 5.2 Check if the cue angle overlaps with the location of the cue ball. # - Check if the distance of the cue ball to the line is not more than the radius of the cue ball. # - Distance is using: abs(ax - by + c) / sqrt(aa + bb) # - a = y2-y1 # - b = x2-x1 # - c = x2y1 + y2x1 # - x, y = point of cue ball center. # - Extra check if the cue ball is not behind the cue tip. # http://www.pygame.org/wiki/IntersectingLineDetection def line_intersect(line1, line2): def gradient(points): if points[0][0] == points[1][0]: return None return (points[0][1] - points[1][1]) / (points[0][0] - points[1][0]) def y_intersect(p, m): return p[1] - (m * p[0]) m1, m2 = gradient(line1), gradient(line2) if m1 == m2: return None elif m1 is not None and m2 is not None: b1 = y_intersect(line1[0], m1) b2 = y_intersect(line2[0], m2) x = (b2 - b1) / (m1 - m2) pt = x, (m1 * x) + b1 elif m1 is None: b2 = y_intersect(line2[0], m2) pt = line2[0][0], (m2 * line1[0][0]) + b2 else: b1 = y_intersect(line1[0], m1) pt = line2[0][0], (m1 * line2[0][0]) + b1 return tuple(int(x) for x in pt) def line_circle_collision(pt1, pt2, center, circle_radius): global img # Point opposite of circle if (min(pt2[0], img.shape[1]) - pt1[0]) < 0 == (max(pt2[0], 0) - center[0] < 0) or (pt2[1] - pt1[1]) < 0 == ( pt2[1] - center[1]) < 0: return False a = (pt2[1] - pt1[1]) b = (pt2[0] - pt1[0]) c = (pt2[0] * pt1[1]) - (pt2[1] * pt1[0]) x, y = center dist = abs(a * x - b * y + c) / math.sqrt(a * a + b * b) if circle_radius >= dist: return True else: return False # https://stackoverflow.com/questions/29384494/the-intersection-between-a-trajectory-and-the-circles-in-the-same-area def line_circle_intersection(pt1, pt2, center, circle_radius): x1, y1 = [int(x) for x in pt1] x2, y2 = [int(x) for x in pt2] xc, yc = [int(x) for x in center] r = circle_radius dx = x1 - x2 dy = y1 - y2 rx = xc - x1 ry = yc - y1 a = dx * dx + dy * dy b = dx * rx + dy * ry c = rx * rx + ry * ry - r * r # Now solve a*t^2 + 2*b*t + c = 0 d = b * b - a * c if d < 0.: # no real intersection return s = math.sqrt(d) t1 = (- b - s) / a t2 = (- b + s) / a points = [] if 0. <= t1 <= 1.: points.append(tuple([round((1 - t1) * x1 + t1 * x2), round((1 - t1) * y1 + t1 * y2)])) if 0. <= t2 <= 1.: points.append(tuple([round((1 - t2) * x1 + t2 * x2), round((1 - t2) * y1 + t2 * y2)])) return points def invert_angle(angle): return (angle + math.pi) % (2 * math.pi) if line_circle_collision(cue_start, cue_tip, cue_ball, cue_ball_radius): # 5.3 Get the angle of the cue ball trajectory. trj_angle = cue_angle start_point = cue_ball collisions = 1 while collisions <= 5: collisions += 1 # 5.4 Use the angle, center and radius of the cue ball to calculate at which point the line starts. # - The point is: x = (x1 + r + cos(radians)), y = (y1 + r + sin(radians)) end_point = (int(start_point[0] + 2000 * np.cos(trj_angle)), int(start_point[1] + 2000 * np.sin(trj_angle))) # 5.5 Draw the trajectory. # - When the edge of the image is released then continue on a new angle or stop after 5 collision. line = np.array([start_point, end_point]) # Filter out balls that are possible to hit selected_balls = [] for i in range(0, len(balls)): if i not in hit_balls: selected_balls.append(balls[i]) # Sort the balls based on distance def point_distance(pt1, pt2): return math.sqrt(math.pow(pt2[0]-pt1[0], 2)+math.pow(pt2[1]-pt1[1], 2)) def point_by_angle(pt, angle, distance): x = pt[0] + (distance * math.cos(angle)) y = pt[1] + (distance * math.sin(angle)) return tuple([round(x), round(y)]) selected_balls.sort(key=lambda ball: point_distance(start_point, ball)) ball_hit = False for ball in selected_balls: if ball in hit_balls: continue if line_circle_collision(start_point, end_point, ball, cue_ball_radius*2): points = line_circle_intersection(start_point, end_point, ball, cue_ball_radius*2) if len(points) <= 0 or start_point == points[0]: continue end_point = points[0] cv.circle(shown_image, end_point, cue_ball_radius, (0, 255,255), thickness=2) ball_hit = True trj_angle = invert_angle(points_to_angle(ball, end_point)) cv.line(shown_image, end_point, point
random_line_split
trajectory_multiple.py
_shape) # src_pt = np.float32(table) # dst_pt = np.float32([(0,0), (dst_shape[0],0), dst_shape, (0, dst_shape[1])]) # warp_matrix = cv.getPerspectiveTransform(src_pt, dst_pt) # img = cv.warpPerspective(img, warp_matrix, dst_shape) # 4. Correct known coordinates with the warp matrix. def transform_point(pt, matrix): is_list = type(pt) is list if not is_list: pt = [[pt]] else: pt = [[p] for p in pt] array = np.array(pt) transformed = cv.transform(array, matrix) squeezed = np.squeeze(np.squeeze(transformed)) if not is_list: return tuple(squeezed[:2]) return [tuple(x[:2]) for x in squeezed] # table = transform_point(table, warp_matrix) table = [(0, 0), (1710, 0), (1710, 768), (0, 768)] cue_ball = transform_point(cue_ball, warp_matrix) cue_start = transform_point(cue_start, warp_matrix) cue_tip = transform_point(cue_tip, warp_matrix) balls = transform_point(balls, warp_matrix) pockets = table.copy() x_center = int(img.shape[1] / 2) pockets.append((x_center, 0)) pockets.append((x_center, img.shape[0])) def update(): global img, table, cue_tip, cue_start, cue_ball, cue_ball_radius hit_balls = [] shown_image = img.copy() def points_to_angle(point1, point2): return math.atan2(point2[1] - point1[1], point2[0] - point1[0]) # 5. Draw the trajectory # 5.1 Get the cue angle based on 2 known points. # - To get the angle in Radian measure use the atan2(y1-y2, x1-x2) function. cue_angle = points_to_angle(cue_start, cue_tip) # 5.2 Check if the cue angle overlaps with the location of the cue ball. # - Check if the distance of the cue ball to the line is not more than the radius of the cue ball. # - Distance is using: abs(ax - by + c) / sqrt(aa + bb) # - a = y2-y1 # - b = x2-x1 # - c = x2y1 + y2x1 # - x, y = point of cue ball center. # - Extra check if the cue ball is not behind the cue tip. # http://www.pygame.org/wiki/IntersectingLineDetection def line_intersect(line1, line2): def gradient(points): if points[0][0] == points[1][0]: return None return (points[0][1] - points[1][1]) / (points[0][0] - points[1][0]) def y_intersect(p, m): return p[1] - (m * p[0]) m1, m2 = gradient(line1), gradient(line2) if m1 == m2: return None elif m1 is not None and m2 is not None: b1 = y_intersect(line1[0], m1) b2 = y_intersect(line2[0], m2) x = (b2 - b1) / (m1 - m2) pt = x, (m1 * x) + b1 elif m1 is None: b2 = y_intersect(line2[0], m2) pt = line2[0][0], (m2 * line1[0][0]) + b2 else: b1 = y_intersect(line1[0], m1) pt = line2[0][0], (m1 * line2[0][0]) + b1 return tuple(int(x) for x in pt) def line_circle_collision(pt1, pt2, center, circle_radius): global img # Point opposite of circle if (min(pt2[0], img.shape[1]) - pt1[0]) < 0 == (max(pt2[0], 0) - center[0] < 0) or (pt2[1] - pt1[1]) < 0 == ( pt2[1] - center[1]) < 0: return False a = (pt2[1] - pt1[1]) b = (pt2[0] - pt1[0]) c = (pt2[0] * pt1[1]) - (pt2[1] * pt1[0]) x, y = center dist = abs(a * x - b * y + c) / math.sqrt(a * a + b * b) if circle_radius >= dist: return True else: return False # https://stackoverflow.com/questions/29384494/the-intersection-between-a-trajectory-and-the-circles-in-the-same-area def line_circle_intersection(pt1, pt2, center, circle_radius): x1, y1 = [int(x) for x in pt1] x2, y2 = [int(x) for x in pt2] xc, yc = [int(x) for x in center] r = circle_radius dx = x1 - x2 dy = y1 - y2 rx = xc - x1 ry = yc - y1 a = dx * dx + dy * dy b = dx * rx + dy * ry c = rx * rx + ry * ry - r * r # Now solve a*t^2 + 2*b*t + c = 0 d = b * b - a * c if d < 0.: # no real intersection return s = math.sqrt(d) t1 = (- b - s) / a t2 = (- b + s) / a points = [] if 0. <= t1 <= 1.: points.append(tuple([round((1 - t1) * x1 + t1 * x2), round((1 - t1) * y1 + t1 * y2)])) if 0. <= t2 <= 1.: points.append(tuple([round((1 - t2) * x1 + t2 * x2), round((1 - t2) * y1 + t2 * y2)])) return points def invert_angle(angle):
if line_circle_collision(cue_start, cue_tip, cue_ball, cue_ball_radius): # 5.3 Get the angle of the cue ball trajectory. trj_angle = cue_angle start_point = cue_ball collisions = 1 while collisions <= 5: collisions += 1 # 5.4 Use the angle, center and radius of the cue ball to calculate at which point the line starts. # - The point is: x = (x1 + r + cos(radians)), y = (y1 + r + sin(radians)) end_point = (int(start_point[0] + 2000 * np.cos(trj_angle)), int(start_point[1] + 2000 * np.sin(trj_angle))) # 5.5 Draw the trajectory. # - When the edge of the image is released then continue on a new angle or stop after 5 collision. line = np.array([start_point, end_point]) # Filter out balls that are possible to hit selected_balls = [] for i in range(0, len(balls)): if i not in hit_balls: selected_balls.append(balls[i]) # Sort the balls based on distance def point_distance(pt1, pt2): return math.sqrt(math.pow(pt2[0]-pt1[0], 2)+math.pow(pt2[1]-pt1[1], 2)) def point_by_angle(pt, angle, distance): x = pt[0] + (distance * math.cos(angle)) y = pt[1] + (distance * math.sin(angle)) return tuple([round(x), round(y)]) selected_balls.sort(key=lambda ball: point_distance(start_point, ball)) ball_hit = False for ball in selected_balls: if ball in hit_balls: continue if line_circle_collision(start_point, end_point, ball, cue_ball_radius*2): points = line_circle_intersection(start_point, end_point, ball, cue_ball_radius*2) if len(points) <= 0 or start_point == points[0]: continue end_point = points[0] cv.circle(shown_image, end_point, cue_ball_radius, (0, 255,255), thickness=2) ball_hit = True trj_angle = invert_angle(points_to_angle(ball, end_point)) cv.line(shown_image, end_point, point
return (angle + math.pi) % (2 * math.pi)
identifier_body
lib.rs
([127, 0, 0, 1], 8000).into(); //! let svc = MitmProxyService::<MyMitm>::new(); //! let server = Server::bind(&addr) //! .serve(svc) //! .map_err(|e| eprintln!("server error: {}", e)); //! println!("noop mitm proxy listening on http://{}", addr); //! hyper::rt::run(server); //! } //! ``` //! //! Other examples can be found at //! <https://github.com/nlevitt/monie/tree/master/examples>. #![deny(warnings)] #![deny(missing_docs)] #![deny(missing_debug_implementations)] #[macro_use] extern crate lazy_static; #[macro_use] extern crate log; pub mod certauth; use std::error::Error; use std::sync::Arc; use bytes::Bytes; use futures::future::{self, Future, FutureResult}; use futures::stream::Stream; use http::method::Method; use http::uri::{Authority, Uri}; use hyper::client::pool::Pooled; use hyper::client::{HttpConnector, PoolClient}; use hyper::server::conn::Http; use hyper::service::{service_fn, NewService, Service}; use hyper::upgrade::Upgraded; use hyper::{Body, Chunk, Client, Request, Response}; use hyper_rustls::HttpsConnector; use tokio_rustls::{Accept, TlsAcceptor, TlsStream}; /// Represents the interception of a single request. Users of the library must /// implement this trait. With it you can observe and manipulate the request /// and response payload and headers. pub trait Mitm { /// Create a new instance of this `Mitm` implementation. The argument `uri` /// is the uri being proxied. Implementations may do with this what they /// wish (log it, stash it, ignore it, etc). fn new(uri: Uri) -> Self; /// Observe and manipulate the request headers. The `req` argument contains /// the original request headers received from the proxy client. The /// request headers returned by this function are sent to the remote /// server. fn request_headers(&self, req: Request<Body>) -> Request<Body>; /// Observe and manipulate a chunk of the request payload. This function /// may be called zero or more times, depending on the size of the request /// payload. It will not be called at all in the common case of a GET /// request with no payload. The `chunk` argument contains an original /// chunk of the request payload as received from the proxy client. The /// return value of this function is sent to the remote server. fn request_body_chunk(&self, chunk: Chunk) -> Chunk; /// Observe and manipulate the response headers. The `res` argument /// contains the original response headers received from the remote server. /// The response headers returned by this function are sent to the proxy /// client. fn response_headers(&self, res: Response<Body>) -> Response<Body>; /// Observe and manipulate a chunk of the response payload. This function /// may be called zero or more times, depending on the size of the payload. /// The `chunk` argument represents an unaltered chunk of the response /// payload as received from the remote server. The return value of this /// function is sent to the remote server. fn response_body_chunk(&self, chunk: Chunk) -> Chunk; } lazy_static! { static ref CLIENT: Client<HttpsConnector<HttpConnector>, Body> = Client::builder().build(HttpsConnector::new(4)); static ref HTTP: Http = Http::new(); } /// The `hyper::service::Service` that does the proxying and calls your `Mitm` /// implementation. #[derive(Debug)] pub struct MitmProxyService<T: Mitm + Sync> { _phantom: std::marker::PhantomData<T>, } impl<T: Mitm + Sync> MitmProxyService<T> { /// Creates a new `MitmProxyService`. #[inline] pub fn new() -> Self
} impl<T: Mitm + Sync + Send + 'static> NewService for MitmProxyService<T> { type ReqBody = Body; type ResBody = Body; type Error = std::io::Error; type Service = MitmProxyService<T>; type InitError = std::io::Error; type Future = FutureResult<Self::Service, Self::InitError>; fn new_service(&self) -> Self::Future { future::ok(MitmProxyService::new()) } } impl<T: Mitm + Sync + Send + 'static> Service for MitmProxyService<T> { type ReqBody = Body; type ResBody = Body; type Error = std::io::Error; type Future = Box<dyn Future<Item = Response<Body>, Error = std::io::Error> + Send>; fn call(&mut self, req: Request<Body>) -> Self::Future { info!("MitmProxyService::call() handling {:?}", req); if *req.method() == Method::CONNECT { Box::new(proxy_connect::<T>(req)) } else { Box::new(proxy_request::<T>(req)) } } } /// Obtains a connection to the scheme://authority of `uri` from the connection /// pool. fn obtain_connection( uri: Uri, ) -> impl Future<Item = Pooled<PoolClient<Body>>, Error = std::io::Error> { let key1 = Arc::new(format!( "{}://{}", uri.scheme_part().unwrap(), uri.authority_part().unwrap() )); let key2 = Arc::clone(&key1); let result = CLIENT.connection_for(uri, key1).map_err(move |e| { std::io::Error::new( std::io::ErrorKind::Other, format!("error obtaining connection to {}: {:?}", key2, e), ) }); result } /// Obtains a connection to the remote server and proxies the request, calling /// the `Mitm` implementation functions, which may manipulate the request and /// reponse. Returns a future that resolves to the response or error. /// /// This function is called for plain http requests, and for https requests /// received "inside" the fake, tapped `CONNECT` tunnel. fn proxy_request<T: Mitm + Sync + Send + 'static>( req: Request<Body>, ) -> impl Future<Item = Response<Body>, Error = std::io::Error> { obtain_connection(req.uri().to_owned()) .map(|mut connection| { let mitm1 = Arc::new(T::new(req.uri().to_owned())); let mitm2 = Arc::clone(&mitm1); let req = mitm1.request_headers(req); let (parts, body) = req.into_parts(); let body = Body::wrap_stream( body.map(move |chunk| mitm1.request_body_chunk(chunk)), ); let req = Request::from_parts(parts, body); info!("proxy_request() sending request {:?}", req); connection .send_request_retryable(req) .map(|response| { let response = mitm2.response_headers(response); let (parts, body) = response.into_parts(); let body = Body::wrap_stream(body.map(move |chunk| { mitm2.response_body_chunk(chunk) })); Response::from_parts(parts, body) }) .map_err(|(e, _f)| { std::io::Error::new(std::io::ErrorKind::Other, e) }) }) .flatten() .or_else(|e| { info!("proxy_request() returning 502 ({})", e); future::ok( Response::builder().status(502).body(Body::empty()).unwrap(), ) }) } /// Handles a CONNECT request. Tries to obtain an https connection to the /// remote server. If that fails, returns 502 Bad Gateway. Otherwise returns /// 200 OK, then attempts to establish a TLS connection with the proxy client, /// masquerading as the remote server. fn proxy_connect<T: Mitm + Sync + Send + 'static>( connect_req: Request<Body>, ) -> impl Future<Item = Response<Body>, Error = std::io::Error> { info!("proxy_connect() impersonating {:?}", connect_req.uri()); let authority = Authority::from_shared(Bytes::from(connect_req.uri().to_string())) .unwrap(); let tls_cfg = certauth::tls_config(&authority); let uri = http::uri::Builder::new() .scheme("https") .authority(authority) .path_and_query("/") .build() .unwrap(); obtain_connection(uri) .map(move |_pooled| { let inner = connect_req.into_body().on_upgrade().map_err(|e| { info!("proxy_connect() on_upgrade error: {:?}", e); std::io::Error::new(std::io::ErrorKind::Other, e) }) .and_then(|upgraded: Upgraded| -> Accept<Upgraded> { TlsAcceptor::from(tls_cfg).accept(upgraded) }) .map(move |stream: T
{ MitmProxyService::<T> { _phantom: std::marker::PhantomData, } }
identifier_body
lib.rs
. fn request_headers(&self, req: Request<Body>) -> Request<Body>; /// Observe and manipulate a chunk of the request payload. This function /// may be called zero or more times, depending on the size of the request /// payload. It will not be called at all in the common case of a GET /// request with no payload. The `chunk` argument contains an original /// chunk of the request payload as received from the proxy client. The /// return value of this function is sent to the remote server. fn request_body_chunk(&self, chunk: Chunk) -> Chunk; /// Observe and manipulate the response headers. The `res` argument /// contains the original response headers received from the remote server. /// The response headers returned by this function are sent to the proxy /// client. fn response_headers(&self, res: Response<Body>) -> Response<Body>; /// Observe and manipulate a chunk of the response payload. This function /// may be called zero or more times, depending on the size of the payload. /// The `chunk` argument represents an unaltered chunk of the response /// payload as received from the remote server. The return value of this /// function is sent to the remote server. fn response_body_chunk(&self, chunk: Chunk) -> Chunk; } lazy_static! { static ref CLIENT: Client<HttpsConnector<HttpConnector>, Body> = Client::builder().build(HttpsConnector::new(4)); static ref HTTP: Http = Http::new(); } /// The `hyper::service::Service` that does the proxying and calls your `Mitm` /// implementation. #[derive(Debug)] pub struct MitmProxyService<T: Mitm + Sync> { _phantom: std::marker::PhantomData<T>, } impl<T: Mitm + Sync> MitmProxyService<T> { /// Creates a new `MitmProxyService`. #[inline] pub fn new() -> Self { MitmProxyService::<T> { _phantom: std::marker::PhantomData, } } } impl<T: Mitm + Sync + Send + 'static> NewService for MitmProxyService<T> { type ReqBody = Body; type ResBody = Body; type Error = std::io::Error; type Service = MitmProxyService<T>; type InitError = std::io::Error; type Future = FutureResult<Self::Service, Self::InitError>; fn new_service(&self) -> Self::Future { future::ok(MitmProxyService::new()) } } impl<T: Mitm + Sync + Send + 'static> Service for MitmProxyService<T> { type ReqBody = Body; type ResBody = Body; type Error = std::io::Error; type Future = Box<dyn Future<Item = Response<Body>, Error = std::io::Error> + Send>; fn call(&mut self, req: Request<Body>) -> Self::Future { info!("MitmProxyService::call() handling {:?}", req); if *req.method() == Method::CONNECT { Box::new(proxy_connect::<T>(req)) } else { Box::new(proxy_request::<T>(req)) } } } /// Obtains a connection to the scheme://authority of `uri` from the connection /// pool. fn obtain_connection( uri: Uri, ) -> impl Future<Item = Pooled<PoolClient<Body>>, Error = std::io::Error> { let key1 = Arc::new(format!( "{}://{}", uri.scheme_part().unwrap(), uri.authority_part().unwrap() )); let key2 = Arc::clone(&key1); let result = CLIENT.connection_for(uri, key1).map_err(move |e| { std::io::Error::new( std::io::ErrorKind::Other, format!("error obtaining connection to {}: {:?}", key2, e), ) }); result } /// Obtains a connection to the remote server and proxies the request, calling /// the `Mitm` implementation functions, which may manipulate the request and /// reponse. Returns a future that resolves to the response or error. /// /// This function is called for plain http requests, and for https requests /// received "inside" the fake, tapped `CONNECT` tunnel. fn proxy_request<T: Mitm + Sync + Send + 'static>( req: Request<Body>, ) -> impl Future<Item = Response<Body>, Error = std::io::Error> { obtain_connection(req.uri().to_owned()) .map(|mut connection| { let mitm1 = Arc::new(T::new(req.uri().to_owned())); let mitm2 = Arc::clone(&mitm1); let req = mitm1.request_headers(req); let (parts, body) = req.into_parts(); let body = Body::wrap_stream( body.map(move |chunk| mitm1.request_body_chunk(chunk)), ); let req = Request::from_parts(parts, body); info!("proxy_request() sending request {:?}", req); connection .send_request_retryable(req) .map(|response| { let response = mitm2.response_headers(response); let (parts, body) = response.into_parts(); let body = Body::wrap_stream(body.map(move |chunk| { mitm2.response_body_chunk(chunk) })); Response::from_parts(parts, body) }) .map_err(|(e, _f)| { std::io::Error::new(std::io::ErrorKind::Other, e) }) }) .flatten() .or_else(|e| { info!("proxy_request() returning 502 ({})", e); future::ok( Response::builder().status(502).body(Body::empty()).unwrap(), ) }) } /// Handles a CONNECT request. Tries to obtain an https connection to the /// remote server. If that fails, returns 502 Bad Gateway. Otherwise returns /// 200 OK, then attempts to establish a TLS connection with the proxy client, /// masquerading as the remote server. fn proxy_connect<T: Mitm + Sync + Send + 'static>( connect_req: Request<Body>, ) -> impl Future<Item = Response<Body>, Error = std::io::Error> { info!("proxy_connect() impersonating {:?}", connect_req.uri()); let authority = Authority::from_shared(Bytes::from(connect_req.uri().to_string())) .unwrap(); let tls_cfg = certauth::tls_config(&authority); let uri = http::uri::Builder::new() .scheme("https") .authority(authority) .path_and_query("/") .build() .unwrap(); obtain_connection(uri) .map(move |_pooled| { let inner = connect_req.into_body().on_upgrade().map_err(|e| { info!("proxy_connect() on_upgrade error: {:?}", e); std::io::Error::new(std::io::ErrorKind::Other, e) }) .and_then(|upgraded: Upgraded| -> Accept<Upgraded> { TlsAcceptor::from(tls_cfg).accept(upgraded) }) .map(move |stream: TlsStream<Upgraded, rustls::ServerSession>| { info!("proxy_connect() tls connection established with proxy \ client: {:?}", stream); service_inner_requests::<T>(stream) }) .map_err(|e: std::io::Error| { error!("proxy_connect() error from somewhere: {}", e); }) .flatten(); hyper::rt::spawn(inner); Response::builder().status(200).body(Body::empty()).unwrap() }) .or_else(|e| { info!("proxy_connect() returning 502, failed to connect: {:?}", e); future::ok( Response::builder().status(502).body(Body::empty()).unwrap(), ) }) } /// Called by `proxy_connect()` once the TLS session has been established with /// the proxy client. Proxies requests received on the TLS connection. fn service_inner_requests<T: Mitm + Sync + Send + 'static>( stream: TlsStream<Upgraded, rustls::ServerSession>, ) -> impl Future<Item = (), Error = ()> { let svc = service_fn(move |req: Request<Body>| { // "host" header is required for http 1.1 // XXX but we could fall back on authority let authority = req.headers().get("host").unwrap().to_str().unwrap(); let uri = http::uri::Builder::new() .scheme("https") .authority(authority) .path_and_query(&req.uri().to_string() as &str) .build() .unwrap(); let (mut parts, body) = req.into_parts(); parts.uri = uri; let req = Request::from_parts(parts, body); proxy_request::<T>(req) }); HTTP.serve_connection(stream, svc) .map_err(|e: hyper::Error| { if match e.source() { Some(source) => source .to_string() .find("Connection reset by peer") .is_some(), None => false, }
{ info!( "service_inner_requests() serve_connection: \ client closed connection" ); }
conditional_block
lib.rs
= ([127, 0, 0, 1], 8000).into(); //! let svc = MitmProxyService::<MyMitm>::new(); //! let server = Server::bind(&addr) //! .serve(svc) //! .map_err(|e| eprintln!("server error: {}", e)); //! println!("noop mitm proxy listening on http://{}", addr); //! hyper::rt::run(server); //! } //! ``` //! //! Other examples can be found at //! <https://github.com/nlevitt/monie/tree/master/examples>. #![deny(warnings)] #![deny(missing_docs)] #![deny(missing_debug_implementations)] #[macro_use] extern crate lazy_static; #[macro_use] extern crate log; pub mod certauth; use std::error::Error; use std::sync::Arc; use bytes::Bytes; use futures::future::{self, Future, FutureResult}; use futures::stream::Stream; use http::method::Method; use http::uri::{Authority, Uri}; use hyper::client::pool::Pooled; use hyper::client::{HttpConnector, PoolClient}; use hyper::server::conn::Http; use hyper::service::{service_fn, NewService, Service}; use hyper::upgrade::Upgraded; use hyper::{Body, Chunk, Client, Request, Response}; use hyper_rustls::HttpsConnector; use tokio_rustls::{Accept, TlsAcceptor, TlsStream}; /// Represents the interception of a single request. Users of the library must /// implement this trait. With it you can observe and manipulate the request /// and response payload and headers. pub trait Mitm { /// Create a new instance of this `Mitm` implementation. The argument `uri` /// is the uri being proxied. Implementations may do with this what they /// wish (log it, stash it, ignore it, etc). fn new(uri: Uri) -> Self; /// Observe and manipulate the request headers. The `req` argument contains /// the original request headers received from the proxy client. The /// request headers returned by this function are sent to the remote /// server. fn request_headers(&self, req: Request<Body>) -> Request<Body>; /// Observe and manipulate a chunk of the request payload. This function /// may be called zero or more times, depending on the size of the request /// payload. It will not be called at all in the common case of a GET /// request with no payload. The `chunk` argument contains an original /// chunk of the request payload as received from the proxy client. The /// return value of this function is sent to the remote server. fn request_body_chunk(&self, chunk: Chunk) -> Chunk; /// Observe and manipulate the response headers. The `res` argument /// contains the original response headers received from the remote server. /// The response headers returned by this function are sent to the proxy /// client. fn response_headers(&self, res: Response<Body>) -> Response<Body>; /// Observe and manipulate a chunk of the response payload. This function /// may be called zero or more times, depending on the size of the payload. /// The `chunk` argument represents an unaltered chunk of the response /// payload as received from the remote server. The return value of this /// function is sent to the remote server. fn response_body_chunk(&self, chunk: Chunk) -> Chunk; } lazy_static! { static ref CLIENT: Client<HttpsConnector<HttpConnector>, Body> = Client::builder().build(HttpsConnector::new(4)); static ref HTTP: Http = Http::new(); } /// The `hyper::service::Service` that does the proxying and calls your `Mitm` /// implementation. #[derive(Debug)] pub struct MitmProxyService<T: Mitm + Sync> { _phantom: std::marker::PhantomData<T>, } impl<T: Mitm + Sync> MitmProxyService<T> { /// Creates a new `MitmProxyService`. #[inline] pub fn new() -> Self { MitmProxyService::<T> { _phantom: std::marker::PhantomData, } } } impl<T: Mitm + Sync + Send + 'static> NewService for MitmProxyService<T> { type ReqBody = Body; type ResBody = Body; type Error = std::io::Error; type Service = MitmProxyService<T>; type InitError = std::io::Error; type Future = FutureResult<Self::Service, Self::InitError>; fn new_service(&self) -> Self::Future { future::ok(MitmProxyService::new()) } } impl<T: Mitm + Sync + Send + 'static> Service for MitmProxyService<T> { type ReqBody = Body; type ResBody = Body; type Error = std::io::Error; type Future = Box<dyn Future<Item = Response<Body>, Error = std::io::Error> + Send>; fn call(&mut self, req: Request<Body>) -> Self::Future { info!("MitmProxyService::call() handling {:?}", req); if *req.method() == Method::CONNECT { Box::new(proxy_connect::<T>(req)) } else { Box::new(proxy_request::<T>(req)) } } } /// Obtains a connection to the scheme://authority of `uri` from the connection /// pool. fn obtain_connection( uri: Uri, ) -> impl Future<Item = Pooled<PoolClient<Body>>, Error = std::io::Error> { let key1 = Arc::new(format!( "{}://{}", uri.scheme_part().unwrap(), uri.authority_part().unwrap() )); let key2 = Arc::clone(&key1); let result = CLIENT.connection_for(uri, key1).map_err(move |e| { std::io::Error::new( std::io::ErrorKind::Other, format!("error obtaining connection to {}: {:?}", key2, e), ) }); result } /// Obtains a connection to the remote server and proxies the request, calling /// the `Mitm` implementation functions, which may manipulate the request and /// reponse. Returns a future that resolves to the response or error. /// /// This function is called for plain http requests, and for https requests /// received "inside" the fake, tapped `CONNECT` tunnel. fn proxy_request<T: Mitm + Sync + Send + 'static>( req: Request<Body>, ) -> impl Future<Item = Response<Body>, Error = std::io::Error> { obtain_connection(req.uri().to_owned()) .map(|mut connection| { let mitm1 = Arc::new(T::new(req.uri().to_owned())); let mitm2 = Arc::clone(&mitm1); let req = mitm1.request_headers(req); let (parts, body) = req.into_parts(); let body = Body::wrap_stream( body.map(move |chunk| mitm1.request_body_chunk(chunk)), ); let req = Request::from_parts(parts, body); info!("proxy_request() sending request {:?}", req); connection .send_request_retryable(req) .map(|response| { let response = mitm2.response_headers(response); let (parts, body) = response.into_parts(); let body = Body::wrap_stream(body.map(move |chunk| { mitm2.response_body_chunk(chunk) })); Response::from_parts(parts, body) }) .map_err(|(e, _f)| { std::io::Error::new(std::io::ErrorKind::Other, e) }) }) .flatten() .or_else(|e| { info!("proxy_request() returning 502 ({})", e); future::ok( Response::builder().status(502).body(Body::empty()).unwrap(), ) }) } /// Handles a CONNECT request. Tries to obtain an https connection to the /// remote server. If that fails, returns 502 Bad Gateway. Otherwise returns /// 200 OK, then attempts to establish a TLS connection with the proxy client, /// masquerading as the remote server. fn
<T: Mitm + Sync + Send + 'static>( connect_req: Request<Body>, ) -> impl Future<Item = Response<Body>, Error = std::io::Error> { info!("proxy_connect() impersonating {:?}", connect_req.uri()); let authority = Authority::from_shared(Bytes::from(connect_req.uri().to_string())) .unwrap(); let tls_cfg = certauth::tls_config(&authority); let uri = http::uri::Builder::new() .scheme("https") .authority(authority) .path_and_query("/") .build() .unwrap(); obtain_connection(uri) .map(move |_pooled| { let inner = connect_req.into_body().on_upgrade().map_err(|e| { info!("proxy_connect() on_upgrade error: {:?}", e); std::io::Error::new(std::io::ErrorKind::Other, e) }) .and_then(|upgraded: Upgraded| -> Accept<Upgraded> { TlsAcceptor::from(tls_cfg).accept(upgraded) }) .map(move |stream: T
proxy_connect
identifier_name
lib.rs
= ([127, 0, 0, 1], 8000).into(); //! let svc = MitmProxyService::<MyMitm>::new(); //! let server = Server::bind(&addr) //! .serve(svc) //! .map_err(|e| eprintln!("server error: {}", e)); //! println!("noop mitm proxy listening on http://{}", addr); //! hyper::rt::run(server); //! } //! ``` //! //! Other examples can be found at //! <https://github.com/nlevitt/monie/tree/master/examples>. #![deny(warnings)] #![deny(missing_docs)] #![deny(missing_debug_implementations)] #[macro_use] extern crate lazy_static; #[macro_use] extern crate log; pub mod certauth; use std::error::Error; use std::sync::Arc; use bytes::Bytes; use futures::future::{self, Future, FutureResult}; use futures::stream::Stream; use http::method::Method; use http::uri::{Authority, Uri}; use hyper::client::pool::Pooled; use hyper::client::{HttpConnector, PoolClient}; use hyper::server::conn::Http; use hyper::service::{service_fn, NewService, Service}; use hyper::upgrade::Upgraded; use hyper::{Body, Chunk, Client, Request, Response}; use hyper_rustls::HttpsConnector; use tokio_rustls::{Accept, TlsAcceptor, TlsStream}; /// Represents the interception of a single request. Users of the library must /// implement this trait. With it you can observe and manipulate the request /// and response payload and headers. pub trait Mitm { /// Create a new instance of this `Mitm` implementation. The argument `uri` /// is the uri being proxied. Implementations may do with this what they /// wish (log it, stash it, ignore it, etc). fn new(uri: Uri) -> Self; /// Observe and manipulate the request headers. The `req` argument contains /// the original request headers received from the proxy client. The /// request headers returned by this function are sent to the remote /// server. fn request_headers(&self, req: Request<Body>) -> Request<Body>; /// Observe and manipulate a chunk of the request payload. This function /// may be called zero or more times, depending on the size of the request /// payload. It will not be called at all in the common case of a GET /// request with no payload. The `chunk` argument contains an original /// chunk of the request payload as received from the proxy client. The /// return value of this function is sent to the remote server. fn request_body_chunk(&self, chunk: Chunk) -> Chunk; /// Observe and manipulate the response headers. The `res` argument /// contains the original response headers received from the remote server. /// The response headers returned by this function are sent to the proxy /// client. fn response_headers(&self, res: Response<Body>) -> Response<Body>; /// Observe and manipulate a chunk of the response payload. This function /// may be called zero or more times, depending on the size of the payload. /// The `chunk` argument represents an unaltered chunk of the response /// payload as received from the remote server. The return value of this /// function is sent to the remote server. fn response_body_chunk(&self, chunk: Chunk) -> Chunk; } lazy_static! { static ref CLIENT: Client<HttpsConnector<HttpConnector>, Body> = Client::builder().build(HttpsConnector::new(4)); static ref HTTP: Http = Http::new(); } /// The `hyper::service::Service` that does the proxying and calls your `Mitm` /// implementation. #[derive(Debug)] pub struct MitmProxyService<T: Mitm + Sync> { _phantom: std::marker::PhantomData<T>, } impl<T: Mitm + Sync> MitmProxyService<T> { /// Creates a new `MitmProxyService`. #[inline] pub fn new() -> Self { MitmProxyService::<T> { _phantom: std::marker::PhantomData, } } } impl<T: Mitm + Sync + Send + 'static> NewService for MitmProxyService<T> { type ReqBody = Body; type ResBody = Body; type Error = std::io::Error; type Service = MitmProxyService<T>; type InitError = std::io::Error; type Future = FutureResult<Self::Service, Self::InitError>; fn new_service(&self) -> Self::Future { future::ok(MitmProxyService::new()) } } impl<T: Mitm + Sync + Send + 'static> Service for MitmProxyService<T> { type ReqBody = Body; type ResBody = Body; type Error = std::io::Error; type Future = Box<dyn Future<Item = Response<Body>, Error = std::io::Error> + Send>; fn call(&mut self, req: Request<Body>) -> Self::Future { info!("MitmProxyService::call() handling {:?}", req); if *req.method() == Method::CONNECT { Box::new(proxy_connect::<T>(req)) } else { Box::new(proxy_request::<T>(req)) } } } /// Obtains a connection to the scheme://authority of `uri` from the connection /// pool. fn obtain_connection( uri: Uri, ) -> impl Future<Item = Pooled<PoolClient<Body>>, Error = std::io::Error> { let key1 = Arc::new(format!( "{}://{}", uri.scheme_part().unwrap(), uri.authority_part().unwrap() )); let key2 = Arc::clone(&key1); let result = CLIENT.connection_for(uri, key1).map_err(move |e| { std::io::Error::new( std::io::ErrorKind::Other, format!("error obtaining connection to {}: {:?}", key2, e), ) }); result } /// Obtains a connection to the remote server and proxies the request, calling /// the `Mitm` implementation functions, which may manipulate the request and /// reponse. Returns a future that resolves to the response or error.
/// received "inside" the fake, tapped `CONNECT` tunnel. fn proxy_request<T: Mitm + Sync + Send + 'static>( req: Request<Body>, ) -> impl Future<Item = Response<Body>, Error = std::io::Error> { obtain_connection(req.uri().to_owned()) .map(|mut connection| { let mitm1 = Arc::new(T::new(req.uri().to_owned())); let mitm2 = Arc::clone(&mitm1); let req = mitm1.request_headers(req); let (parts, body) = req.into_parts(); let body = Body::wrap_stream( body.map(move |chunk| mitm1.request_body_chunk(chunk)), ); let req = Request::from_parts(parts, body); info!("proxy_request() sending request {:?}", req); connection .send_request_retryable(req) .map(|response| { let response = mitm2.response_headers(response); let (parts, body) = response.into_parts(); let body = Body::wrap_stream(body.map(move |chunk| { mitm2.response_body_chunk(chunk) })); Response::from_parts(parts, body) }) .map_err(|(e, _f)| { std::io::Error::new(std::io::ErrorKind::Other, e) }) }) .flatten() .or_else(|e| { info!("proxy_request() returning 502 ({})", e); future::ok( Response::builder().status(502).body(Body::empty()).unwrap(), ) }) } /// Handles a CONNECT request. Tries to obtain an https connection to the /// remote server. If that fails, returns 502 Bad Gateway. Otherwise returns /// 200 OK, then attempts to establish a TLS connection with the proxy client, /// masquerading as the remote server. fn proxy_connect<T: Mitm + Sync + Send + 'static>( connect_req: Request<Body>, ) -> impl Future<Item = Response<Body>, Error = std::io::Error> { info!("proxy_connect() impersonating {:?}", connect_req.uri()); let authority = Authority::from_shared(Bytes::from(connect_req.uri().to_string())) .unwrap(); let tls_cfg = certauth::tls_config(&authority); let uri = http::uri::Builder::new() .scheme("https") .authority(authority) .path_and_query("/") .build() .unwrap(); obtain_connection(uri) .map(move |_pooled| { let inner = connect_req.into_body().on_upgrade().map_err(|e| { info!("proxy_connect() on_upgrade error: {:?}", e); std::io::Error::new(std::io::ErrorKind::Other, e) }) .and_then(|upgraded: Upgraded| -> Accept<Upgraded> { TlsAcceptor::from(tls_cfg).accept(upgraded) }) .map(move |stream: TlsStream
/// /// This function is called for plain http requests, and for https requests
random_line_split
stereo_calibration.py
else: term = (cv2.TERM_CRITERIA_EPS + cv2.TERM_CRITERIA_COUNT, 30, 0.1) cv2.cornerSubPix(img, corners, (5, 5), (-1, -1), term) if displayCorners is True: # print(filename) vis = cv2.cvtColor(img, cv2.COLOR_GRAY2BGR) cv2.drawChessboardCorners(vis, boardSize, corners, found) cv2.imshow("corners", vis) # cv2.waitKey(50) if k is 1: # Image from side camera goodImageList.append(imageList[i*2]) goodImageList.append(imageList[i*2+1]) j = j + 1 sideImagePoints.append(corners.reshape(-1,2)) topImagePoints.append(tempTop) objectPoints.append(pattern_points) print('Added left and right points') else: # Image from top camera # rightImagePoints.append(corners.reshape(-1,2)) tempTop= corners.reshape(-1,2) # imagePoints[k].append(corners.reshape(-1,2)) # objectPoints.append(pattern_points) print('OK') # print(corners) print(str(j) + " chessboard pairs have been detected\n") nImages = j if nImages < 2: print("Too few pairs to run calibration\n") return # print(imagePoints[1]) # print(objectPoints) print("Img count: " + str(len(topImagePoints))) print("Obj count: " + str(len(objectPoints))) # print(np.array(imagePoints[0])) top_calibration = np.load('top_calibration.npz') side_calibration = np.load('side_calibration.npz') top_rms = top_calibration['rms'] top_camera_matrix = top_calibration['camera_matrix'] top_dist_coefs = top_calibration['dist_coefs'] side_rms = side_calibration['rms'] side_camera_matrix = side_calibration['camera_matrix'] side_dist_coefs = side_calibration['dist_coefs'] # camera_transform = np.load('camera_transform.npz') # R = camera_transform['R'] # T = camera_transform['T'] # print('Calibrating top...') # top_rms, top_camera_matrix, top_dist_coefs, rvecs, tvecs = cv2.calibrateCamera(objectPoints, topImagePoints, imageSize, None, None) print("Top Camera\nRMS:" + str(top_rms)) print("camera matrix: " + str(top_camera_matrix)) print("distortion coefficients: " + str(top_dist_coefs.ravel())) # print('Calibrating side...') # side_rms, side_camera_matrix, side_dist_coefs, rvecs, tvecs = cv2.calibrateCamera(objectPoints, sideImagePoints, imageSize, None, None) print("Side Camera\nRMS:", side_rms) print("camera matrix:\n", side_camera_matrix) print("distortion coefficients: ", side_dist_coefs.ravel()) # top_undistorted = cv2.undistort(cv2.imread(imageList[0]), top_camera_matrix, top_dist_coefs) # cv2.imshow("Top Undistorted", top_undistorted) # side_undistorted = cv2.undistort(cv2.imread(imageList[1]), side_camera_matrix, side_dist_coefs) # cv2.imshow("Side Undistorted", side_undistorted) # cameraMatrix[0] = cv2.initCameraMatrix2D(np.array(objectPoints), np.array(imagePoints[0]), imageSize, 0) # cameraMatrix[1] = cv2.initCameraMatrix2D(objectPoints, imagePoints[1], imageSize, 0) # print(objectPoints[0]) # ret, rvec_top, tvec_top = cv2.solvePnP(objectPoints[0], topImagePoints[0], top_camera_matrix, top_dist_coefs) # ret, rvec_side, tvec_side = cv2.solvePnP(objectPoints[0], sideImagePoints[0], side_camera_matrix, side_dist_coefs) # print('\n') # print('rvec top', rvec_top) # print('tvec top', tvec_top) # print('rvec side', rvec_side) # print('tvec side', tvec_side) # print(topImagePoints[0]) print('\n') print('Stereo calibration (cv2)...') retval, cameraMatrix1, distCoeffs1, cameraMatrix2, distCoeffs2, R, T, E, F = cv2.stereoCalibrate(objectPoints, topImagePoints, sideImagePoints, top_camera_matrix, top_dist_coefs, side_camera_matrix, side_dist_coefs, imageSize, (cv2.CALIB_FIX_INTRINSIC)) print("Rotation: ", R) r_vec, jac = cv2.Rodrigues(R) print("R_vec: ", np.multiply(r_vec, 180/math.pi)) print("Translation: ", T) print("Essential: ", E) print("Fundamental: ", F) print('Stereo calibration (DIY)...') topImagePointsConcat = topImagePoints[0] sideImagePointsConcat = sideImagePoints[0] for i in range(1, len(topImagePoints)): topImagePointsConcat = np.concatenate((topImagePointsConcat, topImagePoints[i])) sideImagePointsConcat = np.concatenate((sideImagePointsConcat, sideImagePoints[i])) m1 = np.ones((len(topImagePointsConcat), 3)) m1[:,0:2] = topImagePointsConcat m2 = np.ones((len(sideImagePointsConcat), 3)) m2[:,0:2] = sideImagePointsConcat x1, T1 = normalizePoints(m1) x2, T2 = normalizePoints(m2) # print('Normalized', x1, T) # Normalization matrix # N = np.array([[2.0/w,0.0,-1.0],[0.0,2.0/h,-1.0],[0.0,0.0,1.0]], np.float64) # print('N', N) # x1 = np.dot(N,m1.T).T # print('x1,',x1) # x2 = np.dot(N,m2.T).T # print('x2',x2) A = np.ones((len(topImagePointsConcat),9)) A[:,0] = np.multiply(x1[:,0],x2[:,0]) A[:,1] = np.multiply(x1[:,1],x2[:,0]) A[:,2] = x2[:,0] A[:,3] = np.multiply(x1[:,0],x2[:,1]) A[:,4] = np.multiply(x1[:,1],x2[:,1]) A[:,5] = x2[:,1] A[:,6] = x1[:,0] A[:,7] = x1[:,1] # A[:,0] = np.multiply(x2[:,0],x1[:,0]) # A[:,1] = np.multiply(x2[:,0],x1[:,1]) # A[:,2] = x2[:,0] # A[:,3] = np.multiply(x2[:,1],x1[:,0]) # A[:,4] = np.multiply(x2[:,1],x1[:,1]) # A[:,5] = x2[:,1] # A[:,6] = x1[:,0] # A[:,7] = x1[:,1] print(A) U, D, V = np.linalg.svd(A) # print('U',U) # print('D',D) # print('V',V) V = V.conj().T F_new = V[:,8].reshape(3,3).copy() # make rank 2 U, D, V = np.linalg.svd(F_new); # print('U',U) # print('D',D) # print('V',V) D_diag = np.diag([D[0], D[1], 0]) F_new = np.dot(np.dot(U, D_diag), V) # F_new=np.dot(N.T,np.dot(F_new,N)) F_new = np.dot(np.dot(T2.T, F_new), T1) print(F_new) # R, jac = cv2.Rodrigues(np.dot(np.array([[-90],[0],[0]], dtype = np.float64), math.pi/180)) T = np.array([[0], [-130], [130]], dtype=np.float64) print("Rotation: ", R) # r_vec, jac = cv2.Rodrigues(R) # print("R_vec: ", r_vec) print("Translation: ", T) # print("Fundamental: ", F_new) # F = F_new # top_undistorted = cv2.undistort(cv2.imread(imageList[0]), cameraMatrix1, distCoeffs1) # cv2.imshow("Top Undistorted", top_undistorted) # side_undistorted = cv2.undistort(cv2.imread(imageList[1]), cameraMatrix2, distCoeffs2) # cv2.imshow("Side Undistorted", side_undistorted) # R1, R2, P
break
random_line_split
stereo_calibration.py
_FIX_INTRINSIC)) print("Rotation: ", R) r_vec, jac = cv2.Rodrigues(R) print("R_vec: ", np.multiply(r_vec, 180/math.pi)) print("Translation: ", T) print("Essential: ", E) print("Fundamental: ", F) print('Stereo calibration (DIY)...') topImagePointsConcat = topImagePoints[0] sideImagePointsConcat = sideImagePoints[0] for i in range(1, len(topImagePoints)): topImagePointsConcat = np.concatenate((topImagePointsConcat, topImagePoints[i])) sideImagePointsConcat = np.concatenate((sideImagePointsConcat, sideImagePoints[i])) m1 = np.ones((len(topImagePointsConcat), 3)) m1[:,0:2] = topImagePointsConcat m2 = np.ones((len(sideImagePointsConcat), 3)) m2[:,0:2] = sideImagePointsConcat x1, T1 = normalizePoints(m1) x2, T2 = normalizePoints(m2) # print('Normalized', x1, T) # Normalization matrix # N = np.array([[2.0/w,0.0,-1.0],[0.0,2.0/h,-1.0],[0.0,0.0,1.0]], np.float64) # print('N', N) # x1 = np.dot(N,m1.T).T # print('x1,',x1) # x2 = np.dot(N,m2.T).T # print('x2',x2) A = np.ones((len(topImagePointsConcat),9)) A[:,0] = np.multiply(x1[:,0],x2[:,0]) A[:,1] = np.multiply(x1[:,1],x2[:,0]) A[:,2] = x2[:,0] A[:,3] = np.multiply(x1[:,0],x2[:,1]) A[:,4] = np.multiply(x1[:,1],x2[:,1]) A[:,5] = x2[:,1] A[:,6] = x1[:,0] A[:,7] = x1[:,1] # A[:,0] = np.multiply(x2[:,0],x1[:,0]) # A[:,1] = np.multiply(x2[:,0],x1[:,1]) # A[:,2] = x2[:,0] # A[:,3] = np.multiply(x2[:,1],x1[:,0]) # A[:,4] = np.multiply(x2[:,1],x1[:,1]) # A[:,5] = x2[:,1] # A[:,6] = x1[:,0] # A[:,7] = x1[:,1] print(A) U, D, V = np.linalg.svd(A) # print('U',U) # print('D',D) # print('V',V) V = V.conj().T F_new = V[:,8].reshape(3,3).copy() # make rank 2 U, D, V = np.linalg.svd(F_new); # print('U',U) # print('D',D) # print('V',V) D_diag = np.diag([D[0], D[1], 0]) F_new = np.dot(np.dot(U, D_diag), V) # F_new=np.dot(N.T,np.dot(F_new,N)) F_new = np.dot(np.dot(T2.T, F_new), T1) print(F_new) # R, jac = cv2.Rodrigues(np.dot(np.array([[-90],[0],[0]], dtype = np.float64), math.pi/180)) T = np.array([[0], [-130], [130]], dtype=np.float64) print("Rotation: ", R) # r_vec, jac = cv2.Rodrigues(R) # print("R_vec: ", r_vec) print("Translation: ", T) # print("Fundamental: ", F_new) # F = F_new # top_undistorted = cv2.undistort(cv2.imread(imageList[0]), cameraMatrix1, distCoeffs1) # cv2.imshow("Top Undistorted", top_undistorted) # side_undistorted = cv2.undistort(cv2.imread(imageList[1]), cameraMatrix2, distCoeffs2) # cv2.imshow("Side Undistorted", side_undistorted) # R1, R2, P1, P2, Q, ret1, ret2 = cv2.stereoRectify(cameraMatrix1, distCoeffs1, cameraMatrix2, distCoeffs2, imageSize, R, T) # print('R', R) # R_vec, jac = cv2.Rodrigues(R) # print('R vec', R_vec) # print('T', T) print('\n') # print('Stereo rectification (cv2)...') # R1, R2, P1, P2, Q, ret1, ret2 = cv2.stereoRectify(top_camera_matrix, top_dist_coefs, side_camera_matrix, side_dist_coefs, imageSize, R, T, alpha=1) # print("R1: ", R1) # R1_vec, jac = cv2.Rodrigues(R1) # print("R1 vec: ", R1_vec) # print("R2: ", R2) # R2_vec, jac = cv2.Rodrigues(R2) # print("P1: ", P1) # print("P2: ", P2) # print('Q: ', Q) print('Stereo rectification (DIY)...') P1 = np.concatenate((np.dot(side_camera_matrix,np.eye(3)),np.dot(side_camera_matrix,np.zeros((3,1)))), axis = 1) P2 = np.concatenate((np.dot(side_camera_matrix,R),np.dot(side_camera_matrix,T)), axis = 1) # print("R2 vec: ", R2_vec) print("P1: ", P1) print("P2: ", P2) # np.savez_compressed('calibration.npz', R1=R1, R2=R2, P1=P1, P2=P2, CameraMatrix1=cameraMatrix1, CameraMatrix2=cameraMatrix2, DistCoeffs1=distCoeffs1, DistCoeffs2=distCoeffs2,R=R,T=T,E=E,F=F) # np.savez_compressed('calibration.npz', CameraMatrix1=top_camera_matrix, CameraMatrix2=side_camera_matrix, DistCoeffs1=top_dist_coefs, DistCoeffs2=side_dist_coefs) np.savez_compressed('calibration.npz', P1=P1, P2=P2, CameraMatrix1=top_camera_matrix, CameraMatrix2=side_camera_matrix, DistCoeffs1=top_dist_coefs, DistCoeffs2=side_dist_coefs,R=R,T=T,E=E,F=F) # path = np.load("path.npz") # top_path = path["top_path"] # side_path = path["side_path"] # tip3D_homogeneous = cv2.triangulatePoints(P1, P2, top_path.reshape(2,-1)[:,50:75], side_path.reshape(2,-1)[:,50:75]) # tip3D = (tip3D_homogeneous/tip3D_homogeneous[3])[0:3] # # print("homogeneous coords: " , tip3D_homogeneous) # print("3D coords: ", tip3D) # ax.scatter(np.array(tip3D)[0,:],np.array(tip3D)[1,:],np.array(tip3D)[2,:]) # # plt.show() # leftInputPoints = np.array(leftImagePoints[0]).reshape(2,-1) # rightInputPoints = np.array(rightImagePoints[0]).reshape(2,-1) # np.savez_compressed('points.npz',left=leftInputPoints,right=rightInputPoints) # print("Left inputs: " + str(leftInputPoints)) # points = cv2.triangulatePoints(P1, P2, leftInputPoints[:,50:100], rightInputPoints[:,50:100]) # print('\n') # testPoint = points[:,0] # testPoint3D = testPoint/testPoint[3] # point3D = points/points[3,:] # print("3D points: " + str(point3D)) def main():
size = (9, 7) squareSize = 6 # millimeters sourcePath = '/home/jgschornak/NeedleGuidance/images_converging_cams/' top_img_mask = sourcePath + 'top*.jpg' top_img_names = glob(top_img_mask) side_img_mask = sourcePath + 'side*.jpg' side_img_names = glob(side_img_mask) # print(left_img_names) # print('\n') # print(right_img_names) numPairs = len(top_img_names) imgList = [] for i in range(0, numPairs): imgList.append(sourcePath + 'top%i' % i + '.jpg') imgList.append(sourcePath + 'side%i' % i + '.jpg')
identifier_body
stereo_calibration.py
(imageList, boardSize, squareSize, displayCorners = True, useCalibrated = True, showRectified = True): nImages = int(len(imageList)/2) goodImageList = [] imageSize = None fig = plt.figure() ax = fig.add_subplot(111, projection='3d') topImagePoints = [] sideImagePoints = [] imagePoints = [] imagePoints.append([]) imagePoints.append([]) for n in range(0, nImages): imagePoints[0].append(None) imagePoints[1].append(None) pattern_points = np.zeros((np.prod(boardSize), 3), np.float32) pattern_points[:, :2] = np.indices(boardSize).T.reshape(-1, 2) pattern_points *= squareSize objectPoints = [] j = 0 tempTop = None for i in range(0, nImages): for k in range(0,2): filename = imageList[i*2+k] print('processsing %s... ' % filename) img = cv2.imread(filename, 0) if img is None: break if imageSize is None: imageSize = img.shape[:2] h, w = img.shape[:2] found, corners = cv2.findChessboardCorners(img, boardSize) if not found: print('chessboard not found') break else: term = (cv2.TERM_CRITERIA_EPS + cv2.TERM_CRITERIA_COUNT, 30, 0.1) cv2.cornerSubPix(img, corners, (5, 5), (-1, -1), term) if displayCorners is True: # print(filename) vis = cv2.cvtColor(img, cv2.COLOR_GRAY2BGR) cv2.drawChessboardCorners(vis, boardSize, corners, found) cv2.imshow("corners", vis) # cv2.waitKey(50) if k is 1: # Image from side camera goodImageList.append(imageList[i*2]) goodImageList.append(imageList[i*2+1]) j = j + 1 sideImagePoints.append(corners.reshape(-1,2)) topImagePoints.append(tempTop) objectPoints.append(pattern_points) print('Added left and right points') else: # Image from top camera # rightImagePoints.append(corners.reshape(-1,2)) tempTop= corners.reshape(-1,2) # imagePoints[k].append(corners.reshape(-1,2)) # objectPoints.append(pattern_points) print('OK') # print(corners) print(str(j) + " chessboard pairs have been detected\n") nImages = j if nImages < 2: print("Too few pairs to run calibration\n") return # print(imagePoints[1]) # print(objectPoints) print("Img count: " + str(len(topImagePoints))) print("Obj count: " + str(len(objectPoints))) # print(np.array(imagePoints[0])) top_calibration = np.load('top_calibration.npz') side_calibration = np.load('side_calibration.npz') top_rms = top_calibration['rms'] top_camera_matrix = top_calibration['camera_matrix'] top_dist_coefs = top_calibration['dist_coefs'] side_rms = side_calibration['rms'] side_camera_matrix = side_calibration['camera_matrix'] side_dist_coefs = side_calibration['dist_coefs'] # camera_transform = np.load('camera_transform.npz') # R = camera_transform['R'] # T = camera_transform['T'] # print('Calibrating top...') # top_rms, top_camera_matrix, top_dist_coefs, rvecs, tvecs = cv2.calibrateCamera(objectPoints, topImagePoints, imageSize, None, None) print("Top Camera\nRMS:" + str(top_rms)) print("camera matrix: " + str(top_camera_matrix)) print("distortion coefficients: " + str(top_dist_coefs.ravel())) # print('Calibrating side...') # side_rms, side_camera_matrix, side_dist_coefs, rvecs, tvecs = cv2.calibrateCamera(objectPoints, sideImagePoints, imageSize, None, None) print("Side Camera\nRMS:", side_rms) print("camera matrix:\n", side_camera_matrix) print("distortion coefficients: ", side_dist_coefs.ravel()) # top_undistorted = cv2.undistort(cv2.imread(imageList[0]), top_camera_matrix, top_dist_coefs) # cv2.imshow("Top Undistorted", top_undistorted) # side_undistorted = cv2.undistort(cv2.imread(imageList[1]), side_camera_matrix, side_dist_coefs) # cv2.imshow("Side Undistorted", side_undistorted) # cameraMatrix[0] = cv2.initCameraMatrix2D(np.array(objectPoints), np.array(imagePoints[0]), imageSize, 0) # cameraMatrix[1] = cv2.initCameraMatrix2D(objectPoints, imagePoints[1], imageSize, 0) # print(objectPoints[0]) # ret, rvec_top, tvec_top = cv2.solvePnP(objectPoints[0], topImagePoints[0], top_camera_matrix, top_dist_coefs) # ret, rvec_side, tvec_side = cv2.solvePnP(objectPoints[0], sideImagePoints[0], side_camera_matrix, side_dist_coefs) # print('\n') # print('rvec top', rvec_top) # print('tvec top', tvec_top) # print('rvec side', rvec_side) # print('tvec side', tvec_side) # print(topImagePoints[0]) print('\n') print('Stereo calibration (cv2)...') retval, cameraMatrix1, distCoeffs1, cameraMatrix2, distCoeffs2, R, T, E, F = cv2.stereoCalibrate(objectPoints, topImagePoints, sideImagePoints, top_camera_matrix, top_dist_coefs, side_camera_matrix, side_dist_coefs, imageSize, (cv2.CALIB_FIX_INTRINSIC)) print("Rotation: ", R) r_vec, jac = cv2.Rodrigues(R) print("R_vec: ", np.multiply(r_vec, 180/math.pi)) print("Translation: ", T) print("Essential: ", E) print("Fundamental: ", F) print('Stereo calibration (DIY)...') topImagePointsConcat = topImagePoints[0] sideImagePointsConcat = sideImagePoints[0] for i in range(1, len(topImagePoints)): topImagePointsConcat = np.concatenate((topImagePointsConcat, topImagePoints[i])) sideImagePointsConcat = np.concatenate((sideImagePointsConcat, sideImagePoints[i])) m1 = np.ones((len(topImagePointsConcat), 3)) m1[:,0:2] = topImagePointsConcat m2 = np.ones((len(sideImagePointsConcat), 3)) m2[:,0:2] = sideImagePointsConcat x1, T1 = normalizePoints(m1) x2, T2 = normalizePoints(m2) # print('Normalized', x1, T) # Normalization matrix # N = np.array([[2.0/w,0.0,-1.0],[0.0,2.0/h,-1.0],[0.0,0.0,1.0]], np.float64) # print('N', N) # x1 = np.dot(N,m1.T).T # print('x1,',x1) # x2 = np.dot(N,m2.T).T # print('x2',x2) A = np.ones((len(topImagePointsConcat),9)) A[:,0] = np.multiply(x1[:,0],x2[:,0]) A[:,1] = np.multiply(x1[:,1],x2[:,0]) A[:,2] = x2[:,0] A[:,3] = np.multiply(x1[:,0],x2[:,1]) A[:,4] = np.multiply(x1[:,1],x2[:,1]) A[:,5] = x2[:,1] A[:,6] = x1[:,0] A[:,7] = x1[:,1] # A[:,0] = np.multiply(x2[:,0],x1[:,0]) # A[:,1] = np.multiply(x2[:,0],x1[:,1]) # A[:,2] = x2[:,0] # A[:,3] = np.multiply(x2[:,1],x1[:,0]) # A[:,4] = np.multiply(x2[:,1],x1[:,1]) # A[:,5] = x2[:,1] # A[:,6] = x1[:,0] # A[:,7] = x1[:,1] print(A) U, D, V = np.linalg.svd(A) # print('U',U) # print('D',D) # print('V',V) V = V.conj().T F_new = V[:,8].reshape(3,3).copy() # make rank 2 U, D, V = np.linalg.svd(F_new); # print('U',U)
StereoCalib
identifier_name
stereo_calibration.py
# Image from side camera goodImageList.append(imageList[i*2]) goodImageList.append(imageList[i*2+1]) j = j + 1 sideImagePoints.append(corners.reshape(-1,2)) topImagePoints.append(tempTop) objectPoints.append(pattern_points) print('Added left and right points') else: # Image from top camera # rightImagePoints.append(corners.reshape(-1,2)) tempTop= corners.reshape(-1,2) # imagePoints[k].append(corners.reshape(-1,2)) # objectPoints.append(pattern_points) print('OK') # print(corners) print(str(j) + " chessboard pairs have been detected\n") nImages = j if nImages < 2:
# print(imagePoints[1]) # print(objectPoints) print("Img count: " + str(len(topImagePoints))) print("Obj count: " + str(len(objectPoints))) # print(np.array(imagePoints[0])) top_calibration = np.load('top_calibration.npz') side_calibration = np.load('side_calibration.npz') top_rms = top_calibration['rms'] top_camera_matrix = top_calibration['camera_matrix'] top_dist_coefs = top_calibration['dist_coefs'] side_rms = side_calibration['rms'] side_camera_matrix = side_calibration['camera_matrix'] side_dist_coefs = side_calibration['dist_coefs'] # camera_transform = np.load('camera_transform.npz') # R = camera_transform['R'] # T = camera_transform['T'] # print('Calibrating top...') # top_rms, top_camera_matrix, top_dist_coefs, rvecs, tvecs = cv2.calibrateCamera(objectPoints, topImagePoints, imageSize, None, None) print("Top Camera\nRMS:" + str(top_rms)) print("camera matrix: " + str(top_camera_matrix)) print("distortion coefficients: " + str(top_dist_coefs.ravel())) # print('Calibrating side...') # side_rms, side_camera_matrix, side_dist_coefs, rvecs, tvecs = cv2.calibrateCamera(objectPoints, sideImagePoints, imageSize, None, None) print("Side Camera\nRMS:", side_rms) print("camera matrix:\n", side_camera_matrix) print("distortion coefficients: ", side_dist_coefs.ravel()) # top_undistorted = cv2.undistort(cv2.imread(imageList[0]), top_camera_matrix, top_dist_coefs) # cv2.imshow("Top Undistorted", top_undistorted) # side_undistorted = cv2.undistort(cv2.imread(imageList[1]), side_camera_matrix, side_dist_coefs) # cv2.imshow("Side Undistorted", side_undistorted) # cameraMatrix[0] = cv2.initCameraMatrix2D(np.array(objectPoints), np.array(imagePoints[0]), imageSize, 0) # cameraMatrix[1] = cv2.initCameraMatrix2D(objectPoints, imagePoints[1], imageSize, 0) # print(objectPoints[0]) # ret, rvec_top, tvec_top = cv2.solvePnP(objectPoints[0], topImagePoints[0], top_camera_matrix, top_dist_coefs) # ret, rvec_side, tvec_side = cv2.solvePnP(objectPoints[0], sideImagePoints[0], side_camera_matrix, side_dist_coefs) # print('\n') # print('rvec top', rvec_top) # print('tvec top', tvec_top) # print('rvec side', rvec_side) # print('tvec side', tvec_side) # print(topImagePoints[0]) print('\n') print('Stereo calibration (cv2)...') retval, cameraMatrix1, distCoeffs1, cameraMatrix2, distCoeffs2, R, T, E, F = cv2.stereoCalibrate(objectPoints, topImagePoints, sideImagePoints, top_camera_matrix, top_dist_coefs, side_camera_matrix, side_dist_coefs, imageSize, (cv2.CALIB_FIX_INTRINSIC)) print("Rotation: ", R) r_vec, jac = cv2.Rodrigues(R) print("R_vec: ", np.multiply(r_vec, 180/math.pi)) print("Translation: ", T) print("Essential: ", E) print("Fundamental: ", F) print('Stereo calibration (DIY)...') topImagePointsConcat = topImagePoints[0] sideImagePointsConcat = sideImagePoints[0] for i in range(1, len(topImagePoints)): topImagePointsConcat = np.concatenate((topImagePointsConcat, topImagePoints[i])) sideImagePointsConcat = np.concatenate((sideImagePointsConcat, sideImagePoints[i])) m1 = np.ones((len(topImagePointsConcat), 3)) m1[:,0:2] = topImagePointsConcat m2 = np.ones((len(sideImagePointsConcat), 3)) m2[:,0:2] = sideImagePointsConcat x1, T1 = normalizePoints(m1) x2, T2 = normalizePoints(m2) # print('Normalized', x1, T) # Normalization matrix # N = np.array([[2.0/w,0.0,-1.0],[0.0,2.0/h,-1.0],[0.0,0.0,1.0]], np.float64) # print('N', N) # x1 = np.dot(N,m1.T).T # print('x1,',x1) # x2 = np.dot(N,m2.T).T # print('x2',x2) A = np.ones((len(topImagePointsConcat),9)) A[:,0] = np.multiply(x1[:,0],x2[:,0]) A[:,1] = np.multiply(x1[:,1],x2[:,0]) A[:,2] = x2[:,0] A[:,3] = np.multiply(x1[:,0],x2[:,1]) A[:,4] = np.multiply(x1[:,1],x2[:,1]) A[:,5] = x2[:,1] A[:,6] = x1[:,0] A[:,7] = x1[:,1] # A[:,0] = np.multiply(x2[:,0],x1[:,0]) # A[:,1] = np.multiply(x2[:,0],x1[:,1]) # A[:,2] = x2[:,0] # A[:,3] = np.multiply(x2[:,1],x1[:,0]) # A[:,4] = np.multiply(x2[:,1],x1[:,1]) # A[:,5] = x2[:,1] # A[:,6] = x1[:,0] # A[:,7] = x1[:,1] print(A) U, D, V = np.linalg.svd(A) # print('U',U) # print('D',D) # print('V',V) V = V.conj().T F_new = V[:,8].reshape(3,3).copy() # make rank 2 U, D, V = np.linalg.svd(F_new); # print('U',U) # print('D',D) # print('V',V) D_diag = np.diag([D[0], D[1], 0]) F_new = np.dot(np.dot(U, D_diag), V) # F_new=np.dot(N.T,np.dot(F_new,N)) F_new = np.dot(np.dot(T2.T, F_new), T1) print(F_new) # R, jac = cv2.Rodrigues(np.dot(np.array([[-90],[0],[0]], dtype = np.float64), math.pi/180)) T = np.array([[0], [-130], [130]], dtype=np.float64) print("Rotation: ", R) # r_vec, jac = cv2.Rodrigues(R) # print("R_vec: ", r_vec) print("Translation: ", T) # print("Fundamental: ", F_new) # F = F_new # top_undistorted = cv2.undistort(cv2.imread(imageList[0]), cameraMatrix1, distCoeffs1) # cv2.imshow("Top Undistorted", top_undistorted) # side_undistorted = cv2.undistort(cv2.imread(imageList[1]), cameraMatrix2, distCoeffs2) # cv2.imshow("Side Undistorted", side_undistorted) # R1, R2, P1, P2, Q, ret1, ret2 = cv2.stereoRectify(cameraMatrix1, distCoeffs1, cameraMatrix2, distCoeffs2, imageSize, R, T) # print('R', R) # R_vec, jac = cv2.Rodrigues(R) # print('R vec', R_vec) # print('T', T) print('\n') # print('Stereo rectification (cv2)...') # R1, R2, P1, P2, Q, ret1, ret2 = cv2.stereoRectify(top_camera_matrix, top_dist_coefs
print("Too few pairs to run calibration\n") return
conditional_block
block.rs
const ROW0_TABLE: LookupTable<u8, Block> = lookup_table![ (0x00, 0x7F, Basic), ]; const ROW0_LIMIT: char = '\u{80}'; const PLANE0_TABLE: LookupTable<u16, Block> = lookup_table![ (0x0080, 0x024F, Latin), (0x0250, 0x02AF, IPA), (0x02B0, 0x02FF, Spacing), (0x0300, 0x036F, Combining), (0x0370, 0x03FF, Greek), (0x0400, 0x052F, Cyrillic), (0x0530, 0x058F, Armenian), (0x0590, 0x05FF, Hebrew), (0x0600, 0x06FF, Arabic), (0x0700, 0x074F, Syriac), (0x0750, 0x077F, Arabic), (0x0780, 0x07BF, Thaana), (0x07C0, 0x07FF, NKo), (0x0800, 0x083F, Samaritan), (0x0840, 0x085F, Mandaic), (0x0860, 0x086F, Syriac), (0x08A0, 0x08FF, Arabic), (0x0900, 0x097F, Devanagari), (0x0980, 0x09FF, Bengali), (0x0A00, 0x0A7F, Gurmukhi), (0x0A80, 0x0AFF, Gujarati), (0x0B00, 0x0B7F, Oriya), (0x0B80, 0x0BFF, Tamil), (0x0C00, 0x0C7F, Telugu), (0x0C80, 0x0CFF, Kannada), (0x0D00, 0x0D7F, Malayalam), (0x0D80, 0x0DFF, Sinhala), (0x0E00, 0x0E7F, Thai), (0x0E80, 0x0EFF, Lao), (0x0F00, 0x0FFF, Tibetan), (0x1000, 0x109F, Myanmar), (0x10A0, 0x10FF, Georgian), (0x1100, 0x11FF, Hangul), (0x1200, 0x139F, Ethiopic), (0x13A0, 0x13FF, Cherokee), (0x1400, 0x167F, Unified), (0x1680, 0x169F, Ogham), (0x16A0, 0x16FF, Runic), (0x1700, 0x171F, Tagalog), (0x1720, 0x173F, Hanunoo), (0x1740, 0x175F, Buhid), (0x1760, 0x177F, Tagbanwa), (0x1780, 0x17FF, Khmer), (0x1800, 0x18AF, Mongolian), (0x18B0, 0x18FF, Unified), (0x1900, 0x194F, Limbu), (0x1950, 0x197F, Tai), (0x1980, 0x19DF, New), (0x19E0, 0x19FF, Khmer), (0x1A00, 0x1A1F, Buginese), (0x1A20, 0x1AAF, Tai), (0x1AB0, 0x1AFF, Combining), (0x1B00, 0x1B7F, Balinese), (0x1B80, 0x1BBF, Sundanese), (0x1BC0, 0x1BFF, Batak), (0x1C00, 0x1C4F, Lepcha), (0x1C50, 0x1C7F, Ol), (0x1C80, 0x1C8F, Cyrillic), (0x1C90, 0x1CBF, Georgian), (0x1CC0, 0x1CCF, Sundanese), (0x1CD0, 0x1CFF, Vedic), (0x1D00, 0x1DBF, Phonetic), (0x1DC0, 0x1DFF, Combining), (0x1E00, 0x1EFF, Latin), (0x1F00, 0x1FFF, Greek), (0x2000, 0x206F, General), (0x2070, 0x209F, Superscripts), (0x20A0, 0x20CF, Currency), (0x20D0, 0x20FF, Combining), (0x2100, 0x214F, Letterlike), (0x2150, 0x218F, Number), (0x2190, 0x21FF, Arrows), (0x2200, 0x22FF, Mathematical), (0x2300, 0x23FF, Miscellaneous), (0x2400, 0x243F, Control), (0x2440, 0x245F, Optical), (0x2460, 0x24FF, Enclosed), (0x2500, 0x257F, Box), (0x2580, 0x259F, Block), (0x25A0, 0x25FF, Geometric), (0x2600, 0x26FF, Miscellaneous), (0x2700, 0x27BF, Dingbats), (0x27C0, 0x27EF, Miscellaneous), (0x27F0, 0x27FF, Supplemental), (0x2800, 0x28FF, Braille), (0x2900, 0x297F, Supplemental), (0x2980, 0x29FF, Miscellaneous), (0x2A00, 0x2AFF, Supplemental), (0x2B00, 0x2BFF, Miscellaneous), (0x2C00, 0x2C5F, Glagolitic), (0x2C60, 0x2C7F, Latin), (0x2C80, 0x2CFF, Coptic), (0x2D00, 0x2D2F, Georgian), (0x2D30, 0x2D7F, Tifinagh), (0x2D80, 0x2DDF, Ethiopic), (0x2DE0, 0x2DFF, Cyrillic), (0x2E00, 0x2E7F, Supplemental), (0x2E80, 0x2EFF, CJK), (0x2F00, 0x2FDF
{ use std::convert::TryInto; ROW0_TABLE.validate(); if let Ok(x) = (ROW0_LIMIT as u32).try_into() { assert!(!ROW0_TABLE.contains(&x)); } PLANE0_TABLE.validate(); if let Ok(x) = (PLANE0_LIMIT as u32).try_into() { assert!(!PLANE0_TABLE.contains(&x)); } SUPPLEMENTARY_TABLE.validate(); }
identifier_body
block.rs
, 0x197F, Tai), (0x1980, 0x19DF, New), (0x19E0, 0x19FF, Khmer), (0x1A00, 0x1A1F, Buginese), (0x1A20, 0x1AAF, Tai), (0x1AB0, 0x1AFF, Combining), (0x1B00, 0x1B7F, Balinese), (0x1B80, 0x1BBF, Sundanese), (0x1BC0, 0x1BFF, Batak), (0x1C00, 0x1C4F, Lepcha), (0x1C50, 0x1C7F, Ol),
(0x1CC0, 0x1CCF, Sundanese), (0x1CD0, 0x1CFF, Vedic), (0x1D00, 0x1DBF, Phonetic), (0x1DC0, 0x1DFF, Combining), (0x1E00, 0x1EFF, Latin), (0x1F00, 0x1FFF, Greek), (0x2000, 0x206F, General), (0x2070, 0x209F, Superscripts), (0x20A0, 0x20CF, Currency), (0x20D0, 0x20FF, Combining), (0x2100, 0x214F, Letterlike), (0x2150, 0x218F, Number), (0x2190, 0x21FF, Arrows), (0x2200, 0x22FF, Mathematical), (0x2300, 0x23FF, Miscellaneous), (0x2400, 0x243F, Control), (0x2440, 0x245F, Optical), (0x2460, 0x24FF, Enclosed), (0x2500, 0x257F, Box), (0x2580, 0x259F, Block), (0x25A0, 0x25FF, Geometric), (0x2600, 0x26FF, Miscellaneous), (0x2700, 0x27BF, Dingbats), (0x27C0, 0x27EF, Miscellaneous), (0x27F0, 0x27FF, Supplemental), (0x2800, 0x28FF, Braille), (0x2900, 0x297F, Supplemental), (0x2980, 0x29FF, Miscellaneous), (0x2A00, 0x2AFF, Supplemental), (0x2B00, 0x2BFF, Miscellaneous), (0x2C00, 0x2C5F, Glagolitic), (0x2C60, 0x2C7F, Latin), (0x2C80, 0x2CFF, Coptic), (0x2D00, 0x2D2F, Georgian), (0x2D30, 0x2D7F, Tifinagh), (0x2D80, 0x2DDF, Ethiopic), (0x2DE0, 0x2DFF, Cyrillic), (0x2E00, 0x2E7F, Supplemental), (0x2E80, 0x2EFF, CJK), (0x2F00, 0x2FDF, Kangxi), (0x2FF0, 0x2FFF, Ideographic), (0x3000, 0x303F, CJK), (0x3040, 0x309F, Hiragana), (0x30A0, 0x30FF, Katakana), (0x3100, 0x312F, Bopomofo), (0x3130, 0x318F, Hangul), (0x3190, 0x319F, Kanbun), (0x31A0, 0x31BF, Bopomofo), (0x31C0, 0x31EF, CJK), (0x31F0, 0x31FF, Katakana), (0x3200, 0x32FF, Enclosed), (0x3300, 0x4DBF, CJK), (0x4DC0, 0x4DFF, Yijing), (0x4E00, 0x9FFF, CJK), (0xA000, 0xA4CF, Yi), (0xA4D0, 0xA4FF, Lisu), (0xA500, 0xA63F, Vai), (0xA640, 0xA69F, Cyrillic), (0xA6A0, 0xA6FF, Bamum), (0xA700, 0xA71F, Modifier), (0xA720, 0xA7FF, Latin), (0xA800, 0xA82F, Syloti), (0xA830, 0xA83F, Common), (0xA840, 0xA87F, Phags), (0xA880, 0xA8DF, Saurashtra), (0xA8E0, 0xA8FF, Devanagari), (0xA900, 0xA92F, Kayah), (0xA930, 0xA95F, Rejang), (0xA960, 0xA97F, Hangul), (0xA980, 0xA9DF, Javanese), (0xA9E0, 0xA9FF, Myanmar), (0xAA00, 0xAA5F, Cham), (0xAA60, 0xAA7F, Myanmar), (0xAA80, 0xAADF, Tai), (0xAAE0, 0xAAFF, Meetei), (0xAB00, 0xAB2F, Ethiopic), (0xAB30, 0xAB6F, Latin), (0xAB70, 0xABBF, Cherokee), (0xABC0, 0xABFF, Meetei), (0xAC00, 0xD7FF, Hangul), (0xD800, 0xDBFF, High), (0xDC00, 0xDFFF, Low), (0xE000, 0xF8FF, Private), (0xF900, 0xFAFF, CJK), (0xFB00, 0xFB4F, Alphabetic), (0xFB50, 0xFDFF, Arabic), (0xFE00, 0xFE0F, Variation), (0xFE10, 0xFE1F, Vertical), (0xFE20, 0xFE2F, Combining), (0xFE30, 0xFE4F, CJK), (0xFE50, 0xFE6F, Small), (0xFE70, 0xFEFF, Arabic), (0xFF00, 0xFFEF, Halfwidth), (0xFFF0, 0xFFFF, Specials), ]; const PLANE0_LIMIT: char = '\u{10000}'; const SUPPLEMENTARY_TABLE: LookupTable<u32, Block> = lookup_table![ (0x010000, 0x0100FF, Linear), (0x010100, 0x01013F, Aegean), (0x010140, 0x0101CF, Ancient), (0x0101D0, 0x0101FF, Phaistos), (0x
(0x1C80, 0x1C8F, Cyrillic), (0x1C90, 0x1CBF, Georgian),
random_line_split
block.rs
SUPPLEMENTARY_TABLE.validate(); } const ROW0_TABLE: LookupTable<u8, Block> = lookup_table![ (0x00, 0x7F, Basic), ]; const ROW0_LIMIT: char = '\u{80}'; const PLANE0_TABLE: LookupTable<u16, Block> = lookup_table![ (0x0080, 0x024F, Latin), (0x0250, 0x02AF, IPA), (0x02B0, 0x02FF, Spacing), (0x0300, 0x036F, Combining), (0x0370, 0x03FF, Greek), (0x0400, 0x052F, Cyrillic), (0x0530, 0x058F, Armenian), (0x0590, 0x05FF, Hebrew), (0x0600, 0x06FF, Arabic), (0x0700, 0x074F, Syriac), (0x0750, 0x077F, Arabic), (0x0780, 0x07BF, Thaana), (0x07C0, 0x07FF, NKo), (0x0800, 0x083F, Samaritan), (0x0840, 0x085F, Mandaic), (0x0860, 0x086F, Syriac), (0x08A0, 0x08FF, Arabic), (0x0900, 0x097F, Devanagari), (0x0980, 0x09FF, Bengali), (0x0A00, 0x0A7F, Gurmukhi), (0x0A80, 0x0AFF, Gujarati), (0x0B00, 0x0B7F, Oriya), (0x0B80, 0x0BFF, Tamil), (0x0C00, 0x0C7F, Telugu), (0x0C80, 0x0CFF, Kannada), (0x0D00, 0x0D7F, Malayalam), (0x0D80, 0x0DFF, Sinhala), (0x0E00, 0x0E7F, Thai), (0x0E80, 0x0EFF, Lao), (0x0F00, 0x0FFF, Tibetan), (0x1000, 0x109F, Myanmar), (0x10A0, 0x10FF, Georgian), (0x1100, 0x11FF, Hangul), (0x1200, 0x139F, Ethiopic), (0x13A0, 0x13FF, Cherokee), (0x1400, 0x167F, Unified), (0x1680, 0x169F, Ogham), (0x16A0, 0x16FF, Runic), (0x1700, 0x171F, Tagalog), (0x1720, 0x173F, Hanunoo), (0x1740, 0x175F, Buhid), (0x1760, 0x177F, Tagbanwa), (0x1780, 0x17FF, Khmer), (0x1800, 0x18AF, Mongolian), (0x18B0, 0x18FF, Unified), (0x1900, 0x194F, Limbu), (0x1950, 0x197F, Tai), (0x1980, 0x19DF, New), (0x19E0, 0x19FF, Khmer), (0x1A00, 0x1A1F, Buginese), (0x1A20, 0x1AAF, Tai), (0x1AB0, 0x1AFF, Combining), (0x1B00, 0x1B7F, Balinese), (0x1B80, 0x1BBF, Sundanese), (0x1BC0, 0x1BFF, Batak), (0x1C00, 0x1C4F, Lepcha), (0x1C50, 0x1C7F, Ol), (0x1C80, 0x1C8F, Cyrillic), (0x1C90, 0x1CBF, Georgian), (0x1CC0, 0x1CCF, Sundanese), (0x1CD0, 0x1CFF, Vedic), (0x1D00, 0x1DBF, Phonetic), (0x1DC0, 0x1DFF, Combining), (0x1E00, 0x1EFF, Latin), (0x1F00, 0x1FFF, Greek), (0x2000, 0x206F, General), (0x2070, 0x209F, Superscripts), (0x20A0, 0x20CF, Currency), (0x20D0, 0x20FF, Combining), (0x2100, 0x214F, Letterlike), (0x2150, 0x218F, Number), (0x2190, 0x21FF, Arrows), (0x2200, 0x22FF, Mathematical), (0x2300, 0x23FF, Miscellaneous), (0x2400, 0x243F, Control), (0x2440, 0x245F, Optical), (0x2460, 0x24FF, Enclosed), (0x2500, 0x257F, Box), (0x2580, 0x259F, Block), (0x25A0, 0x25FF, Geometric), (0x2600, 0x26FF, Miscellaneous), (0x2700, 0x27BF, Dingbats), (0x27C0, 0x27EF, Miscellaneous), (0x27F0, 0x27FF, Supplemental), (0x2800, 0x28FF, Braille), (0x2900, 0x297F, Supplemental), (0x2980, 0x29FF, Miscellaneous), (0x2A00, 0x2AFF, Supplemental), (0x2B00, 0x2BFF, Miscellaneous), (0x2C00, 0x2C5F, Glagolitic), (0x2C60, 0x2C7F, Latin), (0x2C80, 0x2CFF, Coptic), (0x2D00, 0x2D2F, Georgian), (0x2D30, 0x2D7F, Tifinagh), (0x2D80, 0x2DDF, Ethiopic), (0x2DE0, 0x2DFF, Cyrillic), (0x2E00, 0x2E7F, Supplemental), (0x2E80, 0x2EFF, CJK), (0x2F00, 0x2FDF, Kangxi), (0x2FF0, 0x2FFF, Ideographic), (0x3000, 0x303F, CJK), (0x3040, 0x309F, Hiragana), (0x30A0, 0x
{ assert!(!PLANE0_TABLE.contains(&x)); }
conditional_block
block.rs
() { use std::convert::TryInto; ROW0_TABLE.validate(); if let Ok(x) = (ROW0_LIMIT as u32).try_into() { assert!(!ROW0_TABLE.contains(&x)); } PLANE0_TABLE.validate(); if let Ok(x) = (PLANE0_LIMIT as u32).try_into() { assert!(!PLANE0_TABLE.contains(&x)); } SUPPLEMENTARY_TABLE.validate(); } const ROW0_TABLE: LookupTable<u8, Block> = lookup_table![ (0x00, 0x7F, Basic), ]; const ROW0_LIMIT: char = '\u{80}'; const PLANE0_TABLE: LookupTable<u16, Block> = lookup_table![ (0x0080, 0x024F, Latin), (0x0250, 0x02AF, IPA), (0x02B0, 0x02FF, Spacing), (0x0300, 0x036F, Combining), (0x0370, 0x03FF, Greek), (0x0400, 0x052F, Cyrillic), (0x0530, 0x058F, Armenian), (0x0590, 0x05FF, Hebrew), (0x0600, 0x06FF, Arabic), (0x0700, 0x074F, Syriac), (0x0750, 0x077F, Arabic), (0x0780, 0x07BF, Thaana), (0x07C0, 0x07FF, NKo), (0x0800, 0x083F, Samaritan), (0x0840, 0x085F, Mandaic), (0x0860, 0x086F, Syriac), (0x08A0, 0x08FF, Arabic), (0x0900, 0x097F, Devanagari), (0x0980, 0x09FF, Bengali), (0x0A00, 0x0A7F, Gurmukhi), (0x0A80, 0x0AFF, Gujarati), (0x0B00, 0x0B7F, Oriya), (0x0B80, 0x0BFF, Tamil), (0x0C00, 0x0C7F, Telugu), (0x0C80, 0x0CFF, Kannada), (0x0D00, 0x0D7F, Malayalam), (0x0D80, 0x0DFF, Sinhala), (0x0E00, 0x0E7F, Thai), (0x0E80, 0x0EFF, Lao), (0x0F00, 0x0FFF, Tibetan), (0x1000, 0x109F, Myanmar), (0x10A0, 0x10FF, Georgian), (0x1100, 0x11FF, Hangul), (0x1200, 0x139F, Ethiopic), (0x13A0, 0x13FF, Cherokee), (0x1400, 0x167F, Unified), (0x1680, 0x169F, Ogham), (0x16A0, 0x16FF, Runic), (0x1700, 0x171F, Tagalog), (0x1720, 0x173F, Hanunoo), (0x1740, 0x175F, Buhid), (0x1760, 0x177F, Tagbanwa), (0x1780, 0x17FF, Khmer), (0x1800, 0x18AF, Mongolian), (0x18B0, 0x18FF, Unified), (0x1900, 0x194F, Limbu), (0x1950, 0x197F, Tai), (0x1980, 0x19DF, New), (0x19E0, 0x19FF, Khmer), (0x1A00, 0x1A1F, Buginese), (0x1A20, 0x1AAF, Tai), (0x1AB0, 0x1AFF, Combining), (0x1B00, 0x1B7F, Balinese), (0x1B80, 0x1BBF, Sundanese), (0x1BC0, 0x1BFF, Batak), (0x1C00, 0x1C4F, Lepcha), (0x1C50, 0x1C7F, Ol), (0x1C80, 0x1C8F, Cyrillic), (0x1C90, 0x1CBF, Georgian), (0x1CC0, 0x1CCF, Sundanese), (0x1CD0, 0x1CFF, Vedic), (0x1D00, 0x1DBF, Phonetic), (0x1DC0, 0x1DFF, Combining), (0x1E00, 0x1EFF, Latin), (0x1F00, 0x1FFF, Greek), (0x2000, 0x206F, General), (0x2070, 0x209F, Superscripts), (0x20A0, 0x20CF, Currency), (0x20D0, 0x20FF, Combining), (0x2100, 0x214F, Letterlike), (0x2150, 0x218F, Number), (0x2190, 0x21FF, Arrows), (0x2200, 0x22FF, Mathematical), (0x2300, 0x23FF, Miscellaneous), (0x2400, 0x243F, Control), (0x2440, 0x245F, Optical), (0x2460, 0x24FF, Enclosed), (0x2500, 0x257F, Box), (0x2580, 0x259F, Block), (0x25A0, 0x25FF, Geometric), (0x2600, 0x26FF, Miscellaneous), (0x2700, 0x27BF, Dingbats), (0x27C0, 0x27EF, Miscellaneous), (0x27F0, 0x27FF, Supplemental), (0x2800, 0x28FF, Braille), (0x2900, 0x297F, Supplemental), (0x2980, 0x29FF, Miscellaneous), (0x2A00, 0x2AFF, Supplemental), (0x2B00, 0x2BFF, Miscellaneous), (0x2C00, 0x2C5F, Glagolitic), (0x2C60, 0x2C7F, Latin), (0x2C80, 0x2CFF, Coptic), (0x2D00, 0x2D2F, Georgian), (0x2D30, 0x2D7F, Tifinagh), (0x2D80, 0x2DDF, Ethiopic), (0x2DE0, 0x2DFF, Cyrillic), (0x2E00, 0x2E7F, Supplemental), (0x2E80, 0x2EFF, CJK), (0x2F00, 0x2
validate_tables
identifier_name
drawing.go
{c[0], a[1]}, f32.Vec2{a[0], c[1]} uvb, uvd := f32.Vec2{uvc[0], uva[1]}, f32.Vec2{uva[0], uvc[1]} dl.VtxWriter[0] = DrawVert{a, uva, color} dl.VtxWriter[1] = DrawVert{b, uvb, color} dl.VtxWriter[2] = DrawVert{c, uvc, color} dl.VtxWriter[3] = DrawVert{d, uvd, color} ii := dl.vtxIndex dl.IdxWriter[0] = DrawIdx(ii+0) dl.IdxWriter[1] = DrawIdx(ii+1) dl.IdxWriter[2] = DrawIdx(ii+2) dl.IdxWriter[3] = DrawIdx(ii+0) dl.IdxWriter[4] = DrawIdx(ii+2) dl.IdxWriter[5] = DrawIdx(ii+3) dl.idxIndex += 6 dl.vtxIndex += 4 } func (dl *DrawList) PrimQuadUV(a, b, c, d f32.Vec2, uva, uvb,uvc, uvd f32.Vec2, color uint32) { // vertex dl.VtxWriter[0] = DrawVert{a, uva, color} dl.VtxWriter[1] = DrawVert{b, uvb, color} dl.VtxWriter[2] = DrawVert{c, uvc, color} dl.VtxWriter[3] = DrawVert{d, uvd, color} ii := dl.vtxIndex dl.IdxWriter[0] = DrawIdx(ii+0) dl.IdxWriter[1] = DrawIdx(ii+1) dl.IdxWriter[2] = DrawIdx(ii+2) dl.IdxWriter[3] = DrawIdx(ii+0) dl.IdxWriter[4] = DrawIdx(ii+2) dl.IdxWriter[5] = DrawIdx(ii+3) dl.vtxIndex += 4 dl.idxIndex += 6 } // 此处生成最终的顶点数据和索引数据 // 当前并不支持抗锯齿!!简单的用顶点生成线段 func (dl *DrawList) AddPolyLine(points []f32.Vec2, color uint32, thickness float32, closed bool) { pointsCount := len(points) if pointsCount < 2 { return } uv := dl.TexUVWhitePixel count := pointsCount if !closed { count = pointsCount - 1 } // Non Anti-aliased Stroke idxCount := count * 6 vtxCount := count * 4 dl.PrimReserve(idxCount, vtxCount) for i1 := 0; i1 < count; i1 ++{ i2 := i1 + 1 if i2 == pointsCount { i2 = 0 } p1, p2 := points[i1], points[i2] diff := p2.Sub(p1) invLength := math.InvLength(diff[0], diff[1], 1.0) diff = diff.Mul(invLength) dx := diff[0] * (thickness * 0.5) dy := diff[1] * (thickness * 0.5) vi := i1*4 dl.VtxWriter[vi+0] = DrawVert{f32.Vec2{p1[0]+dy, p1[1]-dx}, uv, color} dl.VtxWriter[vi+1] = DrawVert{f32.Vec2{p2[0]+dy, p2[1]-dx}, uv, color} dl.VtxWriter[vi+2] = DrawVert{f32.Vec2{p2[0]-dy, p2[1]+dx}, uv, color} dl.VtxWriter[vi+3] = DrawVert{f32.Vec2{p1[0]-dy, p1[1]+dx}, uv, color} ii := i1*6 dl.IdxWriter[ii+0] = DrawIdx(dl.vtxIndex+0) dl.IdxWriter[ii+1] = DrawIdx(dl.vtxIndex+1) dl.IdxWriter[ii+2] = DrawIdx(dl.vtxIndex+2) dl.IdxWriter[ii+3] = DrawIdx(dl.vtxIndex+0) dl.IdxWriter[ii+4] = DrawIdx(dl.vtxIndex+2) dl.IdxWriter[ii+5] = DrawIdx(dl.vtxIndex+3) dl.vtxIndex += 4 dl.idxIndex += 6 } dl.AddCommand(idxCount) } // Non Anti-aliased Fill func (dl *DrawList) AddConvexPolyFilled(points []f32.Vec2, color uint32) { uv := dl.TexUVWhitePixel pointCount := len(points) idxCount := (pointCount-2)*3 vtxCount := pointCount dl.PrimReserve(idxCount, vtxCount) for i := 0; i < vtxCount; i++ { dl.VtxWriter[i] = DrawVert{points[i], uv, color} } for i, ii := 2, 0; i < pointCount; i, ii = i+1, ii+3 { dl.IdxWriter[ii+0] = DrawIdx(dl.vtxIndex+0) dl.IdxWriter[ii+1] = DrawIdx(dl.vtxIndex+i-1) dl.IdxWriter[ii+2] = DrawIdx(dl.vtxIndex+i) } dl.vtxIndex += vtxCount dl.idxIndex += idxCount dl.AddCommand(idxCount) } // 此处圆角的算法: // 使用一个12边形近似圆形,采用中心放射算法,计算出 // 各个角度的sin/cos, 然后通过公式,得到圆圆形顶点 // f(x) = centre.x + cos()*radius // f(y) = centre.y + sin()*radius // 以上, 可以提前算好 sin/cos 加速整个过程 func (dl *DrawList) PathArcToFast(centre f32.Vec2, radius float32, min12, max12 int) { if radius == 0 || min12 > max12 { dl.path[dl.pathUsed] = centre; dl.pathUsed ++ return } for a := min12; a <= max12; a++ { x := centre[0] + dl.CircleVtx12[a%12][0] * radius y := centre[1] + dl.CircleVtx12[a%12][1] * radius dl.path[dl.pathUsed] = f32.Vec2{x, y} dl.pathUsed ++ } } func (dl *DrawList) PathArcTo(centre f32.Vec2, radius float32, min, max float32, segments int) { if radius == 0 { dl.path[dl.pathUsed] = centre; dl.pathUsed++ return } for i := 0; i <= segments; i++ { a := min + (float32(i)/float32(segments)) * (max-min) x := centre[0] + math.Cos(a) * radius y := centre[1] + math.Sin(a) * radius dl.path[dl.pathUsed] = f32.Vec2{x, y} dl.pathUsed ++ } } func (dl *DrawList) PathBezierCurveTo(p2, p3, p4 f32.Vec2, segments int) { } func (dl *DrawList) PathRect(a, b f32.Vec2, rounding float32, corners FlagCorner) { if rounding <= 0 || corners == FlagCornerNone { dl.PathLineTo(a) dl.PathLineTo(f32.Vec2{b[0], a[1]}) dl.PathLineTo(b) dl.PathLineTo(f32.Vec2{a[0], b[1]}) } else { var bl, br, tr, tl float32 if (corners & FlagCornerBottomLeft) != 0 { bl = rounding } if (corners & FlagCornerBottomRight) != 0 { br = rounding } if (corners & FlagCornerTopRight) != 0 { tr = rounding } if (corners & FlagCornerTopLeft) != 0 { tl = rounding } dl.PathArcToFast(f32.Vec2{a[0]+bl, a[1]+bl}, bl, 6, 9) // bottom-left dl.PathArcToFast(f
32.Vec2{b[0]-br, a[1]+
conditional_block
drawing.go
ndex = 0 dl.vtxIndex = 0 } func (dl *DrawList) PathClear() { dl.pathUsed = 0 } func (dl *DrawList) PathLineTo(pos f32.Vec2) { if n := len(dl.path); dl.pathUsed < n-1 { dl.path[dl.pathUsed] = pos dl.pathUsed += 1 } } func (dl *DrawList) PathLineToMergeDuplicate(pos f32.Vec2) { //if (_Path.Size == 0 || memcmp(&_Path[_Path.Size-1], &pos, 8) != 0) // _Path.push_back(pos); } func (dl *DrawList) PathFillConvex(col uint32) { dl.AddConvexPolyFilled(dl.path[:dl.pathUsed], col); dl.pathUsed = 0 } // default: thickness=1.0 func (dl *DrawList) PathStroke(color uint32, thickness float32, closed bool) { dl.AddPolyLine(dl.path[:dl.pathUsed], color, thickness, closed) dl.PathClear() } func (dl *DrawList) CurrentClipRect() (clip f32.Vec4) { if n := len(dl.ClipRectStack); n > 0 { clip = dl.ClipRectStack[n-1] } else { clip = dl.FullScreen } return } func (dl *DrawList) CurrentTextureId() (id uint16) { if n := len(dl.TextureIdStack); n > 0 { id = dl.TextureIdStack[n-1] } return } // will result in new draw-call func (dl *DrawList) UpdateClipRect() { //clip := dl.CurrentClipRect() } func (dl *DrawList) UpdateTextureId() { } // Clip 相关的操作 func (dl *DrawList) PushClipRect(min, max f32.Vec2, intersectCurrentClip bool) { cr := f32.Vec4{min[0], min[1], max[0], max[1]} if intersectCurrentClip && len(dl.ClipRectStack) > 0{ current := dl.ClipRectStack[len(dl.ClipRectStack)-1] if cr[0] < current[0] { cr[0] = current[0] } if cr[1] < current[1] { cr[1] = current[1] } if cr[2] > current[2] { cr[2] = current[2] } if cr[3] > current[3] { cr[3] = current[3] } cr[2] = math.Max(cr[0], cr[2]) cr[3] = math.Max(cr[1], cr[3]) dl.ClipRectStack = append(dl.ClipRectStack, cr) dl.UpdateClipRect() } } func (dl *DrawList) PushClipRectFullScreen() { min := f32.Vec2{dl.FullScreen[0], dl.FullScreen[1]} max := f32.Vec2{dl.FullScreen[2], dl.FullScreen[3]} dl.PushClipRect(min, max, false) } func (dl *DrawList) PopClipRect() { if n := len(dl.ClipRectStack); n > 0 { dl.ClipRectStack = dl.ClipRectStack[:n-1] } } func (dl *DrawList) GetClipRectMin() f32.Vec2 { return f32.Vec2{0, 0 } } func (dl *DrawList) GetClipRectMax() f32.Vec2 { return f32.Vec2{0, 0 } } func (dl *DrawList) PushTextureId(texId uint16) { dl.TextureIdStack = append(dl.TextureIdStack, texId) } func (dl *DrawList) PopTextureId() { if n := len(dl.TextureIdStack); n > 0 { dl.TextureIdStack = dl.TextureIdStack[:n-1] } } // primitive operation, auto scale by 1024 func (dl *DrawList) PrimReserve(idxCount, vtxCount int) { if sz, require := len(dl.VtxBuffer), dl.vtxIndex+vtxCount; require >= sz { vtxBuffer := make([]DrawVert, sz+1024) copy(vtxBuffer, dl.VtxBuffer) dl.VtxBuffer = vtxBuffer } if sz, require := len(dl.IdxBuffer), dl.idxIndex+idxCount; require >= sz { idxBuffer := make([]DrawIdx, sz+1024) copy(idxBuffer, dl.IdxBuffer) dl.IdxBuffer = idxBuffer } dl.VtxWriter = dl.VtxBuffer[dl.vtxIndex:dl.vtxIndex+vtxCount] dl.IdxWriter = dl.IdxBuffer[dl.idxIndex:dl.idxIndex+idxCount] } func (dl *DrawList) PrimRect(min, max f32.Vec2, color uint32) { uv := dl.TexUVWhitePixel a, b, c, d := min, f32.Vec2{max[0], min[1]}, max, f32.Vec2{min[0], max[1]} dl.VtxWriter[0] = DrawVert{a, uv, color} dl.VtxWriter[1] = DrawVert{b, uv, color} dl.VtxWriter[2] = DrawVert{c, uv, color} dl.VtxWriter[3] = DrawVert{d, uv, color} dl.IdxWriter[0] = DrawIdx(dl.vtxIndex+0) dl.IdxWriter[1] = DrawIdx(dl.vtxIndex+1) dl.IdxWriter[2] = DrawIdx(dl.vtxIndex+2) dl.IdxWriter[3] = DrawIdx(dl.vtxIndex+0) dl.IdxWriter[4] = DrawIdx(dl.vtxIndex+2) dl.IdxWriter[5] = DrawIdx(dl.vtxIndex+3) dl.vtxIndex += 4 dl.idxIndex += 6 } func (dl *DrawList) PrimRectUV(a, c f32.Vec2, uva, uvc f32.Vec2, color uint32) { b, d := f32.Vec2{c[0], a[1]}, f32.Vec2{a[0], c[1]} uvb, uvd := f32.Vec2{uvc[0], uva[1]}, f32.Vec2{uva[0], uvc[1]} dl.VtxWriter[0] = DrawVert{a, uva, color} dl.VtxWriter[1] = DrawVert{b, uvb, color} dl.VtxWriter[2] = DrawVert{c, uvc, color} dl.VtxWriter[3] = DrawVert{d, uvd, color} ii := dl.vtxIndex dl.IdxWriter[0] = DrawIdx(ii+0) dl.IdxWriter[1] = DrawIdx(ii+1) dl.IdxWriter[2] = DrawIdx(ii+2) dl.IdxWriter[3] = DrawIdx(ii+0) dl.IdxWriter[4] = DrawIdx(ii+2) dl.IdxWriter[5] = DrawIdx(ii+3) dl.idxIndex += 6 dl.vtxIndex += 4 } func (dl *DrawList) PrimQuadUV(a, b, c, d f32.Vec2, uva, uvb,uvc, uvd f32.Vec2, color uint32) { // vertex dl.VtxWriter[0] = DrawVert{a, uva, color} dl.VtxWriter[1] = DrawVert{b, uvb, color} dl.VtxWriter[2] = DrawVert{c, uvc, color} dl.VtxWriter[3] = DrawVert{d, uvd, color} ii := dl.vtxIndex dl.IdxWriter[0] = DrawIdx(ii+0) dl.IdxWriter[1] = DrawIdx(ii+1) dl.IdxWriter[2] = DrawIdx(ii+2) dl.IdxWriter[3] = DrawIdx(ii+0) dl.IdxWriter[4] = DrawIdx(ii+2) dl.IdxWriter[5] = DrawIdx(ii+3) dl.vtxIndex += 4 dl.idxIndex += 6 } // 此处生成最终的顶点数据和索引数据 // 当前并不支持抗锯齿!!简单的用顶点生成线段 func (dl *DrawList) AddPolyLine(points []f32.Vec2, color uint32, thickness float32, closed bool) { pointsCount := len(points) if pointsCount < 2 { return } uv := dl.TexUVWhitePixel count := pointsCount if !closed { count = points
.idxI
identifier_name
drawing.go
.Vec2 { return f32.Vec2{0, 0 } } func (dl *DrawList) GetClipRectMax() f32.Vec2 { return f32.Vec2{0, 0 } } func (dl *DrawList) PushTextureId(texId uint16) { dl.TextureIdStack = append(dl.TextureIdStack, texId) }
} } // primitive operation, auto scale by 1024 func (dl *DrawList) PrimReserve(idxCount, vtxCount int) { if sz, require := len(dl.VtxBuffer), dl.vtxIndex+vtxCount; require >= sz { vtxBuffer := make([]DrawVert, sz+1024) copy(vtxBuffer, dl.VtxBuffer) dl.VtxBuffer = vtxBuffer } if sz, require := len(dl.IdxBuffer), dl.idxIndex+idxCount; require >= sz { idxBuffer := make([]DrawIdx, sz+1024) copy(idxBuffer, dl.IdxBuffer) dl.IdxBuffer = idxBuffer } dl.VtxWriter = dl.VtxBuffer[dl.vtxIndex:dl.vtxIndex+vtxCount] dl.IdxWriter = dl.IdxBuffer[dl.idxIndex:dl.idxIndex+idxCount] } func (dl *DrawList) PrimRect(min, max f32.Vec2, color uint32) { uv := dl.TexUVWhitePixel a, b, c, d := min, f32.Vec2{max[0], min[1]}, max, f32.Vec2{min[0], max[1]} dl.VtxWriter[0] = DrawVert{a, uv, color} dl.VtxWriter[1] = DrawVert{b, uv, color} dl.VtxWriter[2] = DrawVert{c, uv, color} dl.VtxWriter[3] = DrawVert{d, uv, color} dl.IdxWriter[0] = DrawIdx(dl.vtxIndex+0) dl.IdxWriter[1] = DrawIdx(dl.vtxIndex+1) dl.IdxWriter[2] = DrawIdx(dl.vtxIndex+2) dl.IdxWriter[3] = DrawIdx(dl.vtxIndex+0) dl.IdxWriter[4] = DrawIdx(dl.vtxIndex+2) dl.IdxWriter[5] = DrawIdx(dl.vtxIndex+3) dl.vtxIndex += 4 dl.idxIndex += 6 } func (dl *DrawList) PrimRectUV(a, c f32.Vec2, uva, uvc f32.Vec2, color uint32) { b, d := f32.Vec2{c[0], a[1]}, f32.Vec2{a[0], c[1]} uvb, uvd := f32.Vec2{uvc[0], uva[1]}, f32.Vec2{uva[0], uvc[1]} dl.VtxWriter[0] = DrawVert{a, uva, color} dl.VtxWriter[1] = DrawVert{b, uvb, color} dl.VtxWriter[2] = DrawVert{c, uvc, color} dl.VtxWriter[3] = DrawVert{d, uvd, color} ii := dl.vtxIndex dl.IdxWriter[0] = DrawIdx(ii+0) dl.IdxWriter[1] = DrawIdx(ii+1) dl.IdxWriter[2] = DrawIdx(ii+2) dl.IdxWriter[3] = DrawIdx(ii+0) dl.IdxWriter[4] = DrawIdx(ii+2) dl.IdxWriter[5] = DrawIdx(ii+3) dl.idxIndex += 6 dl.vtxIndex += 4 } func (dl *DrawList) PrimQuadUV(a, b, c, d f32.Vec2, uva, uvb,uvc, uvd f32.Vec2, color uint32) { // vertex dl.VtxWriter[0] = DrawVert{a, uva, color} dl.VtxWriter[1] = DrawVert{b, uvb, color} dl.VtxWriter[2] = DrawVert{c, uvc, color} dl.VtxWriter[3] = DrawVert{d, uvd, color} ii := dl.vtxIndex dl.IdxWriter[0] = DrawIdx(ii+0) dl.IdxWriter[1] = DrawIdx(ii+1) dl.IdxWriter[2] = DrawIdx(ii+2) dl.IdxWriter[3] = DrawIdx(ii+0) dl.IdxWriter[4] = DrawIdx(ii+2) dl.IdxWriter[5] = DrawIdx(ii+3) dl.vtxIndex += 4 dl.idxIndex += 6 } // 此处生成最终的顶点数据和索引数据 // 当前并不支持抗锯齿!!简单的用顶点生成线段 func (dl *DrawList) AddPolyLine(points []f32.Vec2, color uint32, thickness float32, closed bool) { pointsCount := len(points) if pointsCount < 2 { return } uv := dl.TexUVWhitePixel count := pointsCount if !closed { count = pointsCount - 1 } // Non Anti-aliased Stroke idxCount := count * 6 vtxCount := count * 4 dl.PrimReserve(idxCount, vtxCount) for i1 := 0; i1 < count; i1 ++{ i2 := i1 + 1 if i2 == pointsCount { i2 = 0 } p1, p2 := points[i1], points[i2] diff := p2.Sub(p1) invLength := math.InvLength(diff[0], diff[1], 1.0) diff = diff.Mul(invLength) dx := diff[0] * (thickness * 0.5) dy := diff[1] * (thickness * 0.5) vi := i1*4 dl.VtxWriter[vi+0] = DrawVert{f32.Vec2{p1[0]+dy, p1[1]-dx}, uv, color} dl.VtxWriter[vi+1] = DrawVert{f32.Vec2{p2[0]+dy, p2[1]-dx}, uv, color} dl.VtxWriter[vi+2] = DrawVert{f32.Vec2{p2[0]-dy, p2[1]+dx}, uv, color} dl.VtxWriter[vi+3] = DrawVert{f32.Vec2{p1[0]-dy, p1[1]+dx}, uv, color} ii := i1*6 dl.IdxWriter[ii+0] = DrawIdx(dl.vtxIndex+0) dl.IdxWriter[ii+1] = DrawIdx(dl.vtxIndex+1) dl.IdxWriter[ii+2] = DrawIdx(dl.vtxIndex+2) dl.IdxWriter[ii+3] = DrawIdx(dl.vtxIndex+0) dl.IdxWriter[ii+4] = DrawIdx(dl.vtxIndex+2) dl.IdxWriter[ii+5] = DrawIdx(dl.vtxIndex+3) dl.vtxIndex += 4 dl.idxIndex += 6 } dl.AddCommand(idxCount) } // Non Anti-aliased Fill func (dl *DrawList) AddConvexPolyFilled(points []f32.Vec2, color uint32) { uv := dl.TexUVWhitePixel pointCount := len(points) idxCount := (pointCount-2)*3 vtxCount := pointCount dl.PrimReserve(idxCount, vtxCount) for i := 0; i < vtxCount; i++ { dl.VtxWriter[i] = DrawVert{points[i], uv, color} } for i, ii := 2, 0; i < pointCount; i, ii = i+1, ii+3 { dl.IdxWriter[ii+0] = DrawIdx(dl.vtxIndex+0) dl.IdxWriter[ii+1] = DrawIdx(dl.vtxIndex+i-1) dl.IdxWriter[ii+2] = DrawIdx(dl.vtxIndex+i) } dl.vtxIndex += vtxCount dl.idxIndex += idxCount dl.AddCommand(idxCount) } // 此处圆角的算法: // 使用一个12边形近似圆形,采用中心放射算法,计算出 // 各个角度的sin/cos, 然后通过公式,得到圆圆形顶点 // f(x) = centre.x + cos()*
func (dl *DrawList) PopTextureId() { if n := len(dl.TextureIdStack); n > 0 { dl.TextureIdStack = dl.TextureIdStack[:n-1]
random_line_split
drawing.go
ClipRect() (clip f32.Vec4) { if n := len(dl.ClipRectStack); n > 0 { clip = dl.ClipRectStack[n-1] } else { clip = dl.FullScreen } return } func (dl *DrawList) CurrentTextureId() (id uint16) { if n := len(dl.TextureIdStack); n > 0 { id = dl.TextureIdStack[n-1] } return } // will result in new draw-call func (dl *DrawList) UpdateClipRect() { //clip := dl.CurrentClipRect() } func (dl *DrawList) UpdateTextureId() { } // Clip 相关的操作 func (dl *DrawList) PushClipRect(min, max f32.Vec2, intersectCurrentClip bool) { cr := f32.Vec4{min[0], min[1], max[0], max[1]} if intersectCurrentClip && len(dl.ClipRectStack) > 0{ current := dl.ClipRectStack[len(dl.ClipRectStack)-1] if cr[0] < current[0] { cr[0] = current[0] } if cr[1] < current[1] { cr[1] = current[1] } if cr[2] > current[2] { cr[2] = current[2] } if cr[3] > current[3] { cr[3] = current[3] } cr[2] = math.Max(cr[0], cr[2]) cr[3] = math.Max(cr[1], cr[3]) dl.ClipRectStack = append(dl.ClipRectStack, cr) dl.UpdateClipRect() } } func (dl *DrawList) PushClipRectFullScreen() { min := f32.Vec2{dl.FullScreen[0], dl.FullScreen[1]} max := f32.Vec2{dl.FullScreen[2], dl.FullScreen[3]} dl.PushClipRect(min, max, false) } func (dl *DrawList) PopClipRect() { if n := len(dl.ClipRectStack); n > 0 { dl.ClipRectStack = dl.ClipRectStack[:n-1] } } func (dl *DrawList) GetClipRectMin() f32.Vec2 { return f32.Vec2{0, 0 } } func (dl *DrawList) GetClipRectMax() f32.Vec2 { return f32.Vec2{0, 0 } } func (dl *DrawList) PushTextureId(texId uint16) { dl.TextureIdStack = append(dl.TextureIdStack, texId) } func (dl *DrawList) PopTextureId() { if n := len(dl.TextureIdStack); n > 0 { dl.TextureIdStack = dl.TextureIdStack[:n-1] } } // primitive operation, auto scale by 1024 func (dl *DrawList) PrimReserve(idxCount, vtxCount int) { if sz, require := len(dl.VtxBuffer), dl.vtxIndex+vtxCount; require >= sz { vtxBuffer := make([]DrawVert, sz+1024) copy(vtxBuffer, dl.VtxBuffer) dl.VtxBuffer = vtxBuffer } if sz, require := len(dl.IdxBuffer), dl.idxIndex+idxCount; require >= sz { idxBuffer := make([]DrawIdx, sz+1024) copy(idxBuffer, dl.IdxBuffer) dl.IdxBuffer = idxBuffer } dl.VtxWriter = dl.VtxBuffer[dl.vtxIndex:dl.vtxIndex+vtxCount] dl.IdxWriter = dl.IdxBuffer[dl.idxIndex:dl.idxIndex+idxCount] } func (dl *DrawList) PrimRect(min, max f32.Vec2, color uint32) { uv := dl.TexUVWhitePixel a, b, c, d := min, f32.Vec2{max[0], min[1]}, max, f32.Vec2{min[0], max[1]} dl.VtxWriter[0] = DrawVert{a, uv, color} dl.VtxWriter[1] = DrawVert{b, uv, color} dl.VtxWriter[2] = DrawVert{c, uv, color} dl.VtxWriter[3] = DrawVert{d, uv, color} dl.IdxWriter[0] = DrawIdx(dl.vtxIndex+0) dl.IdxWriter[1] = DrawIdx(dl.vtxIndex+1) dl.IdxWriter[2] = DrawIdx(dl.vtxIndex+2) dl.IdxWriter[3] = DrawIdx(dl.vtxIndex+0) dl.IdxWriter[4] = DrawIdx(dl.vtxIndex+2) dl.IdxWriter[5] = DrawIdx(dl.vtxIndex+3) dl.vtxIndex += 4 dl.idxIndex += 6 } func (dl *DrawList) PrimRectUV(a, c f32.Vec2, uva, uvc f32.Vec2, color uint32) { b, d := f32.Vec2{c[0], a[1]}, f32.Vec2{a[0], c[1]} uvb, uvd := f32.Vec2{uvc[0], uva[1]}, f32.Vec2{uva[0], uvc[1]} dl.VtxWriter[0] = DrawVert{a, uva, color} dl.VtxWriter[1] = DrawVert{b, uvb, color} dl.VtxWriter[2] = DrawVert{c, uvc, color} dl.VtxWriter[3] = DrawVert{d, uvd, color} ii := dl.vtxIndex dl.IdxWriter[0] = DrawIdx(ii+0) dl.IdxWriter[1] = DrawIdx(ii+1) dl.IdxWriter[2] = DrawIdx(ii+2) dl.IdxWriter[3] = DrawIdx(ii+0) dl.IdxWriter[4] = DrawIdx(ii+2) dl.IdxWriter[5] = DrawIdx(ii+3) dl.idxIndex += 6 dl.vtxIndex += 4 } func (dl *DrawList) PrimQuadUV(a, b, c, d f32.Vec2, uva, uvb,uvc, uvd f32.Vec2, color uint32) { // vertex dl.VtxWriter[0] = DrawVert{a, uva, color} dl.VtxWriter[1] = DrawVert{b, uvb, color} dl.VtxWriter[2] = DrawVert{c, uvc, color} dl.VtxWriter[3] = DrawVert{d, uvd, color} ii := dl.vtxIndex dl.IdxWriter[0] = DrawIdx(ii+0) dl.IdxWriter[1] = DrawIdx(ii+1) dl.IdxWriter[2] = DrawIdx(ii+2) dl.IdxWriter[3] = DrawIdx(ii+0) dl.IdxWriter[4] = DrawIdx(ii+2) dl.IdxWriter[5] = DrawIdx(ii+3) dl.vtxIndex += 4 dl.idxIndex += 6 } // 此处生成最终的顶点数据和索引数据 // 当前并不支持抗锯齿!!简单的用顶点生成线段 func (dl *DrawList) AddPolyLine(points []f32.Vec2, color uint32, thickness float32, closed bool) { pointsCount := len(points) if pointsCount < 2 { return } uv := dl.TexUVWhitePixel count := pointsCount if !closed { count = pointsCount - 1 } // Non Anti-aliased Stroke idxCount := count * 6 vtxCount := count * 4 dl.PrimReserve(idxCount, vtxCount) for i1 := 0; i1 < count; i1 ++{ i2 := i1 + 1 if i2 == pointsCount { i2 = 0 } p1, p2 := points[i1], points[i2] diff := p2.Sub(p1) invLength := math.InvLength(diff[0], diff[1], 1.0) diff = diff.Mul(invLength) dx := diff[0] * (thickness * 0.5) dy := diff[1] * (thickness * 0.5) vi := i1*4 dl.VtxWriter[vi+0] = DrawVert{f32.Vec2{p1[0]+dy, p1[1]-dx}, uv, color} dl.VtxWriter[vi+1] = DrawVert{f
pathUsed], color, thickness, closed) dl.PathClear() } func (dl *DrawList) Current
identifier_body
server.py
@app.route("/users") def user_list(): """Show list of users.""" users = User.query.all() return render_template("user_list.html", users=users) # This takes to each user's profile from user list @app.route("/users/<int:user_id>") def user_profile(user_id): """Show user information""" # Query by user id to return that record in database about user info user = User.query.filter(User.user_id == user_id).one() # import pdb; pdb.set_trace() # Query to get all movies and scores rated by this user # Needed to join Rating and Movie tables and filter by user id # Sort movie titles alphabetically user_movies = db.session.query(Rating.user_id, Rating.movie_id, Rating.score, Movie.title).join(Movie).filter(Rating.user_id == user_id).order_by(Movie.title).all() # Passed user info into jinja and called on its attributes # Passed user_movies into jinja and called on its attributes to get the info return render_template("user_profile.html", user=user, user_movies = user_movies) # # THIS WORKS, but we want to use /user/<int:user_id>, which we figured out above!! # @app.route("/user-profile") # def user_profile(): # """Show user information""" # # import pdb; pdb.set_trace() # # Get user email to query in User database and get all info about the user # email = session["logged_in_user_email"] # user = User.query.filter(User.email == email).one() # # # Test code to see attributes of user object # # user_id = user.user_id # # age = user.age # # zipcode = user.zipcode # return render_template("user_profile.html", user=user) @app.route("/signup-login", methods=["GET"]) def show_forms(): """Show signup and login forms.""" return render_template("signup_login.html") @app.route("/signup", methods=["POST"]) def signup(): """Check if user exists in database, otherwise add user to database.""" # Get values from signup form signup_email = request.form.get("signup_email") signup_password = request.form.get("signup_password") # If user exists, ask them to log in # Otherwise, add user into database and log them in, redirecting to homepage if db.session.query(User).filter(User.email == signup_email).first(): flash("You already have an account please use login!", "danger") return redirect("/signup-login") else: new_user = User(email=signup_email, password=signup_password, age=None, zipcode=None) db.session.add(new_user) db.session.commit() session["logged_in_user_email"] = signup_email session["logged_in_user"] = new_user.user_id flash("Your account has been created! You now are logged in!", "success") return redirect("/") @app.route("/login", methods=["POST"]) def login(): """Check if user's email matches password, otherwise ask user to try again.""" # Get values from login form login_email = request.form.get("login_email") login_password = request.form.get("login_password") # If user's email and password matches, log them in, redirecting them to homepage # Otherwise, ask them to log in with the correct password if db.session.query(User).filter(User.email == login_email, User.password == login_password).first(): flash("Login SUCCESS.", "success") # Query to get user's user id, in order to redirect user to their user profile user = User.query.filter(User.email == login_email).one() session["logged_in_user_email"] = login_email session["logged_in_user"] = user.user_id # Pass a variable through a string via string formatting # so we can pass user_id into the redirected route, which is a string!! return redirect("/users/%s" % user.user_id) # return redirect("/") else: flash("Incorrect password. Please try again!", "danger") return redirect("/signup-login") @app.route("/logout") def process_logout(): """Log user out.""" del session["logged_in_user_email"] del session["logged_in_user"] flash("Logged out.", "success") return redirect("/") @app.route("/movies") def movie_list(): """Show list of movies.""" # sort movie titles alphbetically movies = Movie.query.order_by(Movie.title).all() return render_template("movie_list.html", movies=movies) @app.route("/movies/<int:movie_id>", methods=['GET']) def movie_profile(movie_id): """Show movie information. If a user is logged in, let them add/edit a rating. """ if not session.get('logged_in_user_email'): flash("Please login or signup to see the movie details and rate the movie!", "danger") return redirect("/signup-login") else: # import pdb; pdb.set_trace(); # Query by movie id to return that record in database about movie info # movie = Movie.query.filter(Movie.movie_id == movie_id).one() movie = Movie.query.get(movie_id) user = User.query.filter(User.email == session.get("logged_in_user_email")).one() user_id = user.user_id if user_id: user_rating = Rating.query.filter_by(movie_id=movie_id, user_id=user_id).first() else: user_rating = None # Prediction code: only predict if the user hasn't rated it prediction = None if (not user_rating) and user_id: user = User.query.get(user_id) if user: prediction = user.predict_rating(movie) # Either use the prediction or their real rating if prediction: # User hasn't scored; use our prediction if we made one effective_rating = prediction elif user_rating: # User has already scored for real; use that effective_rating = user_rating.score else: # User hasn't scored and we couldn't get a prediction effective_rating = None # Get the wizard's rating, either by predicting or using real rating wizard = User.query.filter_by(email="wizard@gmail.com").one() wizard_rating = Rating.query.filter_by(user_id=wizard.user_id, movie_id=movie.movie_id).first() if wizard_rating is None: wizard_rating = wizard.predict_rating(movie) else: wizard_rating = wizard_rating.score if wizard_rating and effective_rating: difference = abs(wizard_rating - effective_rating) else: # We couldn't get a wizard rating, so we'll skip difference difference = None # Depending on how different we are from the Wizard, choose a message BERATEMENT_MESSAGES = [ "I suppose you don't have such bad taste after all.", "I regret every decision that I've ever made that has brought me to listen to your opinion.", "Words fail me, as your taste in movies has clearly failed you.", "That movie is great. For a clown to watch. Idiot.", "Words cannot express the awfulness of your taste." ] if difference is not None: beratement = BERATEMENT_MESSAGES[int(difference)] else: beratement = None # Tallies score of each rating (how many people rated this score per rating) # Returns list of tuples for count_score unordered_ratings = db.session.query(Rating.score, func.count(Rating.score)).filter(Rating.movie_id == movie_id).group_by(Rating.score) ordered_movies = unordered_ratings.order_by(Rating.score) count_score = ordered_movies.all() # Get average score, which returns a tuple-like object, so need to access index 0 to return the number and pass through jinja avg_rating = db.session.query(func.avg(Rating.score)).filter(Rating.movie_id == movie_id).one() # Query to get all ratings for a specific movie # Needed to join Rating and Movie tables and filter by user id # Sort movie titles alphabetically ratings = db.session.query(Rating.movie_id, Rating.score, Movie.title).join(Movie).filter(Rating.movie_id == movie_id).all() # # Pass user info into jinja and called on its attributes # # Pass count_score, avg_rating, and ratings into jinja # return render_template("movie_profile.html", movie=movie, count_score=count_score, avg_rating=avg_rating[0], ratings=ratings) return render_template( "movie_profile.html", movie=movie, user_rating=user_rating, avg_rating=avg_rating[0], count_score=count_score, prediction=prediction, ratings=ratings, beratement=beratement) @app
"""Homepage.""" # We want user profile link to show if user is logged in and clicks on homepage # Check if logged in and get the value or else return None # If there is a value, query to get user information so that user.user_id can be accessed in jinja # Else, pass None value through so that if statement in jinja not executed user_email = session.get("logged_in_user_email", None) if user_email is not None: user = User.query.filter(User.email == user_email).one() return render_template("homepage.html", user=user) else: return render_template("homepage.html", user=None)
identifier_body
server.py
.""" # We want user profile link to show if user is logged in and clicks on homepage # Check if logged in and get the value or else return None # If there is a value, query to get user information so that user.user_id can be accessed in jinja # Else, pass None value through so that if statement in jinja not executed user_email = session.get("logged_in_user_email", None) if user_email is not None: user = User.query.filter(User.email == user_email).one() return render_template("homepage.html", user=user) else: return render_template("homepage.html", user=None) @app.route("/users") def
(): """Show list of users.""" users = User.query.all() return render_template("user_list.html", users=users) # This takes to each user's profile from user list @app.route("/users/<int:user_id>") def user_profile(user_id): """Show user information""" # Query by user id to return that record in database about user info user = User.query.filter(User.user_id == user_id).one() # import pdb; pdb.set_trace() # Query to get all movies and scores rated by this user # Needed to join Rating and Movie tables and filter by user id # Sort movie titles alphabetically user_movies = db.session.query(Rating.user_id, Rating.movie_id, Rating.score, Movie.title).join(Movie).filter(Rating.user_id == user_id).order_by(Movie.title).all() # Passed user info into jinja and called on its attributes # Passed user_movies into jinja and called on its attributes to get the info return render_template("user_profile.html", user=user, user_movies = user_movies) # # THIS WORKS, but we want to use /user/<int:user_id>, which we figured out above!! # @app.route("/user-profile") # def user_profile(): # """Show user information""" # # import pdb; pdb.set_trace() # # Get user email to query in User database and get all info about the user # email = session["logged_in_user_email"] # user = User.query.filter(User.email == email).one() # # # Test code to see attributes of user object # # user_id = user.user_id # # age = user.age # # zipcode = user.zipcode # return render_template("user_profile.html", user=user) @app.route("/signup-login", methods=["GET"]) def show_forms(): """Show signup and login forms.""" return render_template("signup_login.html") @app.route("/signup", methods=["POST"]) def signup(): """Check if user exists in database, otherwise add user to database.""" # Get values from signup form signup_email = request.form.get("signup_email") signup_password = request.form.get("signup_password") # If user exists, ask them to log in # Otherwise, add user into database and log them in, redirecting to homepage if db.session.query(User).filter(User.email == signup_email).first(): flash("You already have an account please use login!", "danger") return redirect("/signup-login") else: new_user = User(email=signup_email, password=signup_password, age=None, zipcode=None) db.session.add(new_user) db.session.commit() session["logged_in_user_email"] = signup_email session["logged_in_user"] = new_user.user_id flash("Your account has been created! You now are logged in!", "success") return redirect("/") @app.route("/login", methods=["POST"]) def login(): """Check if user's email matches password, otherwise ask user to try again.""" # Get values from login form login_email = request.form.get("login_email") login_password = request.form.get("login_password") # If user's email and password matches, log them in, redirecting them to homepage # Otherwise, ask them to log in with the correct password if db.session.query(User).filter(User.email == login_email, User.password == login_password).first(): flash("Login SUCCESS.", "success") # Query to get user's user id, in order to redirect user to their user profile user = User.query.filter(User.email == login_email).one() session["logged_in_user_email"] = login_email session["logged_in_user"] = user.user_id # Pass a variable through a string via string formatting # so we can pass user_id into the redirected route, which is a string!! return redirect("/users/%s" % user.user_id) # return redirect("/") else: flash("Incorrect password. Please try again!", "danger") return redirect("/signup-login") @app.route("/logout") def process_logout(): """Log user out.""" del session["logged_in_user_email"] del session["logged_in_user"] flash("Logged out.", "success") return redirect("/") @app.route("/movies") def movie_list(): """Show list of movies.""" # sort movie titles alphbetically movies = Movie.query.order_by(Movie.title).all() return render_template("movie_list.html", movies=movies) @app.route("/movies/<int:movie_id>", methods=['GET']) def movie_profile(movie_id): """Show movie information. If a user is logged in, let them add/edit a rating. """ if not session.get('logged_in_user_email'): flash("Please login or signup to see the movie details and rate the movie!", "danger") return redirect("/signup-login") else: # import pdb; pdb.set_trace(); # Query by movie id to return that record in database about movie info # movie = Movie.query.filter(Movie.movie_id == movie_id).one() movie = Movie.query.get(movie_id) user = User.query.filter(User.email == session.get("logged_in_user_email")).one() user_id = user.user_id if user_id: user_rating = Rating.query.filter_by(movie_id=movie_id, user_id=user_id).first() else: user_rating = None # Prediction code: only predict if the user hasn't rated it prediction = None if (not user_rating) and user_id: user = User.query.get(user_id) if user: prediction = user.predict_rating(movie) # Either use the prediction or their real rating if prediction: # User hasn't scored; use our prediction if we made one effective_rating = prediction elif user_rating: # User has already scored for real; use that effective_rating = user_rating.score else: # User hasn't scored and we couldn't get a prediction effective_rating = None # Get the wizard's rating, either by predicting or using real rating wizard = User.query.filter_by(email="wizard@gmail.com").one() wizard_rating = Rating.query.filter_by(user_id=wizard.user_id, movie_id=movie.movie_id).first() if wizard_rating is None: wizard_rating = wizard.predict_rating(movie) else: wizard_rating = wizard_rating.score if wizard_rating and effective_rating: difference = abs(wizard_rating - effective_rating) else: # We couldn't get a wizard rating, so we'll skip difference difference = None # Depending on how different we are from the Wizard, choose a message BERATEMENT_MESSAGES = [ "I suppose you don't have such bad taste after all.", "I regret every decision that I've ever made that has brought me to listen to your opinion.", "Words fail me, as your taste in movies has clearly failed you.", "That movie is great. For a clown to watch. Idiot.", "Words cannot express the awfulness of your taste." ] if difference is not None: beratement = BERATEMENT_MESSAGES[int(difference)] else: beratement = None # Tallies score of each rating (how many people rated this score per rating) # Returns list of tuples for count_score unordered_ratings = db.session.query(Rating.score, func.count(Rating.score)).filter(Rating.movie_id == movie_id).group_by(Rating.score) ordered_movies = unordered_ratings.order_by(Rating.score) count_score = ordered_movies.all() # Get average score, which returns a tuple-like object, so need to access index 0 to return the number and pass through jinja avg_rating = db.session.query(func.avg(Rating.score)).filter(Rating.movie_id == movie_id).one() # Query to get all ratings for a specific movie # Needed to join Rating and Movie tables and filter by user id # Sort movie titles alphabetically ratings = db.session.query(Rating.movie_id, Rating.score, Movie.title).join(Movie).filter(Rating.movie_id == movie_id).all() # # Pass user info into jinja and called on its attributes # # Pass count_score, avg_rating, and ratings into jinja # return render_template("movie_profile.html", movie=movie, count_score=count_score, avg_rating=avg_rating[0], ratings=ratings) return render_template( "movie_profile.html", movie=movie, user_rating=user_rating, avg_rating=avg_rating[0], count_score=count_score, prediction=prediction, ratings=ratings, beratement=beratement) @app.route("/
user_list
identifier_name
server.py
.""" # We want user profile link to show if user is logged in and clicks on homepage # Check if logged in and get the value or else return None # If there is a value, query to get user information so that user.user_id can be accessed in jinja # Else, pass None value through so that if statement in jinja not executed user_email = session.get("logged_in_user_email", None) if user_email is not None: user = User.query.filter(User.email == user_email).one() return render_template("homepage.html", user=user) else: return render_template("homepage.html", user=None) @app.route("/users") def user_list(): """Show list of users.""" users = User.query.all() return render_template("user_list.html", users=users) # This takes to each user's profile from user list @app.route("/users/<int:user_id>") def user_profile(user_id): """Show user information""" # Query by user id to return that record in database about user info user = User.query.filter(User.user_id == user_id).one() # import pdb; pdb.set_trace() # Query to get all movies and scores rated by this user # Needed to join Rating and Movie tables and filter by user id # Sort movie titles alphabetically user_movies = db.session.query(Rating.user_id, Rating.movie_id, Rating.score, Movie.title).join(Movie).filter(Rating.user_id == user_id).order_by(Movie.title).all() # Passed user info into jinja and called on its attributes # Passed user_movies into jinja and called on its attributes to get the info return render_template("user_profile.html", user=user, user_movies = user_movies) # # THIS WORKS, but we want to use /user/<int:user_id>, which we figured out above!! # @app.route("/user-profile") # def user_profile(): # """Show user information""" # # import pdb; pdb.set_trace() # # Get user email to query in User database and get all info about the user # email = session["logged_in_user_email"] # user = User.query.filter(User.email == email).one() # # # Test code to see attributes of user object # # user_id = user.user_id # # age = user.age # # zipcode = user.zipcode # return render_template("user_profile.html", user=user) @app.route("/signup-login", methods=["GET"]) def show_forms(): """Show signup and login forms.""" return render_template("signup_login.html") @app.route("/signup", methods=["POST"]) def signup(): """Check if user exists in database, otherwise add user to database.""" # Get values from signup form signup_email = request.form.get("signup_email") signup_password = request.form.get("signup_password") # If user exists, ask them to log in # Otherwise, add user into database and log them in, redirecting to homepage if db.session.query(User).filter(User.email == signup_email).first(): flash("You already have an account please use login!", "danger") return redirect("/signup-login") else: new_user = User(email=signup_email, password=signup_password, age=None, zipcode=None) db.session.add(new_user) db.session.commit() session["logged_in_user_email"] = signup_email session["logged_in_user"] = new_user.user_id flash("Your account has been created! You now are logged in!", "success") return redirect("/") @app.route("/login", methods=["POST"]) def login(): """Check if user's email matches password, otherwise ask user to try again.""" # Get values from login form login_email = request.form.get("login_email") login_password = request.form.get("login_password") # If user's email and password matches, log them in, redirecting them to homepage # Otherwise, ask them to log in with the correct password if db.session.query(User).filter(User.email == login_email, User.password == login_password).first(): flash("Login SUCCESS.", "success") # Query to get user's user id, in order to redirect user to their user profile user = User.query.filter(User.email == login_email).one() session["logged_in_user_email"] = login_email session["logged_in_user"] = user.user_id # Pass a variable through a string via string formatting # so we can pass user_id into the redirected route, which is a string!! return redirect("/users/%s" % user.user_id) # return redirect("/") else: flash("Incorrect password. Please try again!", "danger") return redirect("/signup-login") @app.route("/logout") def process_logout(): """Log user out.""" del session["logged_in_user_email"] del session["logged_in_user"] flash("Logged out.", "success") return redirect("/") @app.route("/movies") def movie_list(): """Show list of movies.""" # sort movie titles alphbetically movies = Movie.query.order_by(Movie.title).all() return render_template("movie_list.html", movies=movies) @app.route("/movies/<int:movie_id>", methods=['GET']) def movie_profile(movie_id): """Show movie information. If a user is logged in, let them add/edit a rating. """ if not session.get('logged_in_user_email'): flash("Please login or signup to see the movie details and rate the movie!", "danger") return redirect("/signup-login") else: # import pdb; pdb.set_trace(); # Query by movie id to return that record in database about movie info # movie = Movie.query.filter(Movie.movie_id == movie_id).one() movie = Movie.query.get(movie_id) user = User.query.filter(User.email == session.get("logged_in_user_email")).one() user_id = user.user_id if user_id: user_rating = Rating.query.filter_by(movie_id=movie_id, user_id=user_id).first() else: user_rating = None # Prediction code: only predict if the user hasn't rated it prediction = None if (not user_rating) and user_id: user = User.query.get(user_id) if user: prediction = user.predict_rating(movie) # Either use the prediction or their real rating if prediction: # User hasn't scored; use our prediction if we made one effective_rating = prediction elif user_rating: # User has already scored for real; use that effective_rating = user_rating.score else: # User hasn't scored and we couldn't get a prediction effective_rating = None # Get the wizard's rating, either by predicting or using real rating wizard = User.query.filter_by(email="wizard@gmail.com").one() wizard_rating = Rating.query.filter_by(user_id=wizard.user_id, movie_id=movie.movie_id).first() if wizard_rating is None:
else: wizard_rating = wizard_rating.score if wizard_rating and effective_rating: difference = abs(wizard_rating - effective_rating) else: # We couldn't get a wizard rating, so we'll skip difference difference = None # Depending on how different we are from the Wizard, choose a message BERATEMENT_MESSAGES = [ "I suppose you don't have such bad taste after all.", "I regret every decision that I've ever made that has brought me to listen to your opinion.", "Words fail me, as your taste in movies has clearly failed you.", "That movie is great. For a clown to watch. Idiot.", "Words cannot express the awfulness of your taste." ] if difference is not None: beratement = BERATEMENT_MESSAGES[int(difference)] else: beratement = None # Tallies score of each rating (how many people rated this score per rating) # Returns list of tuples for count_score unordered_ratings = db.session.query(Rating.score, func.count(Rating.score)).filter(Rating.movie_id == movie_id).group_by(Rating.score) ordered_movies = unordered_ratings.order_by(Rating.score) count_score = ordered_movies.all() # Get average score, which returns a tuple-like object, so need to access index 0 to return the number and pass through jinja avg_rating = db.session.query(func.avg(Rating.score)).filter(Rating.movie_id == movie_id).one() # Query to get all ratings for a specific movie # Needed to join Rating and Movie tables and filter by user id # Sort movie titles alphabetically ratings = db.session.query(Rating.movie_id, Rating.score, Movie.title).join(Movie).filter(Rating.movie_id == movie_id).all() # # Pass user info into jinja and called on its attributes # # Pass count_score, avg_rating, and ratings into jinja # return render_template("movie_profile.html", movie=movie, count_score=count_score, avg_rating=avg_rating[0], ratings=ratings) return render_template( "movie_profile.html", movie=movie, user_rating=user_rating, avg_rating=avg_rating[0], count_score=count_score, prediction=prediction, ratings=ratings, beratement=beratement) @app.route("/
wizard_rating = wizard.predict_rating(movie)
conditional_block
server.py
in jinja not executed user_email = session.get("logged_in_user_email", None) if user_email is not None: user = User.query.filter(User.email == user_email).one() return render_template("homepage.html", user=user) else: return render_template("homepage.html", user=None) @app.route("/users") def user_list(): """Show list of users.""" users = User.query.all() return render_template("user_list.html", users=users) # This takes to each user's profile from user list @app.route("/users/<int:user_id>") def user_profile(user_id): """Show user information""" # Query by user id to return that record in database about user info user = User.query.filter(User.user_id == user_id).one() # import pdb; pdb.set_trace() # Query to get all movies and scores rated by this user # Needed to join Rating and Movie tables and filter by user id # Sort movie titles alphabetically user_movies = db.session.query(Rating.user_id, Rating.movie_id, Rating.score, Movie.title).join(Movie).filter(Rating.user_id == user_id).order_by(Movie.title).all() # Passed user info into jinja and called on its attributes # Passed user_movies into jinja and called on its attributes to get the info return render_template("user_profile.html", user=user, user_movies = user_movies) # # THIS WORKS, but we want to use /user/<int:user_id>, which we figured out above!! # @app.route("/user-profile") # def user_profile(): # """Show user information""" # # import pdb; pdb.set_trace() # # Get user email to query in User database and get all info about the user # email = session["logged_in_user_email"] # user = User.query.filter(User.email == email).one() # # # Test code to see attributes of user object # # user_id = user.user_id # # age = user.age # # zipcode = user.zipcode # return render_template("user_profile.html", user=user) @app.route("/signup-login", methods=["GET"]) def show_forms(): """Show signup and login forms.""" return render_template("signup_login.html") @app.route("/signup", methods=["POST"]) def signup(): """Check if user exists in database, otherwise add user to database.""" # Get values from signup form signup_email = request.form.get("signup_email") signup_password = request.form.get("signup_password") # If user exists, ask them to log in # Otherwise, add user into database and log them in, redirecting to homepage if db.session.query(User).filter(User.email == signup_email).first(): flash("You already have an account please use login!", "danger") return redirect("/signup-login") else: new_user = User(email=signup_email, password=signup_password, age=None, zipcode=None) db.session.add(new_user) db.session.commit() session["logged_in_user_email"] = signup_email session["logged_in_user"] = new_user.user_id flash("Your account has been created! You now are logged in!", "success") return redirect("/") @app.route("/login", methods=["POST"]) def login(): """Check if user's email matches password, otherwise ask user to try again.""" # Get values from login form login_email = request.form.get("login_email") login_password = request.form.get("login_password") # If user's email and password matches, log them in, redirecting them to homepage # Otherwise, ask them to log in with the correct password if db.session.query(User).filter(User.email == login_email, User.password == login_password).first(): flash("Login SUCCESS.", "success") # Query to get user's user id, in order to redirect user to their user profile user = User.query.filter(User.email == login_email).one() session["logged_in_user_email"] = login_email session["logged_in_user"] = user.user_id # Pass a variable through a string via string formatting # so we can pass user_id into the redirected route, which is a string!! return redirect("/users/%s" % user.user_id) # return redirect("/") else: flash("Incorrect password. Please try again!", "danger") return redirect("/signup-login") @app.route("/logout") def process_logout(): """Log user out.""" del session["logged_in_user_email"] del session["logged_in_user"] flash("Logged out.", "success") return redirect("/") @app.route("/movies") def movie_list(): """Show list of movies.""" # sort movie titles alphbetically movies = Movie.query.order_by(Movie.title).all() return render_template("movie_list.html", movies=movies) @app.route("/movies/<int:movie_id>", methods=['GET']) def movie_profile(movie_id): """Show movie information. If a user is logged in, let them add/edit a rating. """ if not session.get('logged_in_user_email'): flash("Please login or signup to see the movie details and rate the movie!", "danger") return redirect("/signup-login") else: # import pdb; pdb.set_trace(); # Query by movie id to return that record in database about movie info # movie = Movie.query.filter(Movie.movie_id == movie_id).one() movie = Movie.query.get(movie_id) user = User.query.filter(User.email == session.get("logged_in_user_email")).one() user_id = user.user_id if user_id: user_rating = Rating.query.filter_by(movie_id=movie_id, user_id=user_id).first() else: user_rating = None # Prediction code: only predict if the user hasn't rated it prediction = None if (not user_rating) and user_id: user = User.query.get(user_id) if user: prediction = user.predict_rating(movie) # Either use the prediction or their real rating if prediction: # User hasn't scored; use our prediction if we made one effective_rating = prediction elif user_rating: # User has already scored for real; use that effective_rating = user_rating.score else: # User hasn't scored and we couldn't get a prediction effective_rating = None # Get the wizard's rating, either by predicting or using real rating wizard = User.query.filter_by(email="wizard@gmail.com").one() wizard_rating = Rating.query.filter_by(user_id=wizard.user_id, movie_id=movie.movie_id).first() if wizard_rating is None: wizard_rating = wizard.predict_rating(movie) else: wizard_rating = wizard_rating.score if wizard_rating and effective_rating: difference = abs(wizard_rating - effective_rating) else: # We couldn't get a wizard rating, so we'll skip difference difference = None # Depending on how different we are from the Wizard, choose a message BERATEMENT_MESSAGES = [ "I suppose you don't have such bad taste after all.", "I regret every decision that I've ever made that has brought me to listen to your opinion.", "Words fail me, as your taste in movies has clearly failed you.", "That movie is great. For a clown to watch. Idiot.", "Words cannot express the awfulness of your taste." ] if difference is not None: beratement = BERATEMENT_MESSAGES[int(difference)] else: beratement = None # Tallies score of each rating (how many people rated this score per rating) # Returns list of tuples for count_score unordered_ratings = db.session.query(Rating.score, func.count(Rating.score)).filter(Rating.movie_id == movie_id).group_by(Rating.score) ordered_movies = unordered_ratings.order_by(Rating.score) count_score = ordered_movies.all() # Get average score, which returns a tuple-like object, so need to access index 0 to return the number and pass through jinja avg_rating = db.session.query(func.avg(Rating.score)).filter(Rating.movie_id == movie_id).one() # Query to get all ratings for a specific movie # Needed to join Rating and Movie tables and filter by user id # Sort movie titles alphabetically ratings = db.session.query(Rating.movie_id, Rating.score, Movie.title).join(Movie).filter(Rating.movie_id == movie_id).all() # # Pass user info into jinja and called on its attributes # # Pass count_score, avg_rating, and ratings into jinja # return render_template("movie_profile.html", movie=movie, count_score=count_score, avg_rating=avg_rating[0], ratings=ratings) return render_template( "movie_profile.html", movie=movie, user_rating=user_rating, avg_rating=avg_rating[0], count_score=count_score, prediction=prediction, ratings=ratings, beratement=beratement) @app.route("/movies/<int:movie_id>/rate-movie") def rate_movie(movie_id): """Get user rating score for movie""" user_rating = request.args.get("user_rating")
# get user id from log in email address user_email = session["logged_in_user_email"] user = User.query.filter(User.email == user_email).one()
random_line_split
broken_telephone.go
// randomSubstituteFor returns a semi-random but usually phonetically // close substitute for w. func (d metaphoneDict) randomSubstituteFor(w *metaphoneWord) *metaphoneWord { mp := w.metaphone if rand.Intn(2) == 0 { mp = w.secondary } match := d.matches(mp).randomNonEqual(w.literal) if match != nil { return match } return w } // Push adds mp to the end of d, reallocating space if necessary. func (d *metaphoneDict) Push(mp metaphoneWord) { dict := *d if len(dict) < cap(dict) { dict = dict[0 : len(dict)+1] } else { dict = make(metaphoneDict, len(dict)+1, cap(dict)+dictCapacity) copy(dict, *d) } dict[len(dict)-1] = mp *d = dict } // readWords creates a metaphone dictionary from input of one word per line. // The dictionary is sorted by phonetic representation. func readWords(input io.Reader) metaphoneDict { dict := newMetaphoneDict() rd := bufio.NewReader(input) for { if word, err := rd.ReadString('\n'); err == nil { word = strings.TrimSpace(word) if !*allowNamesInWordList { rune, _ := utf8.DecodeRuneInString(word) if unicode.IsUpper(rune) { continue } } if strings.Index(word, " ") != -1 || len(word) < minWordLength { continue } mp := *doubleMetaphone(word) dict.Push(mp) if mp.metaphone != mp.secondary { // Secondary phonetic representation dict.Push(metaphoneWord{word, mp.literal, mp.secondary, mp.metaphone}) } } else { if err != os.EOF { fmt.Printf("Error: %s\n", err) } break } } sort.Sort(dict) return dict } // doubleMetaphone is like doubleMetaphoneLimited but uses the default maxLen. func doubleMetaphone(word string) (result *metaphoneWord) { if doubleMetaphoneMaxLen > 0 { return doubleMetaphoneLimited(word, doubleMetaphoneMaxLen) } return doubleMetaphoneLimited(word, len(word)*2) } // doubleMetaphone returns two phonetic representations of an English word. // The secondary representation may equal primary. The implementation is // currently not fully complete with all special cases. The phonetic // representations are limited to maxLen length. func doubleMetaphoneLimited(word string, maxLen int) (result *metaphoneWord) { result = new(metaphoneWord) result.original = word word = strings.ToUpper(word) // TODO: Strip punctuation result.literal = word prev, skip, last, slavoGermanic := 0, 0, len(word)-1, false testSlavoGermanic: for pos, c := range word { switch c { case 'C': if pos == last || word[pos+1] != 'Z' { break } fallthrough case 'W', 'K': slavoGermanic = true break testSlavoGermanic } } word += " " // Allow indexing beyond the end for pos, c := range word { if c == ' ' { break } if skip > 0 { prev = 0 skip-- continue } mp, ms := "", "-" switch c { case 'A', 'E', 'I', 'O', 'U', 'Y', 'Ü', 'Ä', 'Ö', 'Å', 'É', 'È', 'Ï': if pos == 0 { // Initial vowel mp = "A" } else if pos == 1 && prev == 'W' { // W + vowel at the start of the word mp, ms = "A", "F" } case 'B': if prev == 'M' && pos > 1 && word[pos-2] == 'U' && (pos == last || (word[pos+1] == 'E' && word[pos+2] == 'R')) { // e.g. dumb, thumb break } if prev != 'B' { mp = "P" } case 'C': if prev == 'X' { break } if pos == 0 && strings.HasPrefix(word, "CAESAR") { mp = "S" skip = 1 break } next := word[pos+1] if next == 'H' { skip = 1 n2, n3 := word[pos+2], word[pos+3] if pos > 0 { if n2 == 'A' && n3 == 'E' { // michael mp, ms = "K", "X" break } if (pos == 1 && (prev == 'M' || prev == 'S')) || n2 == 'T' || n2 == 'S' { // Mc, Sch, -cht, -chs mp = "K" break } if (prev == 'A' || prev == 'O' || prev == 'U' || prev == 'E') && (n2 == 'L' || n2 == 'R' || n2 == 'N' || n2 == 'M' || n2 == 'B' || n2 == 'B' || n2 == 'H' || n2 == 'F' || n2 == 'V' || n2 == 'W') { // e.g. wachtler, wechsler, but not tichner mp = "K" break } if pos > 1 { p2 := word[pos-2] if prev == 'R' && ((p2 == 'O' && n2 == 'E' && n3 == 'S') || (p2 == 'O' && n2 == 'I' && n3 == 'D') || (p2 == 'A' && n2 == 'I' && n3 == 'T')) { // orchestra, orchid, architect (but not arch) mp = "K" break } } } else { // pos == 0 n4, n5 := word[pos+4], word[pos+5] if (n2 == 'A' && n3 == 'R' && ((n4 == 'A' && n5 == 'C') || (n4 == 'I' && n5 == 'S'))) || (n2 == 'E' && n3 == 'M') || (n2 == 'Y' && n3 == 'M') || (n2 == 'I' && n3 == 'A') || (n2 == 'O' && n3 == 'R' && (n4 != 'O' || n5 != 'E')) { // e.g. character, charisma, chorus, chemistry // but not "chore" mp = "K" } else { switch n2 { case 'L', 'R', 'N', 'M', 'B', 'H', 'F', 'V', 'W', ' ': mp = "K" default: mp = "X" } } break } mp, ms = "X", "K" break } else if next == 'Z' { if pos < 2 || word[pos-1] != 'I' || word[pos-2] == 'W' { // cz, not wicz mp, ms = "S", "X" skip = 1 break } } else if next == 'C' { n2 := word[pos+2] if n2 == 'I' && word[pos+3] == 'A' { // -ccia, e.g. focaccia mp = "X" skip = 2 break } if pos != 2 || prev != 'M' { // -cc, but not e.g. McClellan if n
{ if len(d) == 0 { return nil } i := rand.Intn(len(d)) switch { case d[i].literal != w: case i > 0 && d[i-1].literal != w: i-- case i < len(d)-1 && d[i+1].literal != w: i++ case i == 0 && len(d) > 2 && d[i+2].literal != w: i += 2 case i == len(d)-1 && len(d) > 2 && d[i-2].literal != w: i -= 2 default: return nil } return &d[i] }
identifier_body
broken_telephone.go
(i, j int) bool { return d[i].metaphone < d[j].metaphone } // phoneticLocation returns the index where metaphone is or would be sorted. func (d metaphoneDict) phoneticLocation(metaphone string) (i int) { left, right := 0, len(d) for left < right { i = left + ((right - left) / 2) if d[i].metaphone < metaphone { left = i + 1 } else { right = i } } return } // matches returns a slice containing all exact matches in d for metaphone. func (d metaphoneDict) matches(metaphone string) metaphoneDict { var l, r int i := d.phoneticLocation(metaphone) for r = i; r < len(d) && d[r].metaphone == metaphone; r++ { } for l = i; l >= 0 && d[l].metaphone == metaphone; l-- { } l++ if r-l < fudgeDistance*2 { l -= fudgeDistance r += fudgeDistance } if l < 0 { l = 0 } if r > len(d) { r = len(d) } return d[l:r] } // randomNonEqual returns a randomly selected word in d which is not // literally equal to w. Returns -1 if no such word exists. // (The assumption is that the same word literal appears at most twice.) func (d metaphoneDict) randomNonEqual(w string) *metaphoneWord { if len(d) == 0 { return nil } i := rand.Intn(len(d)) switch { case d[i].literal != w: case i > 0 && d[i-1].literal != w: i-- case i < len(d)-1 && d[i+1].literal != w: i++ case i == 0 && len(d) > 2 && d[i+2].literal != w: i += 2 case i == len(d)-1 && len(d) > 2 && d[i-2].literal != w: i -= 2 default: return nil } return &d[i] } // randomSubstituteFor returns a semi-random but usually phonetically // close substitute for w. func (d metaphoneDict) randomSubstituteFor(w *metaphoneWord) *metaphoneWord { mp := w.metaphone if rand.Intn(2) == 0 { mp = w.secondary } match := d.matches(mp).randomNonEqual(w.literal) if match != nil { return match } return w } // Push adds mp to the end of d, reallocating space if necessary. func (d *metaphoneDict) Push(mp metaphoneWord) { dict := *d if len(dict) < cap(dict) { dict = dict[0 : len(dict)+1] } else { dict = make(metaphoneDict, len(dict)+1, cap(dict)+dictCapacity) copy(dict, *d) } dict[len(dict)-1] = mp *d = dict } // readWords creates a metaphone dictionary from input of one word per line. // The dictionary is sorted by phonetic representation. func readWords(input io.Reader) metaphoneDict { dict := newMetaphoneDict() rd := bufio.NewReader(input) for { if word, err := rd.ReadString('\n'); err == nil { word = strings.TrimSpace(word) if !*allowNamesInWordList { rune, _ := utf8.DecodeRuneInString(word) if unicode.IsUpper(rune) { continue } } if strings.Index(word, " ") != -1 || len(word) < minWordLength { continue } mp := *doubleMetaphone(word) dict.Push(mp) if mp.metaphone != mp.secondary { // Secondary phonetic representation dict.Push(metaphoneWord{word, mp.literal, mp.secondary, mp.metaphone}) } } else { if err != os.EOF { fmt.Printf("Error: %s\n", err) } break } } sort.Sort(dict) return dict } // doubleMetaphone is like doubleMetaphoneLimited but uses the default maxLen. func doubleMetaphone(word string) (result *metaphoneWord) { if doubleMetaphoneMaxLen > 0 { return doubleMetaphoneLimited(word, doubleMetaphoneMaxLen) } return doubleMetaphoneLimited(word, len(word)*2) } // doubleMetaphone returns two phonetic representations of an English word. // The secondary representation may equal primary. The implementation is // currently not fully complete with all special cases. The phonetic // representations are limited to maxLen length. func doubleMetaphoneLimited(word string, maxLen int) (result *metaphoneWord) { result = new(metaphoneWord) result.original = word word = strings.ToUpper(word) // TODO: Strip punctuation result.literal = word prev, skip, last, slavoGermanic := 0, 0, len(word)-1, false testSlavoGermanic: for pos, c := range word { switch c { case 'C': if pos == last || word[pos+1] != 'Z' { break } fallthrough case 'W', 'K': slavoGermanic = true break testSlavoGermanic } } word += " " // Allow indexing beyond the end for pos, c := range word { if c == ' ' { break } if skip > 0 { prev = 0 skip-- continue } mp, ms := "", "-" switch c { case 'A', 'E', 'I', 'O', 'U', 'Y', 'Ü', 'Ä', 'Ö', 'Å', 'É', 'È', 'Ï': if pos == 0 { // Initial vowel mp = "A" } else if pos == 1 && prev == 'W' { // W + vowel at the start of the word mp, ms = "A", "F" } case 'B': if prev == 'M' && pos > 1 && word[pos-2] == 'U' && (pos == last || (word[pos+1] == 'E' && word[pos+2] == 'R')) { // e.g. dumb, thumb break } if prev != 'B' { mp = "P" } case 'C': if prev == 'X' { break } if pos == 0 && strings.HasPrefix(word, "CAESAR") { mp = "S" skip = 1 break } next := word[pos+1] if next == 'H' { skip = 1 n2, n3 := word[pos+2], word[pos+3] if pos > 0 { if n2 == 'A' && n3 == 'E' { // michael mp, ms = "K", "X" break } if (pos == 1 && (prev == 'M' || prev == 'S')) || n2 == 'T' || n2 == 'S' { // Mc, Sch, -cht, -chs mp = "K" break } if (prev == 'A' || prev == 'O' || prev == 'U' || prev == 'E') && (n2 == 'L' || n2 == 'R' || n2 == 'N' || n2 == 'M' || n2 == 'B' || n2 == 'B' || n2 == 'H' || n2 == 'F' || n2 == 'V' || n2 == 'W') { // e.g. wachtler, wechsler, but not tichner mp = "K" break } if pos > 1 { p2 := word[pos-2] if prev == 'R' && ((p2 == 'O' && n2 == 'E' && n3 == 'S') || (p2 == 'O' && n2 == 'I' && n3 == 'D') || (p2 == 'A' && n2 == 'I' && n3 == 'T')) { // orchestra, orchid, architect (but not arch) mp = "K" break } } } else { // pos == 0 n4, n5 := word[pos+4], word[pos+5] if (n2 == 'A' && n3 == 'R' && ((n4 == 'A' && n5 == 'C') || (n4 == 'I' && n5 == 'S'))) || (n2 == 'E' && n3
Less
identifier_name
broken_telephone.go
(dict)+1, cap(dict)+dictCapacity) copy(dict, *d) } dict[len(dict)-1] = mp *d = dict } // readWords creates a metaphone dictionary from input of one word per line. // The dictionary is sorted by phonetic representation. func readWords(input io.Reader) metaphoneDict { dict := newMetaphoneDict() rd := bufio.NewReader(input) for { if word, err := rd.ReadString('\n'); err == nil { word = strings.TrimSpace(word) if !*allowNamesInWordList { rune, _ := utf8.DecodeRuneInString(word) if unicode.IsUpper(rune) { continue } } if strings.Index(word, " ") != -1 || len(word) < minWordLength { continue } mp := *doubleMetaphone(word) dict.Push(mp) if mp.metaphone != mp.secondary { // Secondary phonetic representation dict.Push(metaphoneWord{word, mp.literal, mp.secondary, mp.metaphone}) } } else { if err != os.EOF { fmt.Printf("Error: %s\n", err) } break } } sort.Sort(dict) return dict } // doubleMetaphone is like doubleMetaphoneLimited but uses the default maxLen. func doubleMetaphone(word string) (result *metaphoneWord) { if doubleMetaphoneMaxLen > 0 { return doubleMetaphoneLimited(word, doubleMetaphoneMaxLen) } return doubleMetaphoneLimited(word, len(word)*2) } // doubleMetaphone returns two phonetic representations of an English word. // The secondary representation may equal primary. The implementation is // currently not fully complete with all special cases. The phonetic // representations are limited to maxLen length. func doubleMetaphoneLimited(word string, maxLen int) (result *metaphoneWord) { result = new(metaphoneWord) result.original = word word = strings.ToUpper(word) // TODO: Strip punctuation result.literal = word prev, skip, last, slavoGermanic := 0, 0, len(word)-1, false testSlavoGermanic: for pos, c := range word { switch c { case 'C': if pos == last || word[pos+1] != 'Z' { break } fallthrough case 'W', 'K': slavoGermanic = true break testSlavoGermanic } } word += " " // Allow indexing beyond the end for pos, c := range word { if c == ' ' { break } if skip > 0 { prev = 0 skip-- continue } mp, ms := "", "-" switch c { case 'A', 'E', 'I', 'O', 'U', 'Y', 'Ü', 'Ä', 'Ö', 'Å', 'É', 'È', 'Ï': if pos == 0 { // Initial vowel mp = "A" } else if pos == 1 && prev == 'W' { // W + vowel at the start of the word mp, ms = "A", "F" } case 'B': if prev == 'M' && pos > 1 && word[pos-2] == 'U' && (pos == last || (word[pos+1] == 'E' && word[pos+2] == 'R')) { // e.g. dumb, thumb break } if prev != 'B' { mp = "P" } case 'C': if prev == 'X' { break } if pos == 0 && strings.HasPrefix(word, "CAESAR") { mp = "S" skip = 1 break } next := word[pos+1] if next == 'H' { skip = 1 n2, n3 := word[pos+2], word[pos+3]
// michael mp, ms = "K", "X" break } if (pos == 1 && (prev == 'M' || prev == 'S')) || n2 == 'T' || n2 == 'S' { // Mc, Sch, -cht, -chs mp = "K" break } if (prev == 'A' || prev == 'O' || prev == 'U' || prev == 'E') && (n2 == 'L' || n2 == 'R' || n2 == 'N' || n2 == 'M' || n2 == 'B' || n2 == 'B' || n2 == 'H' || n2 == 'F' || n2 == 'V' || n2 == 'W') { // e.g. wachtler, wechsler, but not tichner mp = "K" break } if pos > 1 { p2 := word[pos-2] if prev == 'R' && ((p2 == 'O' && n2 == 'E' && n3 == 'S') || (p2 == 'O' && n2 == 'I' && n3 == 'D') || (p2 == 'A' && n2 == 'I' && n3 == 'T')) { // orchestra, orchid, architect (but not arch) mp = "K" break } } } else { // pos == 0 n4, n5 := word[pos+4], word[pos+5] if (n2 == 'A' && n3 == 'R' && ((n4 == 'A' && n5 == 'C') || (n4 == 'I' && n5 == 'S'))) || (n2 == 'E' && n3 == 'M') || (n2 == 'Y' && n3 == 'M') || (n2 == 'I' && n3 == 'A') || (n2 == 'O' && n3 == 'R' && (n4 != 'O' || n5 != 'E')) { // e.g. character, charisma, chorus, chemistry // but not "chore" mp = "K" } else { switch n2 { case 'L', 'R', 'N', 'M', 'B', 'H', 'F', 'V', 'W', ' ': mp = "K" default: mp = "X" } } break } mp, ms = "X", "K" break } else if next == 'Z' { if pos < 2 || word[pos-1] != 'I' || word[pos-2] == 'W' { // cz, not wicz mp, ms = "S", "X" skip = 1 break } } else if next == 'C' { n2 := word[pos+2] if n2 == 'I' && word[pos+3] == 'A' { // -ccia, e.g. focaccia mp = "X" skip = 2 break } if pos != 2 || prev != 'M' { // -cc, but not e.g. McClellan if n2 == 'I' || n2 == 'E' || (n2 == 'H' && word[pos+3] != 'U') { // e.g. bellocchio, but not bacchus skip = 3 if pos == 1 && prev == 'A' { // e.g. accident mp = "KS" break } else if prev == 'U' && n2 == 'E' && (word[pos+4] == 'S' || word[pos+4] == 'E') { // succeed, success mp = "KS" break } mp = "X" break } } if n2 != 'I' && n2 != 'E' { skip = 1 } } else if next == 'K' || next == 'Q' { skip = 1 } else if next == 'I' { mp = "S" skip = 1 n2 := word[pos+2] if n2 == 'O' || n2 == 'E' || n2 == 'A' { // cio, cie, cia ms = "X" } break } else if next == 'E' || next == 'Y' { skip = 1 mp = "S
if pos > 0 { if n2 == 'A' && n3 == 'E' {
random_line_split
broken_telephone.go
} } if strings.Index(word, " ") != -1 || len(word) < minWordLength { continue } mp := *doubleMetaphone(word) dict.Push(mp) if mp.metaphone != mp.secondary { // Secondary phonetic representation dict.Push(metaphoneWord{word, mp.literal, mp.secondary, mp.metaphone}) } } else { if err != os.EOF { fmt.Printf("Error: %s\n", err) } break } } sort.Sort(dict) return dict } // doubleMetaphone is like doubleMetaphoneLimited but uses the default maxLen. func doubleMetaphone(word string) (result *metaphoneWord) { if doubleMetaphoneMaxLen > 0 { return doubleMetaphoneLimited(word, doubleMetaphoneMaxLen) } return doubleMetaphoneLimited(word, len(word)*2) } // doubleMetaphone returns two phonetic representations of an English word. // The secondary representation may equal primary. The implementation is // currently not fully complete with all special cases. The phonetic // representations are limited to maxLen length. func doubleMetaphoneLimited(word string, maxLen int) (result *metaphoneWord) { result = new(metaphoneWord) result.original = word word = strings.ToUpper(word) // TODO: Strip punctuation result.literal = word prev, skip, last, slavoGermanic := 0, 0, len(word)-1, false testSlavoGermanic: for pos, c := range word { switch c { case 'C': if pos == last || word[pos+1] != 'Z' { break } fallthrough case 'W', 'K': slavoGermanic = true break testSlavoGermanic } } word += " " // Allow indexing beyond the end for pos, c := range word { if c == ' ' { break } if skip > 0 { prev = 0 skip-- continue } mp, ms := "", "-" switch c { case 'A', 'E', 'I', 'O', 'U', 'Y', 'Ü', 'Ä', 'Ö', 'Å', 'É', 'È', 'Ï': if pos == 0 { // Initial vowel mp = "A" } else if pos == 1 && prev == 'W' { // W + vowel at the start of the word mp, ms = "A", "F" } case 'B': if prev == 'M' && pos > 1 && word[pos-2] == 'U' && (pos == last || (word[pos+1] == 'E' && word[pos+2] == 'R')) { // e.g. dumb, thumb break } if prev != 'B' { mp = "P" } case 'C': if prev == 'X' { break } if pos == 0 && strings.HasPrefix(word, "CAESAR") { mp = "S" skip = 1 break } next := word[pos+1] if next == 'H' { skip = 1 n2, n3 := word[pos+2], word[pos+3] if pos > 0 { if n2 == 'A' && n3 == 'E' { // michael mp, ms = "K", "X" break } if (pos == 1 && (prev == 'M' || prev == 'S')) || n2 == 'T' || n2 == 'S' { // Mc, Sch, -cht, -chs mp = "K" break } if (prev == 'A' || prev == 'O' || prev == 'U' || prev == 'E') && (n2 == 'L' || n2 == 'R' || n2 == 'N' || n2 == 'M' || n2 == 'B' || n2 == 'B' || n2 == 'H' || n2 == 'F' || n2 == 'V' || n2 == 'W') { // e.g. wachtler, wechsler, but not tichner mp = "K" break } if pos > 1 { p2 := word[pos-2] if prev == 'R' && ((p2 == 'O' && n2 == 'E' && n3 == 'S') || (p2 == 'O' && n2 == 'I' && n3 == 'D') || (p2 == 'A' && n2 == 'I' && n3 == 'T')) { // orchestra, orchid, architect (but not arch) mp = "K" break } } } else { // pos == 0 n4, n5 := word[pos+4], word[pos+5] if (n2 == 'A' && n3 == 'R' && ((n4 == 'A' && n5 == 'C') || (n4 == 'I' && n5 == 'S'))) || (n2 == 'E' && n3 == 'M') || (n2 == 'Y' && n3 == 'M') || (n2 == 'I' && n3 == 'A') || (n2 == 'O' && n3 == 'R' && (n4 != 'O' || n5 != 'E')) { // e.g. character, charisma, chorus, chemistry // but not "chore" mp = "K" } else { switch n2 { case 'L', 'R', 'N', 'M', 'B', 'H', 'F', 'V', 'W', ' ': mp = "K" default: mp = "X" } } break } mp, ms = "X", "K" break } else if next == 'Z' { if pos < 2 || word[pos-1] != 'I' || word[pos-2] == 'W' { // cz, not wicz mp, ms = "S", "X" skip = 1 break } } else if next == 'C' { n2 := word[pos+2] if n2 == 'I' && word[pos+3] == 'A' { // -ccia, e.g. focaccia mp = "X" skip = 2 break } if pos != 2 || prev != 'M' { // -cc, but not e.g. McClellan if n2 == 'I' || n2 == 'E' || (n2 == 'H' && word[pos+3] != 'U') { // e.g. bellocchio, but not bacchus skip = 3 if pos == 1 && prev == 'A' { // e.g. accident mp = "KS" break } else if prev == 'U' && n2 == 'E' && (word[pos+4] == 'S' || word[pos+4] == 'E') { // succeed, success mp = "KS" break } mp = "X" break } } if n2 != 'I' && n2 != 'E' { skip = 1 } } else if next == 'K' || next == 'Q' { skip = 1 } else if next == 'I' { mp = "S" skip = 1 n2 := word[pos+2] if n2 == 'O' || n2 == 'E' || n2 == 'A' { // cio, cie, cia ms = "X" } break } else if next == 'E' || next == 'Y' { skip = 1 mp = "S" break } mp = "K" case 'D': if prev != 'D' && prev != 'T' { if word[pos+1] == 'G' { skip = 1 switch word[pos+2] { case 'E', 'I', 'Y': // e.g. "edge" mp = "J" default: // e.g. "edgar" mp = "K" } break } mp = "T" } case 'F', 'V': if prev != c { m
p = "F" } case
conditional_block
mod.rs
.read(); match peer_guard.control_channel.connection() { Some(connection) => connection.clone(), None => return, } }; let command_stream = connection.take_command_stream(); // Limit to 16 since that is the max number of transactions we can process at any one time per // AVCTP match command_stream .map(Ok) .try_for_each_concurrent(16, |command| async { let fut = peer.read().command_handler.handle_command(command.unwrap()); let result: Result<(), PeerError> = fut.await; result }) .await { Ok(_) => fx_log_info!("Peer command stream closed"), Err(e) => fx_log_err!("Peer command returned error {:?}", e), } // Command stream closed/errored. Disconnect the peer. { peer.write().reset_connection(false); } } /// Handles received notifications from the peer from the subscribed notifications streams and /// dispatches the notifications back to the controller listeners fn handle_notification( notif: &NotificationEventId, peer: &Arc<RwLock<RemotePeer>>, data: &[u8], ) -> Result<bool, Error> { fx_vlog!(tag: "avrcp", 2, "received notification for {:?} {:?}", notif, data); let preamble = VendorDependentPreamble::decode(data).map_err(|e| Error::PacketError(e))?; let data = &data[preamble.encoded_len()..]; if data.len() < preamble.parameter_length as usize { return Err(Error::UnexpectedResponse); } match notif { NotificationEventId::EventPlaybackStatusChanged => { let response = PlaybackStatusChangedNotificationResponse::decode(data) .map_err(|e| Error::PacketError(e))?; peer.write().handle_new_controller_notification_event( ControllerEvent::PlaybackStatusChanged(response.playback_status()), ); Ok(false) } NotificationEventId::EventTrackChanged => { let response = TrackChangedNotificationResponse::decode(data) .map_err(|e| Error::PacketError(e))?; peer.write().handle_new_controller_notification_event(ControllerEvent::TrackIdChanged( response.identifier(), )); Ok(false) } NotificationEventId::EventPlaybackPosChanged => { let response = PlaybackPosChangedNotificationResponse::decode(data) .map_err(|e| Error::PacketError(e))?; peer.write().handle_new_controller_notification_event( ControllerEvent::PlaybackPosChanged(response.position()), ); Ok(false) } NotificationEventId::EventVolumeChanged => { let response = VolumeChangedNotificationResponse::decode(data) .map_err(|e| Error::PacketError(e))?; peer.write().handle_new_controller_notification_event(ControllerEvent::VolumeChanged( response.volume(), )); Ok(false) } _ => Ok(true), } } /// Starts a task to attempt an outgoing L2CAP connection to remote's AVRCP control channel. /// The control channel should be in `Connecting` state before spawning this task. /// TODO(BT-2747): Fix a race where an incoming connection can come in while we are making an /// outgoing connection. Properly handle the case where we are attempting to connect to remote /// at the same time they make an incoming connection according to how the the spec says. fn start_make_connection_task(peer: Arc<RwLock<RemotePeer>>) { let peer = peer.clone(); fasync::spawn(async move { let (peer_id, profile_service) = { let peer_guard = peer.read(); // early return if we are not in `Connecting` match peer_guard.control_channel { PeerChannel::Connecting => {} _ => return, } (peer_guard.peer_id.clone(), peer_guard.profile_svc.clone()) }; match profile_service.connect_to_device(&peer_id, PSM_AVCTP as u16).await { Ok(socket) => { let mut peer_guard = peer.write(); match peer_guard.control_channel { PeerChannel::Connecting => match AvcPeer::new(socket) { Ok(peer) => { peer_guard.set_control_connection(peer); } Err(e) => { peer_guard.reset_connection(false); fx_log_err!("Unable to make peer from socket {}: {:?}", peer_id, e); } }, _ => { fx_log_info!( "incoming connection established while making outgoing {:?}", peer_id ); // an incoming l2cap connection was made while we were making an // outgoing one. Drop both connections per spec. peer_guard.reset_connection(false); } }; } Err(e) => { fx_log_err!("connect_to_device error {}: {:?}", peer_id, e); let mut peer_guard = peer.write(); if let PeerChannel::Connecting = peer_guard.control_channel { peer_guard.reset_connection(false); } } } }) } /// Checks for supported notification on the peer and registers for notifications. /// This is started on a remote peer when we have a connection and target profile descriptor. async fn pump_notifications(peer: Arc<RwLock<RemotePeer>>) { // events we support when speaking to a peer that supports the target profile. const SUPPORTED_NOTIFICATIONS: [NotificationEventId; 4] = [ NotificationEventId::EventPlaybackStatusChanged, NotificationEventId::EventTrackChanged, NotificationEventId::EventPlaybackPosChanged, NotificationEventId::EventVolumeChanged, ]; let supported_notifications: Vec<NotificationEventId> = SUPPORTED_NOTIFICATIONS.iter().cloned().collect(); // look up what notifications we support on this peer first. Consider updating this from // time to time. let remote_supported_notifications = match get_supported_events_internal(peer.clone()).await { Ok(x) => x, Err(_) => return, }; let supported_notifications: Vec<NotificationEventId> = remote_supported_notifications .into_iter() .filter(|k| supported_notifications.contains(k)) .collect(); let mut notification_streams = SelectAll::new(); for notif in supported_notifications { fx_vlog!(tag: "avrcp", 2, "creating notification stream for {:#?}", notif); let stream = NotificationStream::new(peer.clone(), notif, 1).map_ok(move |data| (notif, data)); notification_streams.push(stream); } pin_mut!(notification_streams); loop { if futures::select! { event_result = notification_streams.select_next_some() => { match event_result { Ok((notif, data)) => { handle_notification(&notif, &peer, &data[..]) .unwrap_or_else(|e| { fx_log_err!("Error decoding packet from peer {:?}", e); true} ) }, Err(Error::CommandNotSupported) => false, Err(_) => true, _=> true, } } complete => { true } } { break; } } fx_vlog!(tag: "avrcp", 2, "stopping notifications for {:#?}", peer.read().peer_id); } /// Starts a task to poll notifications on the remote peer. Aborted when the peer connection is /// reset. fn start_notifications_processing_task(peer: Arc<RwLock<RemotePeer>>) -> AbortHandle { let (handle, registration) = AbortHandle::new_pair(); fasync::spawn( Abortable::new( async move { pump_notifications(peer).await; }, registration, ) .map(|_| ()), ); handle } /// Starts a task to poll control messages from the peer. Aborted when the peer connection is /// reset. Started when we have a connection to the remote peer and we have any type of valid SDP /// profile from the peer. fn start_control_stream_processing_task(peer: Arc<RwLock<RemotePeer>>) -> AbortHandle { let (handle, registration) = AbortHandle::new_pair(); fasync::spawn( Abortable::new( async move { process_control_stream(peer).await; }, registration,
/// State observer task around a remote peer. Takes a change stream from the remote peer that wakes /// the task whenever some state has changed on the peer. Swaps tasks such as making outgoing /// connections, processing the incoming control messages, and registering for notifications on the /// remote peer. pub(super) async fn state_watcher(peer: Arc<RwLock<RemotePeer>>) { fx_vlog!(tag: "avrcp", 2, "state_watcher starting"); let mut change_stream = peer.read().state_change_listener.take_change_stream(); let peer_weak = Arc::downgrade(&peer); drop(peer); let mut channel_processor_abort_handle: Option<AbortHandle> = None; let mut notification_poll_abort_handle: Option<AbortHandle> = None; while let Some(_) = change_stream.next().await { fx_vlog!(tag: "avrcp", 2, "state_watcher command received"); if let Some(peer) = peer_weak.upgrade() { let mut peer_guard = peer.write(); fx_vlog!(tag: "avrcp", 2, "make_connection control channel {:?}", peer_guard.control_channel); match peer_guard.control_channel { PeerChannel::Connecting => {} PeerChannel::Disconnected => { if let Some(ref abort_handle) = channel_processor_abort_handle { abort_handle.abort(); channel_processor_abort_handle = None; } if let Some(ref abort
) .map(|_| ()), ); handle }
random_line_split
mod.rs
.read(); match peer_guard.control_channel.connection() { Some(connection) => connection.clone(), None => return, } }; let command_stream = connection.take_command_stream(); // Limit to 16 since that is the max number of transactions we can process at any one time per // AVCTP match command_stream .map(Ok) .try_for_each_concurrent(16, |command| async { let fut = peer.read().command_handler.handle_command(command.unwrap()); let result: Result<(), PeerError> = fut.await; result }) .await { Ok(_) => fx_log_info!("Peer command stream closed"), Err(e) => fx_log_err!("Peer command returned error {:?}", e), } // Command stream closed/errored. Disconnect the peer. { peer.write().reset_connection(false); } } /// Handles received notifications from the peer from the subscribed notifications streams and /// dispatches the notifications back to the controller listeners fn handle_notification( notif: &NotificationEventId, peer: &Arc<RwLock<RemotePeer>>, data: &[u8], ) -> Result<bool, Error>
NotificationEventId::EventTrackChanged => { let response = TrackChangedNotificationResponse::decode(data) .map_err(|e| Error::PacketError(e))?; peer.write().handle_new_controller_notification_event(ControllerEvent::TrackIdChanged( response.identifier(), )); Ok(false) } NotificationEventId::EventPlaybackPosChanged => { let response = PlaybackPosChangedNotificationResponse::decode(data) .map_err(|e| Error::PacketError(e))?; peer.write().handle_new_controller_notification_event( ControllerEvent::PlaybackPosChanged(response.position()), ); Ok(false) } NotificationEventId::EventVolumeChanged => { let response = VolumeChangedNotificationResponse::decode(data) .map_err(|e| Error::PacketError(e))?; peer.write().handle_new_controller_notification_event(ControllerEvent::VolumeChanged( response.volume(), )); Ok(false) } _ => Ok(true), } } /// Starts a task to attempt an outgoing L2CAP connection to remote's AVRCP control channel. /// The control channel should be in `Connecting` state before spawning this task. /// TODO(BT-2747): Fix a race where an incoming connection can come in while we are making an /// outgoing connection. Properly handle the case where we are attempting to connect to remote /// at the same time they make an incoming connection according to how the the spec says. fn start_make_connection_task(peer: Arc<RwLock<RemotePeer>>) { let peer = peer.clone(); fasync::spawn(async move { let (peer_id, profile_service) = { let peer_guard = peer.read(); // early return if we are not in `Connecting` match peer_guard.control_channel { PeerChannel::Connecting => {} _ => return, } (peer_guard.peer_id.clone(), peer_guard.profile_svc.clone()) }; match profile_service.connect_to_device(&peer_id, PSM_AVCTP as u16).await { Ok(socket) => { let mut peer_guard = peer.write(); match peer_guard.control_channel { PeerChannel::Connecting => match AvcPeer::new(socket) { Ok(peer) => { peer_guard.set_control_connection(peer); } Err(e) => { peer_guard.reset_connection(false); fx_log_err!("Unable to make peer from socket {}: {:?}", peer_id, e); } }, _ => { fx_log_info!( "incoming connection established while making outgoing {:?}", peer_id ); // an incoming l2cap connection was made while we were making an // outgoing one. Drop both connections per spec. peer_guard.reset_connection(false); } }; } Err(e) => { fx_log_err!("connect_to_device error {}: {:?}", peer_id, e); let mut peer_guard = peer.write(); if let PeerChannel::Connecting = peer_guard.control_channel { peer_guard.reset_connection(false); } } } }) } /// Checks for supported notification on the peer and registers for notifications. /// This is started on a remote peer when we have a connection and target profile descriptor. async fn pump_notifications(peer: Arc<RwLock<RemotePeer>>) { // events we support when speaking to a peer that supports the target profile. const SUPPORTED_NOTIFICATIONS: [NotificationEventId; 4] = [ NotificationEventId::EventPlaybackStatusChanged, NotificationEventId::EventTrackChanged, NotificationEventId::EventPlaybackPosChanged, NotificationEventId::EventVolumeChanged, ]; let supported_notifications: Vec<NotificationEventId> = SUPPORTED_NOTIFICATIONS.iter().cloned().collect(); // look up what notifications we support on this peer first. Consider updating this from // time to time. let remote_supported_notifications = match get_supported_events_internal(peer.clone()).await { Ok(x) => x, Err(_) => return, }; let supported_notifications: Vec<NotificationEventId> = remote_supported_notifications .into_iter() .filter(|k| supported_notifications.contains(k)) .collect(); let mut notification_streams = SelectAll::new(); for notif in supported_notifications { fx_vlog!(tag: "avrcp", 2, "creating notification stream for {:#?}", notif); let stream = NotificationStream::new(peer.clone(), notif, 1).map_ok(move |data| (notif, data)); notification_streams.push(stream); } pin_mut!(notification_streams); loop { if futures::select! { event_result = notification_streams.select_next_some() => { match event_result { Ok((notif, data)) => { handle_notification(&notif, &peer, &data[..]) .unwrap_or_else(|e| { fx_log_err!("Error decoding packet from peer {:?}", e); true} ) }, Err(Error::CommandNotSupported) => false, Err(_) => true, _=> true, } } complete => { true } } { break; } } fx_vlog!(tag: "avrcp", 2, "stopping notifications for {:#?}", peer.read().peer_id); } /// Starts a task to poll notifications on the remote peer. Aborted when the peer connection is /// reset. fn start_notifications_processing_task(peer: Arc<RwLock<RemotePeer>>) -> AbortHandle { let (handle, registration) = AbortHandle::new_pair(); fasync::spawn( Abortable::new( async move { pump_notifications(peer).await; }, registration, ) .map(|_| ()), ); handle } /// Starts a task to poll control messages from the peer. Aborted when the peer connection is /// reset. Started when we have a connection to the remote peer and we have any type of valid SDP /// profile from the peer. fn start_control_stream_processing_task(peer: Arc<RwLock<RemotePeer>>) -> AbortHandle { let (handle, registration) = AbortHandle::new_pair(); fasync::spawn( Abortable::new( async move { process_control_stream(peer).await; }, registration, ) .map(|_| ()), ); handle } /// State observer task around a remote peer. Takes a change stream from the remote peer that wakes /// the task whenever some state has changed on the peer. Swaps tasks such as making outgoing /// connections, processing the incoming control messages, and registering for notifications on the /// remote peer. pub(super) async fn state_watcher(peer: Arc<RwLock<RemotePeer>>) { fx_vlog!(tag: "avrcp", 2, "state_watcher starting"); let mut change_stream = peer.read().state_change_listener.take_change_stream(); let peer_weak = Arc::downgrade(&peer); drop(peer); let mut channel_processor_abort_handle: Option<AbortHandle> = None; let mut notification_poll_abort_handle: Option<AbortHandle> = None; while let Some(_) = change_stream.next().await { fx_vlog!(tag: "avrcp", 2, "state_watcher command received"); if let Some(peer) = peer_weak.upgrade() { let mut peer_guard = peer.write(); fx_vlog!(tag: "avrcp", 2, "make_connection control channel {:?}", peer_guard.control_channel); match peer_guard.control_channel { PeerChannel::Connecting => {} PeerChannel::Disconnected => { if let Some(ref abort_handle) = channel_processor_abort_handle { abort_handle.abort(); channel_processor_abort_handle = None; } if let Some(ref abort
{ fx_vlog!(tag: "avrcp", 2, "received notification for {:?} {:?}", notif, data); let preamble = VendorDependentPreamble::decode(data).map_err(|e| Error::PacketError(e))?; let data = &data[preamble.encoded_len()..]; if data.len() < preamble.parameter_length as usize { return Err(Error::UnexpectedResponse); } match notif { NotificationEventId::EventPlaybackStatusChanged => { let response = PlaybackStatusChangedNotificationResponse::decode(data) .map_err(|e| Error::PacketError(e))?; peer.write().handle_new_controller_notification_event( ControllerEvent::PlaybackStatusChanged(response.playback_status()), ); Ok(false) }
identifier_body
mod.rs
.read(); match peer_guard.control_channel.connection() { Some(connection) => connection.clone(), None => return, } }; let command_stream = connection.take_command_stream(); // Limit to 16 since that is the max number of transactions we can process at any one time per // AVCTP match command_stream .map(Ok) .try_for_each_concurrent(16, |command| async { let fut = peer.read().command_handler.handle_command(command.unwrap()); let result: Result<(), PeerError> = fut.await; result }) .await { Ok(_) => fx_log_info!("Peer command stream closed"), Err(e) => fx_log_err!("Peer command returned error {:?}", e), } // Command stream closed/errored. Disconnect the peer. { peer.write().reset_connection(false); } } /// Handles received notifications from the peer from the subscribed notifications streams and /// dispatches the notifications back to the controller listeners fn handle_notification( notif: &NotificationEventId, peer: &Arc<RwLock<RemotePeer>>, data: &[u8], ) -> Result<bool, Error> { fx_vlog!(tag: "avrcp", 2, "received notification for {:?} {:?}", notif, data); let preamble = VendorDependentPreamble::decode(data).map_err(|e| Error::PacketError(e))?; let data = &data[preamble.encoded_len()..]; if data.len() < preamble.parameter_length as usize { return Err(Error::UnexpectedResponse); } match notif { NotificationEventId::EventPlaybackStatusChanged => { let response = PlaybackStatusChangedNotificationResponse::decode(data) .map_err(|e| Error::PacketError(e))?; peer.write().handle_new_controller_notification_event( ControllerEvent::PlaybackStatusChanged(response.playback_status()), ); Ok(false) } NotificationEventId::EventTrackChanged => { let response = TrackChangedNotificationResponse::decode(data) .map_err(|e| Error::PacketError(e))?; peer.write().handle_new_controller_notification_event(ControllerEvent::TrackIdChanged( response.identifier(), )); Ok(false) } NotificationEventId::EventPlaybackPosChanged => { let response = PlaybackPosChangedNotificationResponse::decode(data) .map_err(|e| Error::PacketError(e))?; peer.write().handle_new_controller_notification_event( ControllerEvent::PlaybackPosChanged(response.position()), ); Ok(false) } NotificationEventId::EventVolumeChanged => { let response = VolumeChangedNotificationResponse::decode(data) .map_err(|e| Error::PacketError(e))?; peer.write().handle_new_controller_notification_event(ControllerEvent::VolumeChanged( response.volume(), )); Ok(false) } _ => Ok(true), } } /// Starts a task to attempt an outgoing L2CAP connection to remote's AVRCP control channel. /// The control channel should be in `Connecting` state before spawning this task. /// TODO(BT-2747): Fix a race where an incoming connection can come in while we are making an /// outgoing connection. Properly handle the case where we are attempting to connect to remote /// at the same time they make an incoming connection according to how the the spec says. fn start_make_connection_task(peer: Arc<RwLock<RemotePeer>>) { let peer = peer.clone(); fasync::spawn(async move { let (peer_id, profile_service) = { let peer_guard = peer.read(); // early return if we are not in `Connecting` match peer_guard.control_channel { PeerChannel::Connecting => {} _ => return, } (peer_guard.peer_id.clone(), peer_guard.profile_svc.clone()) }; match profile_service.connect_to_device(&peer_id, PSM_AVCTP as u16).await { Ok(socket) => { let mut peer_guard = peer.write(); match peer_guard.control_channel { PeerChannel::Connecting => match AvcPeer::new(socket) { Ok(peer) => { peer_guard.set_control_connection(peer); } Err(e) => { peer_guard.reset_connection(false); fx_log_err!("Unable to make peer from socket {}: {:?}", peer_id, e); } }, _ => { fx_log_info!( "incoming connection established while making outgoing {:?}", peer_id ); // an incoming l2cap connection was made while we were making an // outgoing one. Drop both connections per spec. peer_guard.reset_connection(false); } }; } Err(e) => { fx_log_err!("connect_to_device error {}: {:?}", peer_id, e); let mut peer_guard = peer.write(); if let PeerChannel::Connecting = peer_guard.control_channel { peer_guard.reset_connection(false); } } } }) } /// Checks for supported notification on the peer and registers for notifications. /// This is started on a remote peer when we have a connection and target profile descriptor. async fn pump_notifications(peer: Arc<RwLock<RemotePeer>>) { // events we support when speaking to a peer that supports the target profile. const SUPPORTED_NOTIFICATIONS: [NotificationEventId; 4] = [ NotificationEventId::EventPlaybackStatusChanged, NotificationEventId::EventTrackChanged, NotificationEventId::EventPlaybackPosChanged, NotificationEventId::EventVolumeChanged, ]; let supported_notifications: Vec<NotificationEventId> = SUPPORTED_NOTIFICATIONS.iter().cloned().collect(); // look up what notifications we support on this peer first. Consider updating this from // time to time. let remote_supported_notifications = match get_supported_events_internal(peer.clone()).await { Ok(x) => x, Err(_) => return, }; let supported_notifications: Vec<NotificationEventId> = remote_supported_notifications .into_iter() .filter(|k| supported_notifications.contains(k)) .collect(); let mut notification_streams = SelectAll::new(); for notif in supported_notifications { fx_vlog!(tag: "avrcp", 2, "creating notification stream for {:#?}", notif); let stream = NotificationStream::new(peer.clone(), notif, 1).map_ok(move |data| (notif, data)); notification_streams.push(stream); } pin_mut!(notification_streams); loop { if futures::select! { event_result = notification_streams.select_next_some() => { match event_result { Ok((notif, data)) => { handle_notification(&notif, &peer, &data[..]) .unwrap_or_else(|e| { fx_log_err!("Error decoding packet from peer {:?}", e); true} ) }, Err(Error::CommandNotSupported) => false, Err(_) => true, _=> true, } } complete => { true } } { break; } } fx_vlog!(tag: "avrcp", 2, "stopping notifications for {:#?}", peer.read().peer_id); } /// Starts a task to poll notifications on the remote peer. Aborted when the peer connection is /// reset. fn start_notifications_processing_task(peer: Arc<RwLock<RemotePeer>>) -> AbortHandle { let (handle, registration) = AbortHandle::new_pair(); fasync::spawn( Abortable::new( async move { pump_notifications(peer).await; }, registration, ) .map(|_| ()), ); handle } /// Starts a task to poll control messages from the peer. Aborted when the peer connection is /// reset. Started when we have a connection to the remote peer and we have any type of valid SDP /// profile from the peer. fn start_control_stream_processing_task(peer: Arc<RwLock<RemotePeer>>) -> AbortHandle { let (handle, registration) = AbortHandle::new_pair(); fasync::spawn( Abortable::new( async move { process_control_stream(peer).await; }, registration, ) .map(|_| ()), ); handle } /// State observer task around a remote peer. Takes a change stream from the remote peer that wakes /// the task whenever some state has changed on the peer. Swaps tasks such as making outgoing /// connections, processing the incoming control messages, and registering for notifications on the /// remote peer. pub(super) async fn
(peer: Arc<RwLock<RemotePeer>>) { fx_vlog!(tag: "avrcp", 2, "state_watcher starting"); let mut change_stream = peer.read().state_change_listener.take_change_stream(); let peer_weak = Arc::downgrade(&peer); drop(peer); let mut channel_processor_abort_handle: Option<AbortHandle> = None; let mut notification_poll_abort_handle: Option<AbortHandle> = None; while let Some(_) = change_stream.next().await { fx_vlog!(tag: "avrcp", 2, "state_watcher command received"); if let Some(peer) = peer_weak.upgrade() { let mut peer_guard = peer.write(); fx_vlog!(tag: "avrcp", 2, "make_connection control channel {:?}", peer_guard.control_channel); match peer_guard.control_channel { PeerChannel::Connecting => {} PeerChannel::Disconnected => { if let Some(ref abort_handle) = channel_processor_abort_handle { abort_handle.abort(); channel_processor_abort_handle = None; } if let Some(ref abort
state_watcher
identifier_name
global.rs
this peer is a stuck node, and we will kick out such kind of stuck peers. pub const STUCK_PEER_KICK_TIME: i64 = 2 * 3600 * 1000; /// If a peer's last seen time is 2 weeks ago we will forget such kind of defunct peers. const PEER_EXPIRATION_DAYS: i64 = 7 * 2; /// Constant that expresses defunct peer timeout in seconds to be used in checks. pub const PEER_EXPIRATION_REMOVE_TIME: i64 = PEER_EXPIRATION_DAYS * 24 * 3600; /// Trigger compaction check on average every day for all nodes. /// Randomized per node - roll the dice on every block to decide. /// Will compact the txhashset to remove pruned data. /// Will also remove old blocks and associated data from the database. /// For a node configured as "archival_mode = true" only the txhashset will be compacted. pub const COMPACTION_CHECK: u64 = DAY_HEIGHT; /// Number of blocks to reuse a txhashset zip for (automated testing and user testing). pub const TESTING_TXHASHSET_ARCHIVE_INTERVAL: u64 = 10; /// Number of blocks to reuse a txhashset zip for. pub const TXHASHSET_ARCHIVE_INTERVAL: u64 = 12 * 60; /// Types of chain a server can run with, dictates the genesis block and /// and mining parameters used. #[derive(Debug, Clone, Copy, Serialize, Deserialize, PartialEq)] pub enum ChainTypes { /// For CI testing AutomatedTesting, /// For User testing UserTesting, /// Protocol testing network Testnet, /// Main production network Mainnet, } impl ChainTypes { /// Short name representing the chain type ("test", "main", etc.) pub fn shortname(&self) -> String { match *self { ChainTypes::AutomatedTesting => "auto".to_owned(), ChainTypes::UserTesting => "user".to_owned(), ChainTypes::Testnet => "test".to_owned(), ChainTypes::Mainnet => "main".to_owned(), } } } impl Default for ChainTypes { fn default() -> ChainTypes { ChainTypes::Mainnet } } lazy_static! { /// Global chain_type that must be initialized once on node startup. /// This is accessed via get_chain_type() which allows the global value /// to be overridden on a per-thread basis (for testing). pub static ref GLOBAL_CHAIN_TYPE: OneTime<ChainTypes> = OneTime::new(); /// Global acccept fee base that must be initialized once on node startup. /// This is accessed via get_acccept_fee_base() which allows the global value /// to be overridden on a per-thread basis (for testing). pub static ref GLOBAL_ACCEPT_FEE_BASE: OneTime<u64> = OneTime::new(); /// Global future time limit that must be initialized once on node startup. /// This is accessed via get_future_time_limit() which allows the global value /// to be overridden on a per-thread basis (for testing). pub static ref GLOBAL_FUTURE_TIME_LIMIT: OneTime<u64> = OneTime::new(); /// Global feature flag for NRD kernel support. /// If enabled NRD kernels are treated as valid after HF3 (based on header version). /// If disabled NRD kernels are invalid regardless of header version or block height. pub static ref GLOBAL_NRD_FEATURE_ENABLED: OneTime<bool> = OneTime::new(); } thread_local! { /// Mainnet|Testnet|UserTesting|AutomatedTesting pub static CHAIN_TYPE: Cell<Option<ChainTypes>> = Cell::new(None); /// minimum transaction fee per unit of transaction weight for mempool acceptance pub static ACCEPT_FEE_BASE: Cell<Option<u64>> = Cell::new(None); /// maximum number of seconds into future for timestamp of block to be acceptable pub static FUTURE_TIME_LIMIT: Cell<Option<u64>> = Cell::new(None); /// Local feature flag for NRD kernel support. pub static NRD_FEATURE_ENABLED: Cell<Option<bool>> = Cell::new(None); } /// One time initialization of the global chain_type. /// Will panic if we attempt to re-initialize this (via OneTime). pub fn init_global_chain_type(new_type: ChainTypes) { GLOBAL_CHAIN_TYPE.init(new_type) } /// Set the global chain_type using an override pub fn set_global_chain_type(new_type: ChainTypes) { GLOBAL_CHAIN_TYPE.set(new_type, true); } /// Set the chain type on a per-thread basis via thread_local storage. pub fn set_local_chain_type(new_type: ChainTypes) { CHAIN_TYPE.with(|chain_type| chain_type.set(Some(new_type))) } /// Get the chain type via thread_local, fallback to global chain_type. pub fn get_chain_type() -> ChainTypes { CHAIN_TYPE.with(|chain_type| match chain_type.get() { None => { if !GLOBAL_CHAIN_TYPE.is_init() { panic!("GLOBAL_CHAIN_TYPE and CHAIN_TYPE unset. Consider set_local_chain_type() in tests."); } let chain_type = GLOBAL_CHAIN_TYPE.borrow(); set_local_chain_type(chain_type); chain_type } Some(chain_type) => chain_type, }) } /// Return genesis block for the active chain type pub fn get_genesis_block() -> Block { match get_chain_type() { ChainTypes::Mainnet => genesis::genesis_main(), ChainTypes::Testnet => genesis::genesis_test(), _ => genesis::genesis_dev(), } } /// One time initialization of the global future time limit /// Will panic if we attempt to re-initialize this (via OneTime). pub fn init_global_future_time_limit(new_ftl: u64) { GLOBAL_FUTURE_TIME_LIMIT.init(new_ftl) } /// The global future time limit may be reset again using the override pub fn set_global_future_time_limit(new_ftl: u64) { GLOBAL_FUTURE_TIME_LIMIT.set(new_ftl, true) } /// One time initialization of the global accept fee base /// Will panic if we attempt to re-initialize this (via OneTime). pub fn init_global_accept_fee_base(new_base: u64) { GLOBAL_ACCEPT_FEE_BASE.init(new_base) } /// The global accept fee base may be reset using override. pub fn set_global_accept_fee_base(new_base: u64) { GLOBAL_ACCEPT_FEE_BASE.set(new_base, true) } /// Set the accept fee base on a per-thread basis via thread_local storage. pub fn set_local_accept_fee_base(new_base: u64) { ACCEPT_FEE_BASE.with(|base| base.set(Some(new_base))) } /// Accept Fee Base /// Look at thread local config first. If not set fallback to global config. /// Default to grin-cent/20 if global config unset. pub fn get_accept_fee_base() -> u64 { ACCEPT_FEE_BASE.with(|base| match base.get() { None => { let base = if GLOBAL_ACCEPT_FEE_BASE.is_init() { GLOBAL_ACCEPT_FEE_BASE.borrow() } else { DEFAULT_ACCEPT_FEE_BASE }; set_local_accept_fee_base(base); base } Some(base) => base, }) } /// Set the future time limit on a per-thread basis via thread_local storage. pub fn set_local_future_time_limit(new_ftl: u64) { FUTURE_TIME_LIMIT.with(|ftl| ftl.set(Some(new_ftl))) } /// Future Time Limit (FTL) /// Look at thread local config first. If not set fallback to global config. /// Default to false if global config unset. pub fn get_future_time_limit() -> u64 { FUTURE_TIME_LIMIT.with(|ftl| match ftl.get() { None => { let ftl = if GLOBAL_FUTURE_TIME_LIMIT.is_init() { GLOBAL_FUTURE_TIME_LIMIT.borrow() } else { DEFAULT_FUTURE_TIME_LIMIT }; set_local_future_time_limit(ftl); ftl } Some(ftl) => ftl, }) } /// One time initialization of the global NRD feature flag. /// Will panic if we attempt to re-initialize this (via OneTime). pub fn init_global_nrd_enabled(enabled: bool)
/// Set the global NRD feature flag using override. pub fn set_global_nrd_enabled(enabled: bool) { GLOBAL_NRD_FEATURE_ENABLED.set(enabled, true) } /// Explicitly enable the local NRD feature flag. pub fn set_local_nrd_enabled(enabled: bool) { NRD_FEATURE_ENABLED.with(|flag| flag.set(Some(enabled))) } /// Is the NRD feature flag enabled? /// Look at thread local config first. If not set fallback to global config. /// Default to false if global config unset. pub fn is_nrd_enabled() -> bool { NRD_FEATURE_ENABLED.with(|flag| match flag.get() { None => { if GLOBAL_NRD_FEATURE_ENABLED.is_init() { let global_flag = GLOBAL_NRD_FEATURE_ENABLED.borrow(); flag.set(Some(global_flag)); global_flag } else { // Global config unset, default to false. false } } Some(flag) => flag, }) } /// Return either a cuckaroo* context or a cuckatoo context /// Single change point pub fn create_pow_context<T>( height: u64, edge_bits:
{ GLOBAL_NRD_FEATURE_ENABLED.init(enabled) }
identifier_body
global.rs
this peer is a stuck node, and we will kick out such kind of stuck peers. pub const STUCK_PEER_KICK_TIME: i64 = 2 * 3600 * 1000; /// If a peer's last seen time is 2 weeks ago we will forget such kind of defunct peers. const PEER_EXPIRATION_DAYS: i64 = 7 * 2; /// Constant that expresses defunct peer timeout in seconds to be used in checks. pub const PEER_EXPIRATION_REMOVE_TIME: i64 = PEER_EXPIRATION_DAYS * 24 * 3600; /// Trigger compaction check on average every day for all nodes. /// Randomized per node - roll the dice on every block to decide. /// Will compact the txhashset to remove pruned data. /// Will also remove old blocks and associated data from the database. /// For a node configured as "archival_mode = true" only the txhashset will be compacted. pub const COMPACTION_CHECK: u64 = DAY_HEIGHT; /// Number of blocks to reuse a txhashset zip for (automated testing and user testing). pub const TESTING_TXHASHSET_ARCHIVE_INTERVAL: u64 = 10; /// Number of blocks to reuse a txhashset zip for. pub const TXHASHSET_ARCHIVE_INTERVAL: u64 = 12 * 60; /// Types of chain a server can run with, dictates the genesis block and /// and mining parameters used. #[derive(Debug, Clone, Copy, Serialize, Deserialize, PartialEq)] pub enum ChainTypes { /// For CI testing AutomatedTesting, /// For User testing UserTesting, /// Protocol testing network Testnet, /// Main production network Mainnet, } impl ChainTypes { /// Short name representing the chain type ("test", "main", etc.) pub fn shortname(&self) -> String { match *self { ChainTypes::AutomatedTesting => "auto".to_owned(), ChainTypes::UserTesting => "user".to_owned(), ChainTypes::Testnet => "test".to_owned(), ChainTypes::Mainnet => "main".to_owned(), } } } impl Default for ChainTypes { fn default() -> ChainTypes { ChainTypes::Mainnet } } lazy_static! { /// Global chain_type that must be initialized once on node startup. /// This is accessed via get_chain_type() which allows the global value /// to be overridden on a per-thread basis (for testing). pub static ref GLOBAL_CHAIN_TYPE: OneTime<ChainTypes> = OneTime::new(); /// Global acccept fee base that must be initialized once on node startup. /// This is accessed via get_acccept_fee_base() which allows the global value /// to be overridden on a per-thread basis (for testing). pub static ref GLOBAL_ACCEPT_FEE_BASE: OneTime<u64> = OneTime::new(); /// Global future time limit that must be initialized once on node startup. /// This is accessed via get_future_time_limit() which allows the global value /// to be overridden on a per-thread basis (for testing). pub static ref GLOBAL_FUTURE_TIME_LIMIT: OneTime<u64> = OneTime::new(); /// Global feature flag for NRD kernel support. /// If enabled NRD kernels are treated as valid after HF3 (based on header version). /// If disabled NRD kernels are invalid regardless of header version or block height. pub static ref GLOBAL_NRD_FEATURE_ENABLED: OneTime<bool> = OneTime::new(); } thread_local! { /// Mainnet|Testnet|UserTesting|AutomatedTesting pub static CHAIN_TYPE: Cell<Option<ChainTypes>> = Cell::new(None); /// minimum transaction fee per unit of transaction weight for mempool acceptance pub static ACCEPT_FEE_BASE: Cell<Option<u64>> = Cell::new(None); /// maximum number of seconds into future for timestamp of block to be acceptable pub static FUTURE_TIME_LIMIT: Cell<Option<u64>> = Cell::new(None); /// Local feature flag for NRD kernel support. pub static NRD_FEATURE_ENABLED: Cell<Option<bool>> = Cell::new(None); } /// One time initialization of the global chain_type. /// Will panic if we attempt to re-initialize this (via OneTime). pub fn init_global_chain_type(new_type: ChainTypes) { GLOBAL_CHAIN_TYPE.init(new_type) } /// Set the global chain_type using an override pub fn set_global_chain_type(new_type: ChainTypes) { GLOBAL_CHAIN_TYPE.set(new_type, true); } /// Set the chain type on a per-thread basis via thread_local storage. pub fn set_local_chain_type(new_type: ChainTypes) { CHAIN_TYPE.with(|chain_type| chain_type.set(Some(new_type))) } /// Get the chain type via thread_local, fallback to global chain_type. pub fn
() -> ChainTypes { CHAIN_TYPE.with(|chain_type| match chain_type.get() { None => { if !GLOBAL_CHAIN_TYPE.is_init() { panic!("GLOBAL_CHAIN_TYPE and CHAIN_TYPE unset. Consider set_local_chain_type() in tests."); } let chain_type = GLOBAL_CHAIN_TYPE.borrow(); set_local_chain_type(chain_type); chain_type } Some(chain_type) => chain_type, }) } /// Return genesis block for the active chain type pub fn get_genesis_block() -> Block { match get_chain_type() { ChainTypes::Mainnet => genesis::genesis_main(), ChainTypes::Testnet => genesis::genesis_test(), _ => genesis::genesis_dev(), } } /// One time initialization of the global future time limit /// Will panic if we attempt to re-initialize this (via OneTime). pub fn init_global_future_time_limit(new_ftl: u64) { GLOBAL_FUTURE_TIME_LIMIT.init(new_ftl) } /// The global future time limit may be reset again using the override pub fn set_global_future_time_limit(new_ftl: u64) { GLOBAL_FUTURE_TIME_LIMIT.set(new_ftl, true) } /// One time initialization of the global accept fee base /// Will panic if we attempt to re-initialize this (via OneTime). pub fn init_global_accept_fee_base(new_base: u64) { GLOBAL_ACCEPT_FEE_BASE.init(new_base) } /// The global accept fee base may be reset using override. pub fn set_global_accept_fee_base(new_base: u64) { GLOBAL_ACCEPT_FEE_BASE.set(new_base, true) } /// Set the accept fee base on a per-thread basis via thread_local storage. pub fn set_local_accept_fee_base(new_base: u64) { ACCEPT_FEE_BASE.with(|base| base.set(Some(new_base))) } /// Accept Fee Base /// Look at thread local config first. If not set fallback to global config. /// Default to grin-cent/20 if global config unset. pub fn get_accept_fee_base() -> u64 { ACCEPT_FEE_BASE.with(|base| match base.get() { None => { let base = if GLOBAL_ACCEPT_FEE_BASE.is_init() { GLOBAL_ACCEPT_FEE_BASE.borrow() } else { DEFAULT_ACCEPT_FEE_BASE }; set_local_accept_fee_base(base); base } Some(base) => base, }) } /// Set the future time limit on a per-thread basis via thread_local storage. pub fn set_local_future_time_limit(new_ftl: u64) { FUTURE_TIME_LIMIT.with(|ftl| ftl.set(Some(new_ftl))) } /// Future Time Limit (FTL) /// Look at thread local config first. If not set fallback to global config. /// Default to false if global config unset. pub fn get_future_time_limit() -> u64 { FUTURE_TIME_LIMIT.with(|ftl| match ftl.get() { None => { let ftl = if GLOBAL_FUTURE_TIME_LIMIT.is_init() { GLOBAL_FUTURE_TIME_LIMIT.borrow() } else { DEFAULT_FUTURE_TIME_LIMIT }; set_local_future_time_limit(ftl); ftl } Some(ftl) => ftl, }) } /// One time initialization of the global NRD feature flag. /// Will panic if we attempt to re-initialize this (via OneTime). pub fn init_global_nrd_enabled(enabled: bool) { GLOBAL_NRD_FEATURE_ENABLED.init(enabled) } /// Set the global NRD feature flag using override. pub fn set_global_nrd_enabled(enabled: bool) { GLOBAL_NRD_FEATURE_ENABLED.set(enabled, true) } /// Explicitly enable the local NRD feature flag. pub fn set_local_nrd_enabled(enabled: bool) { NRD_FEATURE_ENABLED.with(|flag| flag.set(Some(enabled))) } /// Is the NRD feature flag enabled? /// Look at thread local config first. If not set fallback to global config. /// Default to false if global config unset. pub fn is_nrd_enabled() -> bool { NRD_FEATURE_ENABLED.with(|flag| match flag.get() { None => { if GLOBAL_NRD_FEATURE_ENABLED.is_init() { let global_flag = GLOBAL_NRD_FEATURE_ENABLED.borrow(); flag.set(Some(global_flag)); global_flag } else { // Global config unset, default to false. false } } Some(flag) => flag, }) } /// Return either a cuckaroo* context or a cuckatoo context /// Single change point pub fn create_pow_context<T>( height: u64, edge_bits: u
get_chain_type
identifier_name
global.rs
fn set_global_nrd_enabled(enabled: bool) { GLOBAL_NRD_FEATURE_ENABLED.set(enabled, true) } /// Explicitly enable the local NRD feature flag. pub fn set_local_nrd_enabled(enabled: bool) { NRD_FEATURE_ENABLED.with(|flag| flag.set(Some(enabled))) } /// Is the NRD feature flag enabled? /// Look at thread local config first. If not set fallback to global config. /// Default to false if global config unset. pub fn is_nrd_enabled() -> bool { NRD_FEATURE_ENABLED.with(|flag| match flag.get() { None => { if GLOBAL_NRD_FEATURE_ENABLED.is_init() { let global_flag = GLOBAL_NRD_FEATURE_ENABLED.borrow(); flag.set(Some(global_flag)); global_flag } else { // Global config unset, default to false. false } } Some(flag) => flag, }) } /// Return either a cuckaroo* context or a cuckatoo context /// Single change point pub fn create_pow_context<T>( height: u64, edge_bits: u8, proof_size: usize, max_sols: u32, ) -> Result<Box<dyn PoWContext>, pow::Error> { let chain_type = get_chain_type(); if chain_type == ChainTypes::Mainnet || chain_type == ChainTypes::Testnet { // Mainnet and Testnet have Cuckatoo31+ for AF and Cuckaroo{,d,m,z}29 for AR if edge_bits > 29 { new_cuckatoo_ctx(edge_bits, proof_size, max_sols) } else { match header_version(height) { HeaderVersion(1) => new_cuckaroo_ctx(edge_bits, proof_size), HeaderVersion(2) => new_cuckarood_ctx(edge_bits, proof_size), HeaderVersion(3) => new_cuckaroom_ctx(edge_bits, proof_size), HeaderVersion(4) => new_cuckarooz_ctx(edge_bits, proof_size), _ => no_cuckaroo_ctx(), } } } else { // Everything else is Cuckatoo only new_cuckatoo_ctx(edge_bits, proof_size, max_sols) } } /// The minimum acceptable edge_bits pub fn min_edge_bits() -> u8 { match get_chain_type() { ChainTypes::AutomatedTesting => AUTOMATED_TESTING_MIN_EDGE_BITS, ChainTypes::UserTesting => USER_TESTING_MIN_EDGE_BITS, _ => DEFAULT_MIN_EDGE_BITS, } } /// Reference edge_bits used to compute factor on higher Cuck(at)oo graph sizes, /// while the min_edge_bits can be changed on a soft fork, changing /// base_edge_bits is a hard fork. pub fn base_edge_bits() -> u8 { match get_chain_type() { ChainTypes::AutomatedTesting => AUTOMATED_TESTING_MIN_EDGE_BITS, ChainTypes::UserTesting => USER_TESTING_MIN_EDGE_BITS, _ => BASE_EDGE_BITS, } } /// The proofsize pub fn proofsize() -> usize { match get_chain_type() { ChainTypes::AutomatedTesting => AUTOMATED_TESTING_PROOF_SIZE, ChainTypes::UserTesting => USER_TESTING_PROOF_SIZE, _ => PROOFSIZE, } } /// Coinbase maturity for coinbases to be spent pub fn coinbase_maturity() -> u64 { match get_chain_type() { ChainTypes::AutomatedTesting => AUTOMATED_TESTING_COINBASE_MATURITY, ChainTypes::UserTesting => USER_TESTING_COINBASE_MATURITY, _ => COINBASE_MATURITY, } } /// Initial mining difficulty pub fn initial_block_difficulty() -> u64 { match get_chain_type() { ChainTypes::AutomatedTesting => TESTING_INITIAL_DIFFICULTY, ChainTypes::UserTesting => TESTING_INITIAL_DIFFICULTY, ChainTypes::Testnet => INITIAL_DIFFICULTY, ChainTypes::Mainnet => INITIAL_DIFFICULTY, } } /// Initial mining secondary scale pub fn initial_graph_weight() -> u32 { match get_chain_type() { ChainTypes::AutomatedTesting => graph_weight(0, AUTOMATED_TESTING_MIN_EDGE_BITS) as u32, ChainTypes::UserTesting => graph_weight(0, USER_TESTING_MIN_EDGE_BITS) as u32, ChainTypes::Testnet => graph_weight(0, SECOND_POW_EDGE_BITS) as u32, ChainTypes::Mainnet => graph_weight(0, SECOND_POW_EDGE_BITS) as u32, } } /// Minimum valid graph weight post HF4 pub fn min_wtema_graph_weight() -> u64 { match get_chain_type() { ChainTypes::AutomatedTesting => graph_weight(0, AUTOMATED_TESTING_MIN_EDGE_BITS), ChainTypes::UserTesting => graph_weight(0, USER_TESTING_MIN_EDGE_BITS), ChainTypes::Testnet => graph_weight(0, SECOND_POW_EDGE_BITS), ChainTypes::Mainnet => C32_GRAPH_WEIGHT, } } /// Maximum allowed block weight. pub fn max_block_weight() -> u64 { match get_chain_type() { ChainTypes::AutomatedTesting => TESTING_MAX_BLOCK_WEIGHT, ChainTypes::UserTesting => TESTING_MAX_BLOCK_WEIGHT, ChainTypes::Testnet => MAX_BLOCK_WEIGHT, ChainTypes::Mainnet => MAX_BLOCK_WEIGHT, } } /// Maximum allowed transaction weight (1 weight unit ~= 32 bytes) pub fn max_tx_weight() -> u64 { let coinbase_weight = OUTPUT_WEIGHT + KERNEL_WEIGHT; max_block_weight().saturating_sub(coinbase_weight) as u64 } /// Horizon at which we can cut-through and do full local pruning pub fn cut_through_horizon() -> u32 { match get_chain_type() { ChainTypes::AutomatedTesting => AUTOMATED_TESTING_CUT_THROUGH_HORIZON, ChainTypes::UserTesting => USER_TESTING_CUT_THROUGH_HORIZON, _ => CUT_THROUGH_HORIZON, } } /// Threshold at which we can request a txhashset (and full blocks from) pub fn state_sync_threshold() -> u32 { match get_chain_type() { ChainTypes::AutomatedTesting => TESTING_STATE_SYNC_THRESHOLD, ChainTypes::UserTesting => TESTING_STATE_SYNC_THRESHOLD, _ => STATE_SYNC_THRESHOLD, } } /// Number of blocks to reuse a txhashset zip for. pub fn txhashset_archive_interval() -> u64 { match get_chain_type() { ChainTypes::AutomatedTesting => TESTING_TXHASHSET_ARCHIVE_INTERVAL, ChainTypes::UserTesting => TESTING_TXHASHSET_ARCHIVE_INTERVAL, _ => TXHASHSET_ARCHIVE_INTERVAL, } } /// Are we in production mode? /// Production defined as a live public network, testnet[n] or mainnet. pub fn is_production_mode() -> bool { match get_chain_type() { ChainTypes::Testnet => true, ChainTypes::Mainnet => true, _ => false, } } /// Are we in testnet? /// Note: We do not have a corresponding is_mainnet() as we want any tests to be as close /// as possible to "mainnet" configuration as possible. /// We want to avoid missing any mainnet only code paths. pub fn is_testnet() -> bool { match get_chain_type() { ChainTypes::Testnet => true, _ => false, } } /// Converts an iterator of block difficulty data to more a more manageable /// vector and pads if needed (which will) only be needed for the first few /// blocks after genesis pub fn difficulty_data_to_vector<T>(cursor: T) -> Vec<HeaderDifficultyInfo> where T: IntoIterator<Item = HeaderDifficultyInfo>, { // Convert iterator to vector, so we can append to it if necessary let needed_block_count = DMA_WINDOW as usize + 1; let mut last_n: Vec<HeaderDifficultyInfo> = cursor.into_iter().take(needed_block_count).collect(); // Only needed just after blockchain launch... basically ensures there's // always enough data by simulating perfectly timed pre-genesis // blocks at the genesis difficulty as needed. let n = last_n.len(); if needed_block_count > n { let last_ts_delta = if n > 1 { last_n[0].timestamp - last_n[1].timestamp } else { BLOCK_TIME_SEC }; let last_diff = last_n[0].difficulty; // fill in simulated blocks with values from the previous real block let mut last_ts = last_n.last().unwrap().timestamp; for _ in n..needed_block_count { last_ts = last_ts.saturating_sub(last_ts_delta); last_n.push(HeaderDifficultyInfo::from_ts_diff(last_ts, last_diff)); } } last_n.reverse(); last_n }
/// Calculates the size of a header (in bytes) given a number of edge bits in the PoW #[inline] pub fn header_size_bytes(edge_bits: u8) -> usize { let size = 2 + 2 * 8 + 5 * 32 + 32 + 2 * 8; let proof_size = 8 + 4 + 8 + 1 + Proof::pack_len(edge_bits);
random_line_split
simple-filters.go
n *sam.Alignment) bool { return dictTable[aln.RNAME] } } } // ReplaceReferenceSequenceDictionaryFromSamFile returns a filter for // replacing the reference sequence dictionary in a Header with one // parsed from the given SAM/DICT file. func ReplaceReferenceSequenceDictionaryFromSamFile(samFile string) (f sam.Filter, err error) { input, err := sam.Open(samFile, true) if err != nil { return nil, err } defer func() { nerr := input.Close() if err == nil { err = nerr } }() header, _, err := sam.ParseHeader(input.Reader) if err != nil { return nil, err } return ReplaceReferenceSequenceDictionary(header.SQ), nil } // RemoveUnmappedReads is a filter for removing unmapped sam-alignment // instances, based on FLAG. func RemoveUnmappedReads(_ *sam.Header) sam.AlignmentFilter { return func(aln *sam.Alignment) bool { return (aln.FLAG & sam.Unmapped) == 0 } } // RemoveUnmappedReadsStrict is a filter for removing unmapped // sam-alignment instances, based on FLAG, or POS=0, or RNAME=*. func RemoveUnmappedReadsStrict(_ *sam.Header) sam.AlignmentFilter { return func(aln *sam.Alignment) bool { return ((aln.FLAG & sam.Unmapped) == 0) && (aln.POS != 0) && (aln.RNAME != "*") } } // RemoveNonExactMappingReads is a filter that removes all reads that // are not exact matches with the reference (soft-clipping ok), based // on CIGAR string (only M and S allowed). func RemoveNonExactMappingReads(_ *sam.Header) sam.AlignmentFilter { return func(aln *sam.Alignment) bool { return !strings.ContainsAny(aln.CIGAR, "IDNHPX=") } } // Symbols for optional fields used for determining exact matches. See // http://samtools.github.io/hts-specs/SAMv1.pdf - Section 1.5. var ( X0 = utils.Intern("X0") X1 = utils.Intern("X1") XM = utils.Intern("XM") XO = utils.Intern("XO") XG = utils.Intern("XG") ) // RemoveNonExactMappingReadsStrict is a filter that removes all reads // that are not exact matches with the reference, based on the // optional fields X0=1 (unique mapping), X1=0 (no suboptimal hit), // XM=0 (no mismatch), XO=0 (no gap opening), XG=0 (no gap extension). func RemoveNonExactMappingReadsStrict(header *sam.Header) sam.AlignmentFilter { return func(aln *sam.Alignment) bool { if x0, ok := aln.TAGS.Get(X0); !ok || x0.(int32) != 1 { return false } if x1, ok := aln.TAGS.Get(X1); !ok || x1.(int32) != 0 { return false } if xm, ok := aln.TAGS.Get(XM); !ok || xm.(int32) != 0 { return false } if xo, ok := aln.TAGS.Get(XO); !ok || xo.(int32) != 0
if xg, ok := aln.TAGS.Get(XG); !ok || xg.(int32) != 0 { return false } return true } } // RemoveDuplicateReads is a filter for removing duplicate // sam-alignment instances, based on FLAG. func RemoveDuplicateReads(_ *sam.Header) sam.AlignmentFilter { return func(aln *sam.Alignment) bool { return (aln.FLAG & sam.Duplicate) == 0 } } var sr = utils.Intern("sr") // RemoveOptionalReads is a filter for removing alignments that // represent optional information in elPrep. func RemoveOptionalReads(header *sam.Header) sam.AlignmentFilter { if _, found := header.UserRecords["@sr"]; found { delete(header.UserRecords, "@sr") return func(aln *sam.Alignment) bool { _, found := aln.TAGS.Get(sr); return !found } } return nil } // AddOrReplaceReadGroup returns a filter for adding or replacing the // read group both in the Header and in each Alignment. func AddOrReplaceReadGroup(readGroup utils.StringMap) sam.Filter { return func(header *sam.Header) sam.AlignmentFilter { header.RG = []utils.StringMap{readGroup} id := readGroup["ID"] return func(aln *sam.Alignment) bool { aln.SetRG(id); return true } } } // AddPGLine returns a filter for adding a @PG tag to a Header, and // ensuring that it is the first one in the chain. func AddPGLine(newPG utils.StringMap) sam.Filter { return func(header *sam.Header) sam.AlignmentFilter { id := newPG["ID"] for utils.Find(header.PG, func(entry utils.StringMap) bool { return entry["ID"] == id }) >= 0 { id += " " id += strconv.FormatInt(rand.Int63n(0x10000), 16) } newPG["ID"] = id for _, PG := range header.PG { nextID := PG["ID"] if pos := utils.Find(header.PG, func(entry utils.StringMap) bool { return entry["PP"] == nextID }); pos < 0 { newPG["PP"] = nextID break } } header.PG = append(header.PG, newPG) return nil } } // RenameChromosomes is a filter for prepending "chr" to the reference // sequence names in a Header, and in RNAME and RNEXT in each // Alignment. func RenameChromosomes(header *sam.Header) sam.AlignmentFilter { for _, entry := range header.SQ { if sn, found := entry["SN"]; found { entry["SN"] = "chr" + sn } } return func(aln *sam.Alignment) bool { if (aln.RNAME != "=") && (aln.RNAME != "*") { aln.RNAME = "chr" + aln.RNAME } if (aln.RNEXT != "=") && (aln.RNEXT != "*") { aln.RNEXT = "chr" + aln.RNEXT } return true } } // AddREFID is a filter for adding the refid (index in the reference // sequence dictionary) to alignments as temporary values. func AddREFID(header *sam.Header) sam.AlignmentFilter { dictTable := make(map[string]int32) for index, entry := range header.SQ { dictTable[entry["SN"]] = int32(index) } return func(aln *sam.Alignment) bool { value, found := dictTable[aln.RNAME] if !found { value = -1 } aln.SetREFID(value) return true } } // RemoveOptionalFields returns a filter for removing optional fields // in an alignment. func RemoveOptionalFields(tags []string) sam.Filter { if len(tags) == 0 { return nil } // Intern the tags once. var optionals []utils.Symbol for _, tag := range tags { optionals = append(optionals, utils.Intern(tag)) } return func(header *sam.Header) sam.AlignmentFilter { return func(aln *sam.Alignment) bool { aln.TAGS, _ = aln.TAGS.DeleteIf(func(key utils.Symbol, val interface{}) bool { for _, tag := range optionals { if tag == key { return true } } return false }) return true } } } // KeepOptionalFields returns a filter for removing all but a list of // given optional fields in an alignment. func KeepOptionalFields(tags []string) sam.Filter { if len(tags) == 0 { return func(header *sam.Header) sam.AlignmentFilter { return func(aln *sam.Alignment) bool { aln.TAGS = nil return true } } } // Intern the tags once. var optionals []utils.Symbol for _, tag := range tags { optionals = append(optionals, utils.Intern(tag)) } return func(header *sam.Header) sam.AlignmentFilter { return func(aln *sam.Alignment) bool { aln.TAGS, _ = aln.TAGS.DeleteIf(func(key utils.Symbol, val interface{}) bool { for _, tag := range optionals { if tag == key { return false } } return true }) return true } } } // CleanSam is a filter for soft-clipping an alignment at the end of a // reference sequence, and set MAPQ to 0 if unmapped. func CleanSam(header *sam.Header) sam.AlignmentFilter { referenceSequenceTable := make(map[string]int32) for _, sn := range header.SQ { referenceSequenceTable[sn["SN"]], _ = sam.SQLN(sn) } return
{ return false }
conditional_block
simple-filters.go
(aln *sam.Alignment) bool { return dictTable[aln.RNAME] } } } // ReplaceReferenceSequenceDictionaryFromSamFile returns a filter for // replacing the reference sequence dictionary in a Header with one // parsed from the given SAM/DICT file. func ReplaceReferenceSequenceDictionaryFromSamFile(samFile string) (f sam.Filter, err error) { input, err := sam.Open(samFile, true) if err != nil { return nil, err } defer func() { nerr := input.Close() if err == nil { err = nerr } }() header, _, err := sam.ParseHeader(input.Reader) if err != nil { return nil, err } return ReplaceReferenceSequenceDictionary(header.SQ), nil } // RemoveUnmappedReads is a filter for removing unmapped sam-alignment // instances, based on FLAG. func RemoveUnmappedReads(_ *sam.Header) sam.AlignmentFilter { return func(aln *sam.Alignment) bool { return (aln.FLAG & sam.Unmapped) == 0 } } // RemoveUnmappedReadsStrict is a filter for removing unmapped // sam-alignment instances, based on FLAG, or POS=0, or RNAME=*. func
(_ *sam.Header) sam.AlignmentFilter { return func(aln *sam.Alignment) bool { return ((aln.FLAG & sam.Unmapped) == 0) && (aln.POS != 0) && (aln.RNAME != "*") } } // RemoveNonExactMappingReads is a filter that removes all reads that // are not exact matches with the reference (soft-clipping ok), based // on CIGAR string (only M and S allowed). func RemoveNonExactMappingReads(_ *sam.Header) sam.AlignmentFilter { return func(aln *sam.Alignment) bool { return !strings.ContainsAny(aln.CIGAR, "IDNHPX=") } } // Symbols for optional fields used for determining exact matches. See // http://samtools.github.io/hts-specs/SAMv1.pdf - Section 1.5. var ( X0 = utils.Intern("X0") X1 = utils.Intern("X1") XM = utils.Intern("XM") XO = utils.Intern("XO") XG = utils.Intern("XG") ) // RemoveNonExactMappingReadsStrict is a filter that removes all reads // that are not exact matches with the reference, based on the // optional fields X0=1 (unique mapping), X1=0 (no suboptimal hit), // XM=0 (no mismatch), XO=0 (no gap opening), XG=0 (no gap extension). func RemoveNonExactMappingReadsStrict(header *sam.Header) sam.AlignmentFilter { return func(aln *sam.Alignment) bool { if x0, ok := aln.TAGS.Get(X0); !ok || x0.(int32) != 1 { return false } if x1, ok := aln.TAGS.Get(X1); !ok || x1.(int32) != 0 { return false } if xm, ok := aln.TAGS.Get(XM); !ok || xm.(int32) != 0 { return false } if xo, ok := aln.TAGS.Get(XO); !ok || xo.(int32) != 0 { return false } if xg, ok := aln.TAGS.Get(XG); !ok || xg.(int32) != 0 { return false } return true } } // RemoveDuplicateReads is a filter for removing duplicate // sam-alignment instances, based on FLAG. func RemoveDuplicateReads(_ *sam.Header) sam.AlignmentFilter { return func(aln *sam.Alignment) bool { return (aln.FLAG & sam.Duplicate) == 0 } } var sr = utils.Intern("sr") // RemoveOptionalReads is a filter for removing alignments that // represent optional information in elPrep. func RemoveOptionalReads(header *sam.Header) sam.AlignmentFilter { if _, found := header.UserRecords["@sr"]; found { delete(header.UserRecords, "@sr") return func(aln *sam.Alignment) bool { _, found := aln.TAGS.Get(sr); return !found } } return nil } // AddOrReplaceReadGroup returns a filter for adding or replacing the // read group both in the Header and in each Alignment. func AddOrReplaceReadGroup(readGroup utils.StringMap) sam.Filter { return func(header *sam.Header) sam.AlignmentFilter { header.RG = []utils.StringMap{readGroup} id := readGroup["ID"] return func(aln *sam.Alignment) bool { aln.SetRG(id); return true } } } // AddPGLine returns a filter for adding a @PG tag to a Header, and // ensuring that it is the first one in the chain. func AddPGLine(newPG utils.StringMap) sam.Filter { return func(header *sam.Header) sam.AlignmentFilter { id := newPG["ID"] for utils.Find(header.PG, func(entry utils.StringMap) bool { return entry["ID"] == id }) >= 0 { id += " " id += strconv.FormatInt(rand.Int63n(0x10000), 16) } newPG["ID"] = id for _, PG := range header.PG { nextID := PG["ID"] if pos := utils.Find(header.PG, func(entry utils.StringMap) bool { return entry["PP"] == nextID }); pos < 0 { newPG["PP"] = nextID break } } header.PG = append(header.PG, newPG) return nil } } // RenameChromosomes is a filter for prepending "chr" to the reference // sequence names in a Header, and in RNAME and RNEXT in each // Alignment. func RenameChromosomes(header *sam.Header) sam.AlignmentFilter { for _, entry := range header.SQ { if sn, found := entry["SN"]; found { entry["SN"] = "chr" + sn } } return func(aln *sam.Alignment) bool { if (aln.RNAME != "=") && (aln.RNAME != "*") { aln.RNAME = "chr" + aln.RNAME } if (aln.RNEXT != "=") && (aln.RNEXT != "*") { aln.RNEXT = "chr" + aln.RNEXT } return true } } // AddREFID is a filter for adding the refid (index in the reference // sequence dictionary) to alignments as temporary values. func AddREFID(header *sam.Header) sam.AlignmentFilter { dictTable := make(map[string]int32) for index, entry := range header.SQ { dictTable[entry["SN"]] = int32(index) } return func(aln *sam.Alignment) bool { value, found := dictTable[aln.RNAME] if !found { value = -1 } aln.SetREFID(value) return true } } // RemoveOptionalFields returns a filter for removing optional fields // in an alignment. func RemoveOptionalFields(tags []string) sam.Filter { if len(tags) == 0 { return nil } // Intern the tags once. var optionals []utils.Symbol for _, tag := range tags { optionals = append(optionals, utils.Intern(tag)) } return func(header *sam.Header) sam.AlignmentFilter { return func(aln *sam.Alignment) bool { aln.TAGS, _ = aln.TAGS.DeleteIf(func(key utils.Symbol, val interface{}) bool { for _, tag := range optionals { if tag == key { return true } } return false }) return true } } } // KeepOptionalFields returns a filter for removing all but a list of // given optional fields in an alignment. func KeepOptionalFields(tags []string) sam.Filter { if len(tags) == 0 { return func(header *sam.Header) sam.AlignmentFilter { return func(aln *sam.Alignment) bool { aln.TAGS = nil return true } } } // Intern the tags once. var optionals []utils.Symbol for _, tag := range tags { optionals = append(optionals, utils.Intern(tag)) } return func(header *sam.Header) sam.AlignmentFilter { return func(aln *sam.Alignment) bool { aln.TAGS, _ = aln.TAGS.DeleteIf(func(key utils.Symbol, val interface{}) bool { for _, tag := range optionals { if tag == key { return false } } return true }) return true } } } // CleanSam is a filter for soft-clipping an alignment at the end of a // reference sequence, and set MAPQ to 0 if unmapped. func CleanSam(header *sam.Header) sam.AlignmentFilter { referenceSequenceTable := make(map[string]int32) for _, sn := range header.SQ { referenceSequenceTable[sn["SN"]], _ = sam.SQLN(sn) } return
RemoveUnmappedReadsStrict
identifier_name
simple-filters.go
n *sam.Alignment) bool { return dictTable[aln.RNAME] } } } // ReplaceReferenceSequenceDictionaryFromSamFile returns a filter for // replacing the reference sequence dictionary in a Header with one // parsed from the given SAM/DICT file. func ReplaceReferenceSequenceDictionaryFromSamFile(samFile string) (f sam.Filter, err error) { input, err := sam.Open(samFile, true) if err != nil { return nil, err } defer func() { nerr := input.Close() if err == nil { err = nerr } }() header, _, err := sam.ParseHeader(input.Reader) if err != nil { return nil, err } return ReplaceReferenceSequenceDictionary(header.SQ), nil } // RemoveUnmappedReads is a filter for removing unmapped sam-alignment // instances, based on FLAG. func RemoveUnmappedReads(_ *sam.Header) sam.AlignmentFilter { return func(aln *sam.Alignment) bool { return (aln.FLAG & sam.Unmapped) == 0 } } // RemoveUnmappedReadsStrict is a filter for removing unmapped // sam-alignment instances, based on FLAG, or POS=0, or RNAME=*. func RemoveUnmappedReadsStrict(_ *sam.Header) sam.AlignmentFilter { return func(aln *sam.Alignment) bool { return ((aln.FLAG & sam.Unmapped) == 0) && (aln.POS != 0) && (aln.RNAME != "*") } } // RemoveNonExactMappingReads is a filter that removes all reads that // are not exact matches with the reference (soft-clipping ok), based // on CIGAR string (only M and S allowed). func RemoveNonExactMappingReads(_ *sam.Header) sam.AlignmentFilter { return func(aln *sam.Alignment) bool { return !strings.ContainsAny(aln.CIGAR, "IDNHPX=") } } // Symbols for optional fields used for determining exact matches. See // http://samtools.github.io/hts-specs/SAMv1.pdf - Section 1.5. var ( X0 = utils.Intern("X0") X1 = utils.Intern("X1") XM = utils.Intern("XM") XO = utils.Intern("XO") XG = utils.Intern("XG") ) // RemoveNonExactMappingReadsStrict is a filter that removes all reads // that are not exact matches with the reference, based on the // optional fields X0=1 (unique mapping), X1=0 (no suboptimal hit), // XM=0 (no mismatch), XO=0 (no gap opening), XG=0 (no gap extension). func RemoveNonExactMappingReadsStrict(header *sam.Header) sam.AlignmentFilter { return func(aln *sam.Alignment) bool { if x0, ok := aln.TAGS.Get(X0); !ok || x0.(int32) != 1 { return false } if x1, ok := aln.TAGS.Get(X1); !ok || x1.(int32) != 0 { return false } if xm, ok := aln.TAGS.Get(XM); !ok || xm.(int32) != 0 { return false } if xo, ok := aln.TAGS.Get(XO); !ok || xo.(int32) != 0 { return false } if xg, ok := aln.TAGS.Get(XG); !ok || xg.(int32) != 0 { return false } return true } } // RemoveDuplicateReads is a filter for removing duplicate // sam-alignment instances, based on FLAG. func RemoveDuplicateReads(_ *sam.Header) sam.AlignmentFilter { return func(aln *sam.Alignment) bool { return (aln.FLAG & sam.Duplicate) == 0 } } var sr = utils.Intern("sr") // RemoveOptionalReads is a filter for removing alignments that // represent optional information in elPrep. func RemoveOptionalReads(header *sam.Header) sam.AlignmentFilter { if _, found := header.UserRecords["@sr"]; found { delete(header.UserRecords, "@sr") return func(aln *sam.Alignment) bool { _, found := aln.TAGS.Get(sr); return !found } } return nil } // AddOrReplaceReadGroup returns a filter for adding or replacing the // read group both in the Header and in each Alignment. func AddOrReplaceReadGroup(readGroup utils.StringMap) sam.Filter { return func(header *sam.Header) sam.AlignmentFilter { header.RG = []utils.StringMap{readGroup} id := readGroup["ID"] return func(aln *sam.Alignment) bool { aln.SetRG(id); return true } } } // AddPGLine returns a filter for adding a @PG tag to a Header, and // ensuring that it is the first one in the chain. func AddPGLine(newPG utils.StringMap) sam.Filter
// RenameChromosomes is a filter for prepending "chr" to the reference // sequence names in a Header, and in RNAME and RNEXT in each // Alignment. func RenameChromosomes(header *sam.Header) sam.AlignmentFilter { for _, entry := range header.SQ { if sn, found := entry["SN"]; found { entry["SN"] = "chr" + sn } } return func(aln *sam.Alignment) bool { if (aln.RNAME != "=") && (aln.RNAME != "*") { aln.RNAME = "chr" + aln.RNAME } if (aln.RNEXT != "=") && (aln.RNEXT != "*") { aln.RNEXT = "chr" + aln.RNEXT } return true } } // AddREFID is a filter for adding the refid (index in the reference // sequence dictionary) to alignments as temporary values. func AddREFID(header *sam.Header) sam.AlignmentFilter { dictTable := make(map[string]int32) for index, entry := range header.SQ { dictTable[entry["SN"]] = int32(index) } return func(aln *sam.Alignment) bool { value, found := dictTable[aln.RNAME] if !found { value = -1 } aln.SetREFID(value) return true } } // RemoveOptionalFields returns a filter for removing optional fields // in an alignment. func RemoveOptionalFields(tags []string) sam.Filter { if len(tags) == 0 { return nil } // Intern the tags once. var optionals []utils.Symbol for _, tag := range tags { optionals = append(optionals, utils.Intern(tag)) } return func(header *sam.Header) sam.AlignmentFilter { return func(aln *sam.Alignment) bool { aln.TAGS, _ = aln.TAGS.DeleteIf(func(key utils.Symbol, val interface{}) bool { for _, tag := range optionals { if tag == key { return true } } return false }) return true } } } // KeepOptionalFields returns a filter for removing all but a list of // given optional fields in an alignment. func KeepOptionalFields(tags []string) sam.Filter { if len(tags) == 0 { return func(header *sam.Header) sam.AlignmentFilter { return func(aln *sam.Alignment) bool { aln.TAGS = nil return true } } } // Intern the tags once. var optionals []utils.Symbol for _, tag := range tags { optionals = append(optionals, utils.Intern(tag)) } return func(header *sam.Header) sam.AlignmentFilter { return func(aln *sam.Alignment) bool { aln.TAGS, _ = aln.TAGS.DeleteIf(func(key utils.Symbol, val interface{}) bool { for _, tag := range optionals { if tag == key { return false } } return true }) return true } } } // CleanSam is a filter for soft-clipping an alignment at the end of a // reference sequence, and set MAPQ to 0 if unmapped. func CleanSam(header *sam.Header) sam.AlignmentFilter { referenceSequenceTable := make(map[string]int32) for _, sn := range header.SQ { referenceSequenceTable[sn["SN"]], _ = sam.SQLN(sn) } return
{ return func(header *sam.Header) sam.AlignmentFilter { id := newPG["ID"] for utils.Find(header.PG, func(entry utils.StringMap) bool { return entry["ID"] == id }) >= 0 { id += " " id += strconv.FormatInt(rand.Int63n(0x10000), 16) } newPG["ID"] = id for _, PG := range header.PG { nextID := PG["ID"] if pos := utils.Find(header.PG, func(entry utils.StringMap) bool { return entry["PP"] == nextID }); pos < 0 { newPG["PP"] = nextID break } } header.PG = append(header.PG, newPG) return nil } }
identifier_body
simple-filters.go
(aln *sam.Alignment) bool { return dictTable[aln.RNAME] } } } // ReplaceReferenceSequenceDictionaryFromSamFile returns a filter for // replacing the reference sequence dictionary in a Header with one // parsed from the given SAM/DICT file. func ReplaceReferenceSequenceDictionaryFromSamFile(samFile string) (f sam.Filter, err error) { input, err := sam.Open(samFile, true) if err != nil {
nerr := input.Close() if err == nil { err = nerr } }() header, _, err := sam.ParseHeader(input.Reader) if err != nil { return nil, err } return ReplaceReferenceSequenceDictionary(header.SQ), nil } // RemoveUnmappedReads is a filter for removing unmapped sam-alignment // instances, based on FLAG. func RemoveUnmappedReads(_ *sam.Header) sam.AlignmentFilter { return func(aln *sam.Alignment) bool { return (aln.FLAG & sam.Unmapped) == 0 } } // RemoveUnmappedReadsStrict is a filter for removing unmapped // sam-alignment instances, based on FLAG, or POS=0, or RNAME=*. func RemoveUnmappedReadsStrict(_ *sam.Header) sam.AlignmentFilter { return func(aln *sam.Alignment) bool { return ((aln.FLAG & sam.Unmapped) == 0) && (aln.POS != 0) && (aln.RNAME != "*") } } // RemoveNonExactMappingReads is a filter that removes all reads that // are not exact matches with the reference (soft-clipping ok), based // on CIGAR string (only M and S allowed). func RemoveNonExactMappingReads(_ *sam.Header) sam.AlignmentFilter { return func(aln *sam.Alignment) bool { return !strings.ContainsAny(aln.CIGAR, "IDNHPX=") } } // Symbols for optional fields used for determining exact matches. See // http://samtools.github.io/hts-specs/SAMv1.pdf - Section 1.5. var ( X0 = utils.Intern("X0") X1 = utils.Intern("X1") XM = utils.Intern("XM") XO = utils.Intern("XO") XG = utils.Intern("XG") ) // RemoveNonExactMappingReadsStrict is a filter that removes all reads // that are not exact matches with the reference, based on the // optional fields X0=1 (unique mapping), X1=0 (no suboptimal hit), // XM=0 (no mismatch), XO=0 (no gap opening), XG=0 (no gap extension). func RemoveNonExactMappingReadsStrict(header *sam.Header) sam.AlignmentFilter { return func(aln *sam.Alignment) bool { if x0, ok := aln.TAGS.Get(X0); !ok || x0.(int32) != 1 { return false } if x1, ok := aln.TAGS.Get(X1); !ok || x1.(int32) != 0 { return false } if xm, ok := aln.TAGS.Get(XM); !ok || xm.(int32) != 0 { return false } if xo, ok := aln.TAGS.Get(XO); !ok || xo.(int32) != 0 { return false } if xg, ok := aln.TAGS.Get(XG); !ok || xg.(int32) != 0 { return false } return true } } // RemoveDuplicateReads is a filter for removing duplicate // sam-alignment instances, based on FLAG. func RemoveDuplicateReads(_ *sam.Header) sam.AlignmentFilter { return func(aln *sam.Alignment) bool { return (aln.FLAG & sam.Duplicate) == 0 } } var sr = utils.Intern("sr") // RemoveOptionalReads is a filter for removing alignments that // represent optional information in elPrep. func RemoveOptionalReads(header *sam.Header) sam.AlignmentFilter { if _, found := header.UserRecords["@sr"]; found { delete(header.UserRecords, "@sr") return func(aln *sam.Alignment) bool { _, found := aln.TAGS.Get(sr); return !found } } return nil } // AddOrReplaceReadGroup returns a filter for adding or replacing the // read group both in the Header and in each Alignment. func AddOrReplaceReadGroup(readGroup utils.StringMap) sam.Filter { return func(header *sam.Header) sam.AlignmentFilter { header.RG = []utils.StringMap{readGroup} id := readGroup["ID"] return func(aln *sam.Alignment) bool { aln.SetRG(id); return true } } } // AddPGLine returns a filter for adding a @PG tag to a Header, and // ensuring that it is the first one in the chain. func AddPGLine(newPG utils.StringMap) sam.Filter { return func(header *sam.Header) sam.AlignmentFilter { id := newPG["ID"] for utils.Find(header.PG, func(entry utils.StringMap) bool { return entry["ID"] == id }) >= 0 { id += " " id += strconv.FormatInt(rand.Int63n(0x10000), 16) } newPG["ID"] = id for _, PG := range header.PG { nextID := PG["ID"] if pos := utils.Find(header.PG, func(entry utils.StringMap) bool { return entry["PP"] == nextID }); pos < 0 { newPG["PP"] = nextID break } } header.PG = append(header.PG, newPG) return nil } } // RenameChromosomes is a filter for prepending "chr" to the reference // sequence names in a Header, and in RNAME and RNEXT in each // Alignment. func RenameChromosomes(header *sam.Header) sam.AlignmentFilter { for _, entry := range header.SQ { if sn, found := entry["SN"]; found { entry["SN"] = "chr" + sn } } return func(aln *sam.Alignment) bool { if (aln.RNAME != "=") && (aln.RNAME != "*") { aln.RNAME = "chr" + aln.RNAME } if (aln.RNEXT != "=") && (aln.RNEXT != "*") { aln.RNEXT = "chr" + aln.RNEXT } return true } } // AddREFID is a filter for adding the refid (index in the reference // sequence dictionary) to alignments as temporary values. func AddREFID(header *sam.Header) sam.AlignmentFilter { dictTable := make(map[string]int32) for index, entry := range header.SQ { dictTable[entry["SN"]] = int32(index) } return func(aln *sam.Alignment) bool { value, found := dictTable[aln.RNAME] if !found { value = -1 } aln.SetREFID(value) return true } } // RemoveOptionalFields returns a filter for removing optional fields // in an alignment. func RemoveOptionalFields(tags []string) sam.Filter { if len(tags) == 0 { return nil } // Intern the tags once. var optionals []utils.Symbol for _, tag := range tags { optionals = append(optionals, utils.Intern(tag)) } return func(header *sam.Header) sam.AlignmentFilter { return func(aln *sam.Alignment) bool { aln.TAGS, _ = aln.TAGS.DeleteIf(func(key utils.Symbol, val interface{}) bool { for _, tag := range optionals { if tag == key { return true } } return false }) return true } } } // KeepOptionalFields returns a filter for removing all but a list of // given optional fields in an alignment. func KeepOptionalFields(tags []string) sam.Filter { if len(tags) == 0 { return func(header *sam.Header) sam.AlignmentFilter { return func(aln *sam.Alignment) bool { aln.TAGS = nil return true } } } // Intern the tags once. var optionals []utils.Symbol for _, tag := range tags { optionals = append(optionals, utils.Intern(tag)) } return func(header *sam.Header) sam.AlignmentFilter { return func(aln *sam.Alignment) bool { aln.TAGS, _ = aln.TAGS.DeleteIf(func(key utils.Symbol, val interface{}) bool { for _, tag := range optionals { if tag == key { return false } } return true }) return true } } } // CleanSam is a filter for soft-clipping an alignment at the end of a // reference sequence, and set MAPQ to 0 if unmapped. func CleanSam(header *sam.Header) sam.AlignmentFilter { referenceSequenceTable := make(map[string]int32) for _, sn := range header.SQ { referenceSequenceTable[sn["SN"]], _ = sam.SQLN(sn) } return func
return nil, err } defer func() {
random_line_split
scout.py
100 * res.score) label = '{}% {}'.format(percent, labels[res[0]]) cv2_im = cv2.putText(cv2_im, label, (600, 20+ii*30), cv2.FONT_HERSHEY_PLAIN, 2, (255, 0, 255), 1) return cv2_im # define callbacks def human_callback(witch, arg_path, arg_name, arg_recfile, data_sens): # choosw from witch label = "HUMAN" timetag = time.strftime("%Y%m%d_%H%M%S") # log to record file record_file = open(arg_recfile, 'a+') if (witch==1): record_file.write("[scout.detection]: <{0}> <{1}>\n".format(timetag, arg_name[0])) elif (witch==2): record_file.write("[scout.detection]: <{0}> <{1}>\n".format(timetag, arg_name[1])) record_file.write(">> [label]: {}\n".format(label)); record_file.write('>> [sensor.name]:{0}\n'.format(arg_name[1])) record_file.write('>> [mean_temperature]: {0:.2f} C\n'.format(data_sens[3])) record_file.write('>> [data]: \n') record_file.write(data_sens[1]) record_file.write('>> [sensor.name]:{0}\n'.format(arg_name[0])) record_file.write('>> [mean_temperature]: {0:.2f} C\n'.format(data_sens[2])) record_file.write('>> [data]: \n') record_file.write(data_sens[0]) # cmd exck out_filename = '' if (witch==1): out_filename = arg_path + timetag +"_"+label+"_"+ arg_name[0] +".mkv" #cmd = "ffmpeg -i /dev/video6 -vcodec h264_omx -b:v 2M -t 15 " + out_filename cmd = "ffmpeg -i /dev/video6 -t 15 -vcodec copy " + out_filename elif(witch==2): out_filename = arg_path + timetag +"_"+label+"_"+ arg_name[1] +".mkv" #cmd = "ffmpeg -i /dev/video2 -vcodec h264_omx -b:v 2M -t 15 " + out_filename cmd = "ffmpeg -i /dev/video2 -t 15 -vcodec copy " + out_filename else: pass list_cmd = cmd.split(' ') # actualiza y cierra registro record_file.write('>> [video.capture]:{0}\n\n'.format(out_filename)) record_file.close() # ejecuta cmd_out = subprocess.run(list_cmd, stdout=subprocess.PIPE) # print(cmd_out.stdout.decode('utf-8')) return cmd_out.stdout.decode('utf-8') def label_callback(label, witch, arg_path, arg_name, arg_recfile, data_sens): # choosw from witch #label = "HUMAN" timetag = time.strftime("%Y%m%d_%H%M%S") # log to record file record_file = open(arg_recfile, 'a+') if (witch==1): record_file.write("[scout.detection]: <{0}> <{1}>\n".format(timetag, arg_name[0])) elif (witch==2): record_file.write("[scout.detection]: <{0}> <{1}>\n".format(timetag, arg_name[1])) record_file.write(">> [label]: {}\n".format(label)); record_file.write('>> [sensor.name]:[{0}]\n'.format(arg_name[1])) record_file.write('>> [mean_temperature]: {0:.2f} C\n'.format(data_sens[3])) record_file.write('>> [data]: \n') record_file.write(data_sens[1]) record_file.write('>> [sensor.name]:[{0}]\n'.format(arg_name[0])) record_file.write('>> [mean_temperature]: {0:.2f} C\n'.format(data_sens[2])) record_file.write('>> [data]: \n') record_file.write(data_sens[0]) # cmd exck out_filename = '' if (witch==1): out_filename = arg_path + timetag +"_"+label+"_"+ arg_name[0] +".mkv" #cmd = "ffmpeg -i /dev/video6 -vcodec h264_omx -b:v 2M -t 15 " + out_filename cmd = "ffmpeg -i /dev/video6 -t 15 -vcodec copy " + out_filename elif(witch==2): out_filename = arg_path + timetag +"_"+label+"_"+ arg_name[1] +".mkv" #cmd = "ffmpeg -i /dev/video2 -vcodec h264_omx -b:v 2M -t 15 " + out_filename cmd = "ffmpeg -i /dev/video2 -t 15 -vcodec copy " + out_filename else: pass list_cmd = cmd.split(' ') # actualiza y cierra registro record_file.write('>> [video.capture]:{0}\n\n\n'.format(out_filename)) record_file.close() # ejecuta cmd_out = subprocess.run(list_cmd, stdout=subprocess.PIPE) # print(cmd_out.stdout.decode('utf-8')) return cmd_out.stdout.decode('utf-8') # soundsys def update_soundsystem(arg_recfile, arg_name, osc_c): """ envía mensajes a sc que disparan notas aleatorias en los rangos establecidos registra las notas en el archivo de log """ # generate note and send osc message note_val = random.randint(0,6) synthnames = ['A','B', 'C', 'D', 'E', 'F', 'G'] ruta = '/scout/note/'+arg_name+'/' + synthnames[note_val] ruta = ruta.encode() osc_c.send_message(ruta, [1]) # log to record file timetag = time.strftime("%Y%m%d_%H%M%S") record_file = open(arg_recfile, 'a+') record_file.write("\n[scout.note]: <{0}> {1}\n".format(timetag, ruta.decode())) record_file.close() return # -main def main(): # -parse arguments parser = argparse.ArgumentParser() parser.add_argument('--path', help='Path of capture folder', default="/media/pi/DATA/capture/video/") parser.add_argument('--recfile', help='Path of capture folder', default="/media/pi/DATA/capture/record/") parser.add_argument('--name', help='Name of the directions to scout [NE || SW]', default="NE") parser.add_argument('--verbose', help='Show additional info for debugging', default=False) parser.add_argument('--show', help='Show video', default=False) parser.add_argument('--ip', help='OSC ip', default="192.168.1.207") parser.add_argument('--port', help='OSC port', default="57120") args = parser.parse_args() # -init osc client osc_addr = args.ip osc_port = int(args.port) osc_client = OSCClient(osc_addr, osc_port) # -load model and labels for detection default_model_dir = '/home/pi/Dev/animals/train' default_model = 'animals_duo_model.tflite' default_labels = 'animals_duo_model.txt' args_model = os.path.join(default_model_dir, default_model) args_labels = os.path.join(default_model_dir, default_labels) args_top_k = 1 args_camera_idx = 0 args_threshold = 0.1 os.makedirs(args.path, exist_ok=True) os.makedirs(args.recfile, exist_ok=True) # -create the detection interpreter print('Cargando {} con {} categorias de objetos.'.format(args_model, args_labels)) interpreter = cvtf.make_interpreter(args_model) interpreter.allocate_tensors() labels = load_labels(args_labels) # -record file timetag = time.strftime("%Y%m%d_%H%M%S") arg_recfile = args.recfile + timetag + ".log" record_file = open(arg_recfile, 'w+') record_file.write("[scout.record.start]:\t----\t----\t-- <{0}>: \n".format(timetag)); record_file.close() # -create a capture object and connect to cam cam = None witch = 0 empty = create_blank(640, 480, rgb_color=(0,0,0)) buffstream = '' # -the loop (hole) t0 = time.time()
t2 = time.time() nc_a, nc_b, data_sens = dual_detect(args.name, args.verbose) while True:
random_line_split
scout.py
) m_tb, arr_tb = read_sensor_pixels(sensor_b) na = len(list(filter(lambda x: (x - m_ta) >= MIN_TEMP_DIFF, arr_ta))) nb = len(list(filter(lambda x: (x - m_tb) >= MIN_TEMP_DIFF, arr_tb))) if verbose: print("\n") print ('[t1]:{0:.1f}\t[t2]:{1:.1f}'.format(m_tb, m_ta)) for ix in range(8): la = ''.join(['.' if (arr_ta[iy * 8 + ix] - m_ta) < MIN_TEMP_DIFF else '+' for iy in range(8)]) lb = ''.join(['.' if (arr_tb[iy * 8 + ix] - m_tb) < MIN_TEMP_DIFF else '+' for iy in range(8)]) print(lb,'\t',la) print ('[o1]:{0:d}\t\t[o2]:{1:d}'.format(nb, na)) print("\n") return na, nb def dual_detect(arg_name, verbose=False): """ Llama a read_sensor_pixels una vez por cada sensor Devuelve el número de celdas ocupadas en cada sensor (mas los data_sens para log) Con verbose muestra paneles de detección """ m_ta, arr_ta = read_sensor_pixels(sensor_a) m_tb, arr_tb = read_sensor_pixels(sensor_b) na = len(list(filter(lambda x: (x - m_ta) >= MIN_TEMP_DIFF, arr_ta))) nb = len(list(filter(lambda x: (x - m_tb) >= MIN_TEMP_DIFF, arr_tb))) if verbose: print("\n") print ('[t{2}]:{0:.1f}\t[t{3}]:{1:.1f}'.format(m_tb, m_ta, arg_name[1], arg_name[0])) sens_a = "" sens_b = "" for ix in range(8): la = ''.join(['.' if (arr_ta[iy * 8 + ix] - m_ta) < MIN_TEMP_DIFF else '+' for iy in range(8)]) lb = ''.join(['.' if (arr_tb[iy * 8 + ix] - m_tb) < MIN_TEMP_DIFF else '+' for iy in range(8)]) sens_a+=la+'\n' sens_b+=lb+'\n' if verbose: print(lb,'\t',la) if verbose: print ('[o{2}]:{0:d}\t\t[o{3}]:{1:d}'.format(nb, na, arg_name[1], arg_name[0])) print("\n") return na, nb, [sens_a, sens_b, m_ta, m_tb] # detection functions def load_labels(path): p = re.compile(r'\s*(\d+)(.+)') with open(path, 'r', encoding='utf-8') as f: lines = (p.match(line).groups() for line in f.readlines()) return {int(num): text.strip() for num, text in lines} def get_output(interpreter, top_k, score_threshold): """Returns no more than top_k categories with score >= score_threshold.""" scores = cvtf.output_tensor(interpreter, 0) categories = [ Category(i, scores[i]) for i in np.argpartition(scores, -top_k)[-top_k:] if scores[i] >= score_threshold ] return sorted(categories, key=operator.itemgetter(1), reverse=True) def append_results_to_img(cv2_im, results, labels): height, width, channels = cv2_im.shape for ii, res in enumerate(results): percent = int(100 * res.score) label = '{}% {}'.format(percent, labels[res[0]]) cv2_im = cv2.putText(cv2_im, label, (600, 20+ii*30), cv2.FONT_HERSHEY_PLAIN, 2, (255, 0, 255), 1) return cv2_im def parse_results(cv2_im, results, labels): height, width, channels = cv2_im.shape for ii, res in enumerate(results): percent = int(100 * res.score) label = '{}% {}'.format(percent, labels[res[0]]) cv2_im = cv2.putText(cv2_im, label, (600, 20+ii*30), cv2.FONT_HERSHEY_PLAIN, 2, (255, 0, 255), 1) return cv2_im # define callbacks def human_ca
arg_path, arg_name, arg_recfile, data_sens): # choosw from witch label = "HUMAN" timetag = time.strftime("%Y%m%d_%H%M%S") # log to record file record_file = open(arg_recfile, 'a+') if (witch==1): record_file.write("[scout.detection]: <{0}> <{1}>\n".format(timetag, arg_name[0])) elif (witch==2): record_file.write("[scout.detection]: <{0}> <{1}>\n".format(timetag, arg_name[1])) record_file.write(">> [label]: {}\n".format(label)); record_file.write('>> [sensor.name]:{0}\n'.format(arg_name[1])) record_file.write('>> [mean_temperature]: {0:.2f} C\n'.format(data_sens[3])) record_file.write('>> [data]: \n') record_file.write(data_sens[1]) record_file.write('>> [sensor.name]:{0}\n'.format(arg_name[0])) record_file.write('>> [mean_temperature]: {0:.2f} C\n'.format(data_sens[2])) record_file.write('>> [data]: \n') record_file.write(data_sens[0]) # cmd exck out_filename = '' if (witch==1): out_filename = arg_path + timetag +"_"+label+"_"+ arg_name[0] +".mkv" #cmd = "ffmpeg -i /dev/video6 -vcodec h264_omx -b:v 2M -t 15 " + out_filename cmd = "ffmpeg -i /dev/video6 -t 15 -vcodec copy " + out_filename elif(witch==2): out_filename = arg_path + timetag +"_"+label+"_"+ arg_name[1] +".mkv" #cmd = "ffmpeg -i /dev/video2 -vcodec h264_omx -b:v 2M -t 15 " + out_filename cmd = "ffmpeg -i /dev/video2 -t 15 -vcodec copy " + out_filename else: pass list_cmd = cmd.split(' ') # actualiza y cierra registro record_file.write('>> [video.capture]:{0}\n\n'.format(out_filename)) record_file.close() # ejecuta cmd_out = subprocess.run(list_cmd, stdout=subprocess.PIPE) # print(cmd_out.stdout.decode('utf-8')) return cmd_out.stdout.decode('utf-8') def label_callback(label, witch, arg_path, arg_name, arg_recfile, data_sens): # choosw from witch #label = "HUMAN" timetag = time.strftime("%Y%m%d_%H%M%S") # log to record file record_file = open(arg_recfile, 'a+') if (witch==1): record_file.write("[scout.detection]: <{0}> <{1}>\n".format(timetag, arg_name[0])) elif (witch==2): record_file.write("[scout.detection]: <{0}> <{1}>\n".format(timetag, arg_name[1])) record_file.write(">> [label]: {}\n".format(label)); record_file.write('>> [sensor.name]:[{0}]\n'.format(arg_name[1])) record_file.write('>> [mean_temperature]: {0:.2f} C\n'.format(data_sens[3])) record_file.write('>> [data]: \n') record_file.write(data_sens[1]) record_file.write('>> [sensor.name]:[{0}]\n'.format(arg_name[0])) record_file.write('>> [mean_temperature]: {0:.2f} C\n'.format(data_sens[2])) record_file.write('>> [data]: \n') record_file.write(data_sens[0]) # cmd exck out_filename = '' if (witch==1): out_filename = arg_path + timetag +"_"+label+"_"+ arg_name[0] +".mkv" #cmd = "ffmpeg -i /dev/video6 -vcodec h264_omx -b:v 2M -t 15 " + out_filename cmd = "ffmpeg -i /dev/video6 -t 15 -vcodec copy " + out_filename elif(witch==2): out_filename = arg_path + timetag +"_"+label+"_"+ arg_name[1] +".mkv" #cmd = "ffmpeg -i /dev/video2 -vcodec h2
llback(witch,
identifier_name
scout.py
) m_tb, arr_tb = read_sensor_pixels(sensor_b) na = len(list(filter(lambda x: (x - m_ta) >= MIN_TEMP_DIFF, arr_ta))) nb = len(list(filter(lambda x: (x - m_tb) >= MIN_TEMP_DIFF, arr_tb))) if verbose: print(
eturn na, nb def dual_detect(arg_name, verbose=False): """ Llama a read_sensor_pixels una vez por cada sensor Devuelve el número de celdas ocupadas en cada sensor (mas los data_sens para log) Con verbose muestra paneles de detección """ m_ta, arr_ta = read_sensor_pixels(sensor_a) m_tb, arr_tb = read_sensor_pixels(sensor_b) na = len(list(filter(lambda x: (x - m_ta) >= MIN_TEMP_DIFF, arr_ta))) nb = len(list(filter(lambda x: (x - m_tb) >= MIN_TEMP_DIFF, arr_tb))) if verbose: print("\n") print ('[t{2}]:{0:.1f}\t[t{3}]:{1:.1f}'.format(m_tb, m_ta, arg_name[1], arg_name[0])) sens_a = "" sens_b = "" for ix in range(8): la = ''.join(['.' if (arr_ta[iy * 8 + ix] - m_ta) < MIN_TEMP_DIFF else '+' for iy in range(8)]) lb = ''.join(['.' if (arr_tb[iy * 8 + ix] - m_tb) < MIN_TEMP_DIFF else '+' for iy in range(8)]) sens_a+=la+'\n' sens_b+=lb+'\n' if verbose: print(lb,'\t',la) if verbose: print ('[o{2}]:{0:d}\t\t[o{3}]:{1:d}'.format(nb, na, arg_name[1], arg_name[0])) print("\n") return na, nb, [sens_a, sens_b, m_ta, m_tb] # detection functions def load_labels(path): p = re.compile(r'\s*(\d+)(.+)') with open(path, 'r', encoding='utf-8') as f: lines = (p.match(line).groups() for line in f.readlines()) return {int(num): text.strip() for num, text in lines} def get_output(interpreter, top_k, score_threshold): """Returns no more than top_k categories with score >= score_threshold.""" scores = cvtf.output_tensor(interpreter, 0) categories = [ Category(i, scores[i]) for i in np.argpartition(scores, -top_k)[-top_k:] if scores[i] >= score_threshold ] return sorted(categories, key=operator.itemgetter(1), reverse=True) def append_results_to_img(cv2_im, results, labels): height, width, channels = cv2_im.shape for ii, res in enumerate(results): percent = int(100 * res.score) label = '{}% {}'.format(percent, labels[res[0]]) cv2_im = cv2.putText(cv2_im, label, (600, 20+ii*30), cv2.FONT_HERSHEY_PLAIN, 2, (255, 0, 255), 1) return cv2_im def parse_results(cv2_im, results, labels): height, width, channels = cv2_im.shape for ii, res in enumerate(results): percent = int(100 * res.score) label = '{}% {}'.format(percent, labels[res[0]]) cv2_im = cv2.putText(cv2_im, label, (600, 20+ii*30), cv2.FONT_HERSHEY_PLAIN, 2, (255, 0, 255), 1) return cv2_im # define callbacks def human_callback(witch, arg_path, arg_name, arg_recfile, data_sens): # choosw from witch label = "HUMAN" timetag = time.strftime("%Y%m%d_%H%M%S") # log to record file record_file = open(arg_recfile, 'a+') if (witch==1): record_file.write("[scout.detection]: <{0}> <{1}>\n".format(timetag, arg_name[0])) elif (witch==2): record_file.write("[scout.detection]: <{0}> <{1}>\n".format(timetag, arg_name[1])) record_file.write(">> [label]: {}\n".format(label)); record_file.write('>> [sensor.name]:{0}\n'.format(arg_name[1])) record_file.write('>> [mean_temperature]: {0:.2f} C\n'.format(data_sens[3])) record_file.write('>> [data]: \n') record_file.write(data_sens[1]) record_file.write('>> [sensor.name]:{0}\n'.format(arg_name[0])) record_file.write('>> [mean_temperature]: {0:.2f} C\n'.format(data_sens[2])) record_file.write('>> [data]: \n') record_file.write(data_sens[0]) # cmd exck out_filename = '' if (witch==1): out_filename = arg_path + timetag +"_"+label+"_"+ arg_name[0] +".mkv" #cmd = "ffmpeg -i /dev/video6 -vcodec h264_omx -b:v 2M -t 15 " + out_filename cmd = "ffmpeg -i /dev/video6 -t 15 -vcodec copy " + out_filename elif(witch==2): out_filename = arg_path + timetag +"_"+label+"_"+ arg_name[1] +".mkv" #cmd = "ffmpeg -i /dev/video2 -vcodec h264_omx -b:v 2M -t 15 " + out_filename cmd = "ffmpeg -i /dev/video2 -t 15 -vcodec copy " + out_filename else: pass list_cmd = cmd.split(' ') # actualiza y cierra registro record_file.write('>> [video.capture]:{0}\n\n'.format(out_filename)) record_file.close() # ejecuta cmd_out = subprocess.run(list_cmd, stdout=subprocess.PIPE) # print(cmd_out.stdout.decode('utf-8')) return cmd_out.stdout.decode('utf-8') def label_callback(label, witch, arg_path, arg_name, arg_recfile, data_sens): # choosw from witch #label = "HUMAN" timetag = time.strftime("%Y%m%d_%H%M%S") # log to record file record_file = open(arg_recfile, 'a+') if (witch==1): record_file.write("[scout.detection]: <{0}> <{1}>\n".format(timetag, arg_name[0])) elif (witch==2): record_file.write("[scout.detection]: <{0}> <{1}>\n".format(timetag, arg_name[1])) record_file.write(">> [label]: {}\n".format(label)); record_file.write('>> [sensor.name]:[{0}]\n'.format(arg_name[1])) record_file.write('>> [mean_temperature]: {0:.2f} C\n'.format(data_sens[3])) record_file.write('>> [data]: \n') record_file.write(data_sens[1]) record_file.write('>> [sensor.name]:[{0}]\n'.format(arg_name[0])) record_file.write('>> [mean_temperature]: {0:.2f} C\n'.format(data_sens[2])) record_file.write('>> [data]: \n') record_file.write(data_sens[0]) # cmd exck out_filename = '' if (witch==1): out_filename = arg_path + timetag +"_"+label+"_"+ arg_name[0] +".mkv" #cmd = "ffmpeg -i /dev/video6 -vcodec h264_omx -b:v 2M -t 15 " + out_filename cmd = "ffmpeg -i /dev/video6 -t 15 -vcodec copy " + out_filename elif(witch==2): out_filename = arg_path + timetag +"_"+label+"_"+ arg_name[1] +".mkv" #cmd = "ffmpeg -i /dev/video2 -vcodec h26
"\n") print ('[t1]:{0:.1f}\t[t2]:{1:.1f}'.format(m_tb, m_ta)) for ix in range(8): la = ''.join(['.' if (arr_ta[iy * 8 + ix] - m_ta) < MIN_TEMP_DIFF else '+' for iy in range(8)]) lb = ''.join(['.' if (arr_tb[iy * 8 + ix] - m_tb) < MIN_TEMP_DIFF else '+' for iy in range(8)]) print(lb,'\t',la) print ('[o1]:{0:d}\t\t[o2]:{1:d}'.format(nb, na)) print("\n") r
conditional_block
scout.py
) m_tb, arr_tb = read_sensor_pixels(sensor_b) na = len(list(filter(lambda x: (x - m_ta) >= MIN_TEMP_DIFF, arr_ta))) nb = len(list(filter(lambda x: (x - m_tb) >= MIN_TEMP_DIFF, arr_tb))) if verbose: print("\n") print ('[t1]:{0:.1f}\t[t2]:{1:.1f}'.format(m_tb, m_ta)) for ix in range(8): la = ''.join(['.' if (arr_ta[iy * 8 + ix] - m_ta) < MIN_TEMP_DIFF else '+' for iy in range(8)]) lb = ''.join(['.' if (arr_tb[iy * 8 + ix] - m_tb) < MIN_TEMP_DIFF else '+' for iy in range(8)]) print(lb,'\t',la) print ('[o1]:{0:d}\t\t[o2]:{1:d}'.format(nb, na)) print("\n") return na, nb def dual_detect(arg_name, verbose=False): """ Llama a read_sensor_pixels una vez por cada sensor Devuelve el número de celdas ocupadas en cada sensor (mas los data_sens para log) Con verbose muestra paneles de detección """ m_ta, arr_ta = read_sensor_pixels(sensor_a) m_tb, arr_tb = read_sensor_pixels(sensor_b) na = len(list(filter(lambda x: (x - m_ta) >= MIN_TEMP_DIFF, arr_ta))) nb = len(list(filter(lambda x: (x - m_tb) >= MIN_TEMP_DIFF, arr_tb))) if verbose: print("\n") print ('[t{2}]:{0:.1f}\t[t{3}]:{1:.1f}'.format(m_tb, m_ta, arg_name[1], arg_name[0])) sens_a = "" sens_b = "" for ix in range(8): la = ''.join(['.' if (arr_ta[iy * 8 + ix] - m_ta) < MIN_TEMP_DIFF else '+' for iy in range(8)]) lb = ''.join(['.' if (arr_tb[iy * 8 + ix] - m_tb) < MIN_TEMP_DIFF else '+' for iy in range(8)]) sens_a+=la+'\n' sens_b+=lb+'\n' if verbose: print(lb,'\t',la) if verbose: print ('[o{2}]:{0:d}\t\t[o{3}]:{1:d}'.format(nb, na, arg_name[1], arg_name[0])) print("\n") return na, nb, [sens_a, sens_b, m_ta, m_tb] # detection functions def load_labels(path): p = re.compile(r'\s*(\d+)(.+)') with open(path, 'r', encoding='utf-8') as f: lines = (p.match(line).groups() for line in f.readlines()) return {int(num): text.strip() for num, text in lines} def get_output(interpreter, top_k, score_threshold): """Retur
pend_results_to_img(cv2_im, results, labels): height, width, channels = cv2_im.shape for ii, res in enumerate(results): percent = int(100 * res.score) label = '{}% {}'.format(percent, labels[res[0]]) cv2_im = cv2.putText(cv2_im, label, (600, 20+ii*30), cv2.FONT_HERSHEY_PLAIN, 2, (255, 0, 255), 1) return cv2_im def parse_results(cv2_im, results, labels): height, width, channels = cv2_im.shape for ii, res in enumerate(results): percent = int(100 * res.score) label = '{}% {}'.format(percent, labels[res[0]]) cv2_im = cv2.putText(cv2_im, label, (600, 20+ii*30), cv2.FONT_HERSHEY_PLAIN, 2, (255, 0, 255), 1) return cv2_im # define callbacks def human_callback(witch, arg_path, arg_name, arg_recfile, data_sens): # choosw from witch label = "HUMAN" timetag = time.strftime("%Y%m%d_%H%M%S") # log to record file record_file = open(arg_recfile, 'a+') if (witch==1): record_file.write("[scout.detection]: <{0}> <{1}>\n".format(timetag, arg_name[0])) elif (witch==2): record_file.write("[scout.detection]: <{0}> <{1}>\n".format(timetag, arg_name[1])) record_file.write(">> [label]: {}\n".format(label)); record_file.write('>> [sensor.name]:{0}\n'.format(arg_name[1])) record_file.write('>> [mean_temperature]: {0:.2f} C\n'.format(data_sens[3])) record_file.write('>> [data]: \n') record_file.write(data_sens[1]) record_file.write('>> [sensor.name]:{0}\n'.format(arg_name[0])) record_file.write('>> [mean_temperature]: {0:.2f} C\n'.format(data_sens[2])) record_file.write('>> [data]: \n') record_file.write(data_sens[0]) # cmd exck out_filename = '' if (witch==1): out_filename = arg_path + timetag +"_"+label+"_"+ arg_name[0] +".mkv" #cmd = "ffmpeg -i /dev/video6 -vcodec h264_omx -b:v 2M -t 15 " + out_filename cmd = "ffmpeg -i /dev/video6 -t 15 -vcodec copy " + out_filename elif(witch==2): out_filename = arg_path + timetag +"_"+label+"_"+ arg_name[1] +".mkv" #cmd = "ffmpeg -i /dev/video2 -vcodec h264_omx -b:v 2M -t 15 " + out_filename cmd = "ffmpeg -i /dev/video2 -t 15 -vcodec copy " + out_filename else: pass list_cmd = cmd.split(' ') # actualiza y cierra registro record_file.write('>> [video.capture]:{0}\n\n'.format(out_filename)) record_file.close() # ejecuta cmd_out = subprocess.run(list_cmd, stdout=subprocess.PIPE) # print(cmd_out.stdout.decode('utf-8')) return cmd_out.stdout.decode('utf-8') def label_callback(label, witch, arg_path, arg_name, arg_recfile, data_sens): # choosw from witch #label = "HUMAN" timetag = time.strftime("%Y%m%d_%H%M%S") # log to record file record_file = open(arg_recfile, 'a+') if (witch==1): record_file.write("[scout.detection]: <{0}> <{1}>\n".format(timetag, arg_name[0])) elif (witch==2): record_file.write("[scout.detection]: <{0}> <{1}>\n".format(timetag, arg_name[1])) record_file.write(">> [label]: {}\n".format(label)); record_file.write('>> [sensor.name]:[{0}]\n'.format(arg_name[1])) record_file.write('>> [mean_temperature]: {0:.2f} C\n'.format(data_sens[3])) record_file.write('>> [data]: \n') record_file.write(data_sens[1]) record_file.write('>> [sensor.name]:[{0}]\n'.format(arg_name[0])) record_file.write('>> [mean_temperature]: {0:.2f} C\n'.format(data_sens[2])) record_file.write('>> [data]: \n') record_file.write(data_sens[0]) # cmd exck out_filename = '' if (witch==1): out_filename = arg_path + timetag +"_"+label+"_"+ arg_name[0] +".mkv" #cmd = "ffmpeg -i /dev/video6 -vcodec h264_omx -b:v 2M -t 15 " + out_filename cmd = "ffmpeg -i /dev/video6 -t 15 -vcodec copy " + out_filename elif(witch==2): out_filename = arg_path + timetag +"_"+label+"_"+ arg_name[1] +".mkv" #cmd = "ffmpeg -i /dev/video2 -vcodec h2
ns no more than top_k categories with score >= score_threshold.""" scores = cvtf.output_tensor(interpreter, 0) categories = [ Category(i, scores[i]) for i in np.argpartition(scores, -top_k)[-top_k:] if scores[i] >= score_threshold ] return sorted(categories, key=operator.itemgetter(1), reverse=True) def ap
identifier_body
mla_1c_v0.py
0.010741 #item_price 0.067416 -0.228345 1.000000 0.022186 #month_cnt 0.521578 -0.010741 0.022186 1.000000 # Transform it in a links data frame (3 columns only): links = C.stack().reset_index() links.columns =["var1","var2","corr_val"] # remove self correlation (cor(A,A)=1) links_filtered=links.loc[ (links['var1'] != links['var2']) ] links_filtered # Build your graph #G = nx.Graph() G = nx.path_graph(0) graph = {"freq":["price","count"],"price":["freq","count"], "count":["price","freq"]} leng=1 #[('freq', 'price'), ('freq', 'count'), ('price', 'count')] values=[0.067,0.522,0.022] for vertex, edges in graph.items(): G.add_node("%s" % vertex) # leng+=1 for edge in edges: G.add_node("%s" % edge) G.add_edge("%s" % vertex, "%s" % edge, weight = leng) # print("'%s' connects with '%s'" % (vertex,edge)) # Create positions of all nodes and save them #pos = nx.spring_layout(G) pos={"price": [1.5,1.5],"freq": [0.5,1.5],"count": [1,1]} labels ={('freq', 'price'): values[0], ('freq', 'count'): values[1], ('price', 'count'): values[2]} # Draw the graph according to node positions nx.draw(G, pos, with_labels=True,node_size=3000) # Create edge labels #labels = {edg: str(values[G.edges[edg]]) for edg in G.edges} # Draw edge labels according to node positions pos_lab={"price": [1.25,1.25],"freq": [0.75,1.25],"count": [1,1.5]} nx.draw_networkx_edge_labels(G, pos,font_color='red',edge_labels=labels) plt.axis('off') plt.show() ################ #import statsmodels.formula.api as smf # Instantiate a gamma family model with the default link function. #poisson_model = sm.GLM(data.endog, data.exog, family=sm.families.Gamma()) #form="month_cnt ~ date_block_num + item_id + shop_id + item_freq + category_id + month + item_price" #form="month_cnt ~ date_block_num + item_freq + month + item_price" #poisson_model = smf.glm(formula=form, data=dataTrain, family=sm.families.Poisson()) #poisson_fit = poisson_model.fit() #dir(poisson_fit.mle_settings) #poisson_fit.use_t #print(poisson_fit.summary()) # # Generalized Linear Model Regression Results #============================================================================== #Dep. Variable: month_cnt No. Observations: 921400 #Model: GLM Df Residuals: 921392 #Model Family: Poisson Df Model: 7 #Link Function: log Scale: 1.0000 #Method: IRLS Log-Likelihood:* -inf #Date: Fri, 15 Nov 2019 Deviance: * 8.7344e+05 #Time: 18:15:41 Pearson chi2: 3.83e+06 #No. Iterations: 7 *non-defined for Poisson family #Covariance Type: nonrobust * non defined for scale=1 #================================================================================== # coef std err z P>|z| [0.025 0.975] #---------------------------------------------------------------------------------- #Intercept 0.5517 0.003 163.637 0.000 0.545 0.558 #date_block_num 0.0013 0.000 10.540 0.000 0.001 0.002 #item_id -9.174e-06 1.23e-07 -74.511 0.000 -9.41e-06 -8.93e-06 #shop_id -0.0012 4.26e-05 -27.026 0.000 -0.001 -0.001 #item_freq 0.1936 8.63e-05 2244.772 0.000 0.193 0.194 #category_id -0.0055 4.5e-05 -123.243 0.000 -0.006 -0.005 #month 0.0017 0.000 7.667 0.000 0.001 0.002 #item_price 1.289e-05 3.19e-07 40.347 0.000 1.23e-05 1.35e-05 #================================================================================== # item_id, category_id have small weight # Generalized Linear Model Regression Results #============================================================================== #Dep. Variable: month_cnt No. Observations: 921400 #Model: GLM Df Residuals: 921395 #Model Family: Poisson Df Model: 4 #Link Function: log Scale: 1.0000 #Method: IRLS Log-Likelihood: -inf #Date: Fri, 15 Nov 2019 Deviance: 9.1019e+05 #Time: 18:40:30 Pearson chi2: 3.78e+06 #No. Iterations: 7 #Covariance Type: nonrobust #================================================================================== # coef std err z P>|z| [0.025 0.975] #---------------------------------------------------------------------------------- #Intercept 0.2137 0.003 81.395 0.000 0.209 0.219 #date_block_num 0.0004 0.000 3.000 0.003 0.000 0.001 #item_freq 0.1881 8.02e-05 2346.055 0.000 0.188 0.188 #month 0.0024 0.000 11.216 0.000 0.002 0.003 #item_price 2.899e-05 2.82e-07 102.951 0.000 2.84e-05 2.95e-05 #================================================================================== # item_freq is obviously the larger coeff filePath="working_data/"+"1C_ctrl_training.csv" data2=pd.read_csv(filePath, index_col=False) data2=data2.drop(["Unnamed: 0",'Unnamed: 0.1', 'Unnamed: 0.1.1'],axis=1) data2.keys() data2.head() dataCtrl=ct.my_prepareTrain(data2) dataCtrl.keys() dataCtrlHM=ct.my_summaryHistoricFunc(dataCtrl,f_mean=True,f_sum=False) #takes almost 10 minutes #dataCtrl=pd.get_dummies(dataCtrl) dataCtrl.reset_index() dataCtrlHM.reset_index()
C=pd.merge(dataCtrl,dataCtrlHM,how="left",on=["date_block_num","item_id","shop_id"]) #target=dataCtrl["month_cnt"] #dataCtrl=dataCtrl.drop("month_cnt",axis=1) #predictions=poisson_fit.predict(exog=dataCtrl, transform=True)
random_line_split
mla_1c_v0.py
73820 0.067416 0.521578 #category_id -0.073820 1.000000 -0.228345 -0.010741 #item_price 0.067416 -0.228345 1.000000 0.022186 #month_cnt 0.521578 -0.010741 0.022186 1.000000 # Transform it in a links data frame (3 columns only): links = C.stack().reset_index() links.columns =["var1","var2","corr_val"] # remove self correlation (cor(A,A)=1) links_filtered=links.loc[ (links['var1'] != links['var2']) ] links_filtered # Build your graph #G = nx.Graph() G = nx.path_graph(0) graph = {"freq":["price","count"],"price":["freq","count"], "count":["price","freq"]} leng=1 #[('freq', 'price'), ('freq', 'count'), ('price', 'count')] values=[0.067,0.522,0.022] for vertex, edges in graph.items():
# print("'%s' connects with '%s'" % (vertex,edge)) # Create positions of all nodes and save them #pos = nx.spring_layout(G) pos={"price": [1.5,1.5],"freq": [0.5,1.5],"count": [1,1]} labels ={('freq', 'price'): values[0], ('freq', 'count'): values[1], ('price', 'count'): values[2]} # Draw the graph according to node positions nx.draw(G, pos, with_labels=True,node_size=3000) # Create edge labels #labels = {edg: str(values[G.edges[edg]]) for edg in G.edges} # Draw edge labels according to node positions pos_lab={"price": [1.25,1.25],"freq": [0.75,1.25],"count": [1,1.5]} nx.draw_networkx_edge_labels(G, pos,font_color='red',edge_labels=labels) plt.axis('off') plt.show() ################ #import statsmodels.formula.api as smf # Instantiate a gamma family model with the default link function. #poisson_model = sm.GLM(data.endog, data.exog, family=sm.families.Gamma()) #form="month_cnt ~ date_block_num + item_id + shop_id + item_freq + category_id + month + item_price" #form="month_cnt ~ date_block_num + item_freq + month + item_price" #poisson_model = smf.glm(formula=form, data=dataTrain, family=sm.families.Poisson()) #poisson_fit = poisson_model.fit() #dir(poisson_fit.mle_settings) #poisson_fit.use_t #print(poisson_fit.summary()) # # Generalized Linear Model Regression Results #============================================================================== #Dep. Variable: month_cnt No. Observations: 921400 #Model: GLM Df Residuals: 921392 #Model Family: Poisson Df Model: 7 #Link Function: log Scale: 1.0000 #Method: IRLS Log-Likelihood:* -inf #Date: Fri, 15 Nov 2019 Deviance: * 8.7344e+05 #Time: 18:15:41 Pearson chi2: 3.83e+06 #No. Iterations: 7 *non-defined for Poisson family #Covariance Type: nonrobust * non defined for scale=1 #================================================================================== # coef std err z P>|z| [0.025 0.975] #---------------------------------------------------------------------------------- #Intercept 0.5517 0.003 163.637 0.000 0.545 0.558 #date_block_num 0.0013 0.000 10.540 0.000 0.001 0.002 #item_id -9.174e-06 1.23e-07 -74.511 0.000 -9.41e-06 -8.93e-06 #shop_id -0.0012 4.26e-05 -27.026 0.000 -0.001 -0.001 #item_freq 0.1936 8.63e-05 2244.772 0.000 0.193 0.194 #category_id -0.0055 4.5e-05 -123.243 0.000 -0.006 -0.005 #month 0.0017 0.000 7.667 0.000 0.001 0.002 #item_price 1.289e-05 3.19e-07 40.347 0.000 1.23e-05 1.35e-05 #================================================================================== # item_id, category_id have small weight # Generalized Linear Model Regression Results #============================================================================== #Dep. Variable: month_cnt No. Observations: 921400 #Model: GLM Df Residuals: 921395 #Model Family: Poisson Df Model: 4 #Link Function: log Scale: 1.0000 #Method: IRLS Log-Likelihood: -inf #Date: Fri, 15 Nov 2019 Deviance: 9.1019e+05 #Time: 18:40:30 Pearson chi2: 3.78e+06 #No. Iterations: 7 #Covariance Type: nonrobust #================================================================================== # coef std err z P>|z| [0.025 0.975] #---------------------------------------------------------------------------------- #Intercept 0.2137 0.003 81.395 0.000 0.209 0.219 #date_block_num 0.0004 0.000 3.000 0.003 0.000 0.001 #item_freq 0.1881 8.02e-05 2346.055 0.000 0.188 0.188 #month 0.0024 0.000 11.216 0.000 0.002 0.003 #item_price 2.899e-05 2.82e-07 102.951 0.000 2.84e-05 2.95e-05 #================================================================================== # item_freq is obviously the larger coeff filePath="working_data/"+"1C_ctrl_training.csv" data2=pd.read_csv(filePath, index_col=False) data2=data2.drop(["Unnamed: 0",'Unnamed: 0.1', 'Unnamed: 0.1.1'],axis=1) data2.keys() data2.head() dataCtrl=ct.my_prepareTrain(data2) dataCtrl.keys() dataCtrlHM=ct.my_summaryHistoricFunc(dataCtrl,f_mean=True,f_sum=False) #takes almost 10 minutes #dataCtrl=pd.get_dummies(dataCtrl) dataCtrl.reset_index() dataCtrlHM.reset_index()
G.add_node("%s" % vertex) # leng+=1 for edge in edges: G.add_node("%s" % edge) G.add_edge("%s" % vertex, "%s" % edge, weight = leng)
conditional_block
lib.rs
pub struct RequestHeader { pub method: Method, pub uri: Uri, pub version: Version, pub fields: HeaderMap, } #[derive(Debug)] pub struct ResponseHeader { pub status_code: StatusCode, pub fields: HeaderMap, } #[derive(Debug)] pub enum InvalidRequestHeader { Format, RequestLine(InvalidRequestLine), HeaderField(InvalidHeaderField), Io(io::Error), } impl From<InvalidRequestLine> for InvalidRequestHeader { fn from(e: InvalidRequestLine) -> Self { InvalidRequestHeader::RequestLine(e) } } impl From<InvalidHeaderField> for InvalidRequestHeader { fn from(e: InvalidHeaderField) -> Self { InvalidRequestHeader::HeaderField(e) } } impl From<io::Error> for InvalidRequestHeader { fn from(e: io::Error) -> Self { InvalidRequestHeader::Io(e) } } const LINE_CAP: usize = 16384; pub fn parse_request_header<B: BufRead>(mut stream: B) -> Result<RequestHeader, InvalidRequestHeader> { // TODO: Why does removing the type from `line` here cause errors? let next_line = |stream: &mut B, line: &mut Vec<u8>| { line.clear(); let count = stream .take(LINE_CAP as u64) .read_until('\n' as u8, line)?; match count { 0 => Err(InvalidRequestHeader::Format), // FIXME? LINE_CAP => Err(InvalidRequestHeader::Format), // FIXME _ => Ok(()), } }; let mut line = Vec::with_capacity(LINE_CAP); next_line(&mut stream, &mut line)?; if !line.ends_with(b"\r\n") { return Err(InvalidRequestHeader::Format); } line.truncate(line.len() - 2); let (method, uri, version) = parse_request_line(&line[..])?; let mut fields = HeaderMap::new(); loop { next_line(&mut stream, &mut line)?; if !line.ends_with(b"\r\n") { return Err(InvalidRequestHeader::Format); } line.truncate(line.len() - 2); if line == b"" { return Ok(RequestHeader { method, uri, version, fields }); } let (name, value) = parse_header_field(&line)?; // TODO: append is okay, right? No syntax issues because we haven't // seralized anything yet. fields.append(name, value); // TODO: we should care about result, right? } } pub fn write_response_header<W: Write>(header: &ResponseHeader, stream: W) -> io::Result<()> { let mut stream = BufWriter::new(stream); // TODO: Is this the way you're supposed to format bytes? stream.write_all(b"HTTP/1.1")?; stream.write_all(b" ")?; stream.write_all(header.status_code.as_str().as_bytes())?; stream.write_all(b" ")?; stream.write_all( header .status_code .canonical_reason() .unwrap_or("Unknown Reason") .as_bytes() )?; stream.write_all(b"\r\n")?; for key in header.fields.keys() { let mut values = header.fields.get_all(key).into_iter().peekable(); stream.write_all(key.as_str().as_bytes())?; stream.write_all(b": ")?; match values.next() { Some(v) => stream.write_all(v.as_bytes())?, None => panic!("what?"), } if values.peek().is_some() { let separate_fields = key == "set-cookie"; for v in values { if separate_fields { stream.write_all(b"\r\n")?; stream.write_all(key.as_str().as_bytes())?; stream.write_all(b": ")?; } else { stream.write_all(b",")?; } stream.write_all(v.as_bytes())?; } } stream.write_all(b"\r\n")?; } stream.write_all(b"\r\n")?; Ok(()) } #[derive(Debug)] pub enum InvalidRequestLine { Format, Method(InvalidMethod), Uri(InvalidUriBytes), Version, } impl From<InvalidMethod> for InvalidRequestLine { fn from(e: InvalidMethod) -> Self { InvalidRequestLine::Method(e) } } impl From<InvalidUriBytes> for InvalidRequestLine { fn from(e: InvalidUriBytes) -> Self { InvalidRequestLine::Uri(e) } } pub fn parse_request_line(s: &[u8]) -> Result<(Method, Uri, Version), InvalidRequestLine> { lazy_static! { static ref R: Regex = Regex::new( // method SP request-target SP HTTP-version r"(?-u)^(\S+) (\S+) (\S+)$" ).unwrap(); } let cap = R.captures(s).ok_or(InvalidRequestLine::Format)?; Ok(( Method::from_bytes(&cap[1])?, Uri::from_shared(cap[2].into())?, match &cap[3] { // rfc 7230 section A: "Any server that implements name-based // virtual hosts ought to disable support for HTTP/0.9." b"HTTP/1.0" => Version::HTTP_10, b"HTTP/1.1" => Version::HTTP_11, // We don't support HTTP 0.9 or 2.0. 2.0 support may be added later. // FIXME: Can we respond to an invalid version with 505 HTTP // Version Not Supported? If not, unsupported major versions need a // different error than invalid versions. // FIXME: We should probably accept requests with version 1.2 and // higher. Check the spec. _ => return Err(InvalidRequestLine::Version), }, )) } #[derive(Debug)] pub enum InvalidHeaderField { Format, Name(InvalidHeaderName), Value(InvalidHeaderValue), } impl From<InvalidHeaderName> for InvalidHeaderField { fn from(e: InvalidHeaderName) -> Self { InvalidHeaderField::Name(e) } } impl From<InvalidHeaderValue> for InvalidHeaderField { fn from(e: InvalidHeaderValue) -> Self { InvalidHeaderField::Value(e) } } pub fn parse_header_field(s: &[u8]) -> Result<(HeaderName, HeaderValue), InvalidHeaderField> { // TODO: support obs-fold e.g. within message/http // (see rfc7230 section 3.2.4) // rfc7230 section 3.2.4: Server MUST return 400 if there's whitespace // between field name and colon. // rfc7230 section 3.2.4: If obs-fold is used outside a message/http body, // server MUST either return 400 or replace each such obs-fold with one or // more SP chars. lazy_static! { static ref R: Regex = Regex::new(&(String::new() // token ":" OWS *field-content OWS + r"(?-u)^(" + TOKEN + "):" + r"[\t ]*" + r"([!-~\x80-\xFF]([\t !-~\x80-\xFF]*[!-~\x80-\xFF])?)" + r"[\t ]*$" )).unwrap(); } let cap = R.captures(s).ok_or(InvalidHeaderField::Format)?; Ok(( HeaderName::from_bytes(&cap[1])?, // TODO: HeaderValue might not fully validate input. HeaderValue::from_bytes(&cap[2])?, )) } #[cfg(test)] mod test { use crate::{ parse_request_header, parse_request_line, parse_header_field, ResponseHeader, write_response_header, }; use http::header::{ HeaderMap, HeaderValue, }; use http::{ Method, StatusCode, Version, }; #[test] fn test_parse_request_header() { let mut s = Vec::new(); // TODO: There's a better way to do this, right? s.extend( &b"POST http://foo.example.com/bar?qux=19&qux=xyz HTTP/1.1\r\n"[..] ); s.extend(&b"Host: foo.example.com\r\n"[..]); s.extend(&b"Content-Type: application/json\r\n"[..]); s.extend(&b"\r\n"[..]); let h = parse_request_header(&s[..]).unwrap(); assert_eq!(h.method, Method::POST); assert_eq!(h.uri.scheme_str().unwrap(), "http"); assert_eq!(h.uri.host().unwrap(), "foo.example.com"); assert_eq!(h.uri.port_part(), None); assert_eq!(h.uri.path(), "/bar"); assert_eq!(h.uri.query().unwrap(), "qux=19&qux=xyz"); assert_eq!(h.version, Version::HTTP_11); assert_eq!(h.fields["host"], "foo.example.com"); assert_eq!(h.fields["content-type"], "application/json"); }
random_line_split
lib.rs
InvalidRequestHeader::Format); } line.truncate(line.len() - 2); if line == b"" { return Ok(RequestHeader { method, uri, version, fields }); } let (name, value) = parse_header_field(&line)?; // TODO: append is okay, right? No syntax issues because we haven't // seralized anything yet. fields.append(name, value); // TODO: we should care about result, right? } } pub fn write_response_header<W: Write>(header: &ResponseHeader, stream: W) -> io::Result<()> { let mut stream = BufWriter::new(stream); // TODO: Is this the way you're supposed to format bytes? stream.write_all(b"HTTP/1.1")?; stream.write_all(b" ")?; stream.write_all(header.status_code.as_str().as_bytes())?; stream.write_all(b" ")?; stream.write_all( header .status_code .canonical_reason() .unwrap_or("Unknown Reason") .as_bytes() )?; stream.write_all(b"\r\n")?; for key in header.fields.keys() { let mut values = header.fields.get_all(key).into_iter().peekable(); stream.write_all(key.as_str().as_bytes())?; stream.write_all(b": ")?; match values.next() { Some(v) => stream.write_all(v.as_bytes())?, None => panic!("what?"), } if values.peek().is_some() { let separate_fields = key == "set-cookie"; for v in values { if separate_fields { stream.write_all(b"\r\n")?; stream.write_all(key.as_str().as_bytes())?; stream.write_all(b": ")?; } else { stream.write_all(b",")?; } stream.write_all(v.as_bytes())?; } } stream.write_all(b"\r\n")?; } stream.write_all(b"\r\n")?; Ok(()) } #[derive(Debug)] pub enum InvalidRequestLine { Format, Method(InvalidMethod), Uri(InvalidUriBytes), Version, } impl From<InvalidMethod> for InvalidRequestLine { fn from(e: InvalidMethod) -> Self { InvalidRequestLine::Method(e) } } impl From<InvalidUriBytes> for InvalidRequestLine { fn from(e: InvalidUriBytes) -> Self { InvalidRequestLine::Uri(e) } } pub fn parse_request_line(s: &[u8]) -> Result<(Method, Uri, Version), InvalidRequestLine> { lazy_static! { static ref R: Regex = Regex::new( // method SP request-target SP HTTP-version r"(?-u)^(\S+) (\S+) (\S+)$" ).unwrap(); } let cap = R.captures(s).ok_or(InvalidRequestLine::Format)?; Ok(( Method::from_bytes(&cap[1])?, Uri::from_shared(cap[2].into())?, match &cap[3] { // rfc 7230 section A: "Any server that implements name-based // virtual hosts ought to disable support for HTTP/0.9." b"HTTP/1.0" => Version::HTTP_10, b"HTTP/1.1" => Version::HTTP_11, // We don't support HTTP 0.9 or 2.0. 2.0 support may be added later. // FIXME: Can we respond to an invalid version with 505 HTTP // Version Not Supported? If not, unsupported major versions need a // different error than invalid versions. // FIXME: We should probably accept requests with version 1.2 and // higher. Check the spec. _ => return Err(InvalidRequestLine::Version), }, )) } #[derive(Debug)] pub enum InvalidHeaderField { Format, Name(InvalidHeaderName), Value(InvalidHeaderValue), } impl From<InvalidHeaderName> for InvalidHeaderField { fn from(e: InvalidHeaderName) -> Self { InvalidHeaderField::Name(e) } } impl From<InvalidHeaderValue> for InvalidHeaderField { fn from(e: InvalidHeaderValue) -> Self { InvalidHeaderField::Value(e) } } pub fn parse_header_field(s: &[u8]) -> Result<(HeaderName, HeaderValue), InvalidHeaderField> { // TODO: support obs-fold e.g. within message/http // (see rfc7230 section 3.2.4) // rfc7230 section 3.2.4: Server MUST return 400 if there's whitespace // between field name and colon. // rfc7230 section 3.2.4: If obs-fold is used outside a message/http body, // server MUST either return 400 or replace each such obs-fold with one or // more SP chars. lazy_static! { static ref R: Regex = Regex::new(&(String::new() // token ":" OWS *field-content OWS + r"(?-u)^(" + TOKEN + "):" + r"[\t ]*" + r"([!-~\x80-\xFF]([\t !-~\x80-\xFF]*[!-~\x80-\xFF])?)" + r"[\t ]*$" )).unwrap(); } let cap = R.captures(s).ok_or(InvalidHeaderField::Format)?; Ok(( HeaderName::from_bytes(&cap[1])?, // TODO: HeaderValue might not fully validate input. HeaderValue::from_bytes(&cap[2])?, )) } #[cfg(test)] mod test { use crate::{ parse_request_header, parse_request_line, parse_header_field, ResponseHeader, write_response_header, }; use http::header::{ HeaderMap, HeaderValue, }; use http::{ Method, StatusCode, Version, }; #[test] fn test_parse_request_header() { let mut s = Vec::new(); // TODO: There's a better way to do this, right? s.extend( &b"POST http://foo.example.com/bar?qux=19&qux=xyz HTTP/1.1\r\n"[..] ); s.extend(&b"Host: foo.example.com\r\n"[..]); s.extend(&b"Content-Type: application/json\r\n"[..]); s.extend(&b"\r\n"[..]); let h = parse_request_header(&s[..]).unwrap(); assert_eq!(h.method, Method::POST); assert_eq!(h.uri.scheme_str().unwrap(), "http"); assert_eq!(h.uri.host().unwrap(), "foo.example.com"); assert_eq!(h.uri.port_part(), None); assert_eq!(h.uri.path(), "/bar"); assert_eq!(h.uri.query().unwrap(), "qux=19&qux=xyz"); assert_eq!(h.version, Version::HTTP_11); assert_eq!(h.fields["host"], "foo.example.com"); assert_eq!(h.fields["content-type"], "application/json"); } #[test] fn test_write_response_header() { let mut s = Vec::new(); let mut h = ResponseHeader { status_code: StatusCode::from_u16(404).unwrap(), fields: HeaderMap::new(), }; write_response_header(&h, &mut s).unwrap(); assert_eq!(s, b"HTTP/1.1 404 Not Found\r\n\r\n"); h.fields.append("set-cookie", HeaderValue::from_static( "FOO=\"some text\"" )); h.fields.append("Set-cookie", HeaderValue::from_static( "BAR=\"some other text\"" )); h.fields.append("LOCATION", HeaderValue::from_static( "http://example.com:3180/foo&bar" )); h.fields.append("Content-Language", HeaderValue::from_static( "en" )); h.fields.append("Content-Language", HeaderValue::from_static( "de" )); s.clear(); write_response_header(&h, &mut s).unwrap(); assert!(s.starts_with(b"HTTP/1.1 404 Not Found\r\n")); } #[test] fn test_parse_request_line()
{ let s = b"OPTIONS * HTTP/1.1"; let (m, u, v) = parse_request_line(s).unwrap(); assert_eq!(m, Method::OPTIONS); assert_eq!(u.path(), "*"); assert_eq!(v, Version::HTTP_11); let s = b"POST http://foo.example.com/bar?qux=19&qux=xyz HTTP/1.0"; let (m, u, v) = parse_request_line(s).unwrap(); assert_eq!(m, Method::POST); assert_eq!(u.scheme_str().unwrap(), "http"); assert_eq!(u.host().unwrap(), "foo.example.com"); assert_eq!(u.port_part(), None); assert_eq!(u.path(), "/bar"); assert_eq!(u.query().unwrap(), "qux=19&qux=xyz"); assert_eq!(v, Version::HTTP_10); }
identifier_body
lib.rs
pub fields: HeaderMap, } #[derive(Debug)] pub enum InvalidRequestHeader { Format, RequestLine(InvalidRequestLine), HeaderField(InvalidHeaderField), Io(io::Error), } impl From<InvalidRequestLine> for InvalidRequestHeader { fn from(e: InvalidRequestLine) -> Self { InvalidRequestHeader::RequestLine(e) } } impl From<InvalidHeaderField> for InvalidRequestHeader { fn
(e: InvalidHeaderField) -> Self { InvalidRequestHeader::HeaderField(e) } } impl From<io::Error> for InvalidRequestHeader { fn from(e: io::Error) -> Self { InvalidRequestHeader::Io(e) } } const LINE_CAP: usize = 16384; pub fn parse_request_header<B: BufRead>(mut stream: B) -> Result<RequestHeader, InvalidRequestHeader> { // TODO: Why does removing the type from `line` here cause errors? let next_line = |stream: &mut B, line: &mut Vec<u8>| { line.clear(); let count = stream .take(LINE_CAP as u64) .read_until('\n' as u8, line)?; match count { 0 => Err(InvalidRequestHeader::Format), // FIXME? LINE_CAP => Err(InvalidRequestHeader::Format), // FIXME _ => Ok(()), } }; let mut line = Vec::with_capacity(LINE_CAP); next_line(&mut stream, &mut line)?; if !line.ends_with(b"\r\n") { return Err(InvalidRequestHeader::Format); } line.truncate(line.len() - 2); let (method, uri, version) = parse_request_line(&line[..])?; let mut fields = HeaderMap::new(); loop { next_line(&mut stream, &mut line)?; if !line.ends_with(b"\r\n") { return Err(InvalidRequestHeader::Format); } line.truncate(line.len() - 2); if line == b"" { return Ok(RequestHeader { method, uri, version, fields }); } let (name, value) = parse_header_field(&line)?; // TODO: append is okay, right? No syntax issues because we haven't // seralized anything yet. fields.append(name, value); // TODO: we should care about result, right? } } pub fn write_response_header<W: Write>(header: &ResponseHeader, stream: W) -> io::Result<()> { let mut stream = BufWriter::new(stream); // TODO: Is this the way you're supposed to format bytes? stream.write_all(b"HTTP/1.1")?; stream.write_all(b" ")?; stream.write_all(header.status_code.as_str().as_bytes())?; stream.write_all(b" ")?; stream.write_all( header .status_code .canonical_reason() .unwrap_or("Unknown Reason") .as_bytes() )?; stream.write_all(b"\r\n")?; for key in header.fields.keys() { let mut values = header.fields.get_all(key).into_iter().peekable(); stream.write_all(key.as_str().as_bytes())?; stream.write_all(b": ")?; match values.next() { Some(v) => stream.write_all(v.as_bytes())?, None => panic!("what?"), } if values.peek().is_some() { let separate_fields = key == "set-cookie"; for v in values { if separate_fields { stream.write_all(b"\r\n")?; stream.write_all(key.as_str().as_bytes())?; stream.write_all(b": ")?; } else { stream.write_all(b",")?; } stream.write_all(v.as_bytes())?; } } stream.write_all(b"\r\n")?; } stream.write_all(b"\r\n")?; Ok(()) } #[derive(Debug)] pub enum InvalidRequestLine { Format, Method(InvalidMethod), Uri(InvalidUriBytes), Version, } impl From<InvalidMethod> for InvalidRequestLine { fn from(e: InvalidMethod) -> Self { InvalidRequestLine::Method(e) } } impl From<InvalidUriBytes> for InvalidRequestLine { fn from(e: InvalidUriBytes) -> Self { InvalidRequestLine::Uri(e) } } pub fn parse_request_line(s: &[u8]) -> Result<(Method, Uri, Version), InvalidRequestLine> { lazy_static! { static ref R: Regex = Regex::new( // method SP request-target SP HTTP-version r"(?-u)^(\S+) (\S+) (\S+)$" ).unwrap(); } let cap = R.captures(s).ok_or(InvalidRequestLine::Format)?; Ok(( Method::from_bytes(&cap[1])?, Uri::from_shared(cap[2].into())?, match &cap[3] { // rfc 7230 section A: "Any server that implements name-based // virtual hosts ought to disable support for HTTP/0.9." b"HTTP/1.0" => Version::HTTP_10, b"HTTP/1.1" => Version::HTTP_11, // We don't support HTTP 0.9 or 2.0. 2.0 support may be added later. // FIXME: Can we respond to an invalid version with 505 HTTP // Version Not Supported? If not, unsupported major versions need a // different error than invalid versions. // FIXME: We should probably accept requests with version 1.2 and // higher. Check the spec. _ => return Err(InvalidRequestLine::Version), }, )) } #[derive(Debug)] pub enum InvalidHeaderField { Format, Name(InvalidHeaderName), Value(InvalidHeaderValue), } impl From<InvalidHeaderName> for InvalidHeaderField { fn from(e: InvalidHeaderName) -> Self { InvalidHeaderField::Name(e) } } impl From<InvalidHeaderValue> for InvalidHeaderField { fn from(e: InvalidHeaderValue) -> Self { InvalidHeaderField::Value(e) } } pub fn parse_header_field(s: &[u8]) -> Result<(HeaderName, HeaderValue), InvalidHeaderField> { // TODO: support obs-fold e.g. within message/http // (see rfc7230 section 3.2.4) // rfc7230 section 3.2.4: Server MUST return 400 if there's whitespace // between field name and colon. // rfc7230 section 3.2.4: If obs-fold is used outside a message/http body, // server MUST either return 400 or replace each such obs-fold with one or // more SP chars. lazy_static! { static ref R: Regex = Regex::new(&(String::new() // token ":" OWS *field-content OWS + r"(?-u)^(" + TOKEN + "):" + r"[\t ]*" + r"([!-~\x80-\xFF]([\t !-~\x80-\xFF]*[!-~\x80-\xFF])?)" + r"[\t ]*$" )).unwrap(); } let cap = R.captures(s).ok_or(InvalidHeaderField::Format)?; Ok(( HeaderName::from_bytes(&cap[1])?, // TODO: HeaderValue might not fully validate input. HeaderValue::from_bytes(&cap[2])?, )) } #[cfg(test)] mod test { use crate::{ parse_request_header, parse_request_line, parse_header_field, ResponseHeader, write_response_header, }; use http::header::{ HeaderMap, HeaderValue, }; use http::{ Method, StatusCode, Version, }; #[test] fn test_parse_request_header() { let mut s = Vec::new(); // TODO: There's a better way to do this, right? s.extend( &b"POST http://foo.example.com/bar?qux=19&qux=xyz HTTP/1.1\r\n"[..] ); s.extend(&b"Host: foo.example.com\r\n"[..]); s.extend(&b"Content-Type: application/json\r\n"[..]); s.extend(&b"\r\n"[..]); let h = parse_request_header(&s[..]).unwrap(); assert_eq!(h.method, Method::POST); assert_eq!(h.uri.scheme_str().unwrap(), "http"); assert_eq!(h.uri.host().unwrap(), "foo.example.com"); assert_eq!(h.uri.port_part(), None); assert_eq!(h.uri.path(), "/bar"); assert_eq!(h.uri.query().unwrap(), "qux=19&qux=xyz"); assert_eq!(h.version, Version::HTTP_11); assert_eq!(h.fields["host"], "foo.example.com"); assert_eq!(h.fields["content-type"], "application/json"); } #[test] fn test_write_response_header() { let mut s = Vec::new(); let mut h = ResponseHeader { status_code: StatusCode::from_u16(404).unwrap(),
from
identifier_name
api_export_tools.py
's xform :return True or False """ return bool(set(hxl_columns).intersection(set(dv_columns))) def _get_export_type(export_type): if export_type not in EXPORT_EXT or ( export_type == Export.GOOGLE_SHEETS_EXPORT and not getattr(settings, "GOOGLE_EXPORT", False) ): raise exceptions.ParseError( _(f"'{export_type}' format not known or not implemented!") ) return EXPORT_EXT[export_type] # pylint: disable=too-many-arguments, too-many-locals, too-many-branches def custom_response_handler( # noqa: C0901 request, xform, query, export_type, token=None, meta=None, dataview=False, filename=None, metadata=None, ): """ Returns a HTTP response with export file for download. """ export_type = _get_export_type(export_type) if ( export_type in EXTERNAL_EXPORT_TYPES and (token is not None) or (meta is not None) ): export_type = Export.EXTERNAL_EXPORT options = parse_request_export_options(request.query_params) dataview_pk = hasattr(dataview, "pk") and dataview.pk options["dataview_pk"] = dataview_pk options["host"] = request.get_host() if dataview: columns_with_hxl = get_columns_with_hxl(xform.survey.get("children")) if columns_with_hxl: options["include_hxl"] = include_hxl_row( dataview.columns, list(columns_with_hxl) ) try: query = filter_queryset_xform_meta_perms_sql(xform, request.user, query) except NoRecordsPermission: return Response( data=json.dumps({"details": _("You don't have permission")}), status=status.HTTP_403_FORBIDDEN, content_type="application/json", ) if query: options["query"] = query remove_group_name = options.get("remove_group_name") export_id = request.query_params.get("export_id") if export_id: export = get_object_or_404(Export, id=export_id, xform=xform) else: if export_type == Export.GOOGLE_SHEETS_EXPORT: return Response( data=json.dumps( {"details": _("Sheets export only supported in async mode")} ), status=status.HTTP_403_FORBIDDEN, content_type="application/json", ) # check if we need to re-generate, # we always re-generate if a filter is specified def _new_export(): return _generate_new_export( request, xform, query, export_type, dataview_pk=dataview_pk, metadata=metadata, ) if should_create_new_export(xform, export_type, options, request=request): export = _new_export() else: export = newest_export_for(xform, export_type, options) if not export.filename and not export.error_message: export = _new_export() log_export(request, xform, export_type) if export_type == Export.EXTERNAL_EXPORT: return external_export_response(export) if export.filename is None and export.error_message:
# get extension from file_path, exporter could modify to # xlsx if it exceeds limits _path, ext = os.path.splitext(export.filename) ext = ext[1:] show_date = True if filename is None and export.status == Export.SUCCESSFUL: filename = _generate_filename( request, xform, remove_group_name, dataview_pk=dataview_pk ) else: show_date = False response = response_with_mimetype_and_name( Export.EXPORT_MIMES[ext], filename, extension=ext, show_date=show_date, file_path=export.filepath, ) return response def _generate_new_export( # noqa: C0901 request, xform, query, export_type, dataview_pk=False, metadata=None ): query = _set_start_end_params(request, query) extension = _get_extension_from_export_type(export_type) options = { "extension": extension, "username": xform.user.username, "id_string": xform.id_string, "host": request.get_host(), "sort": request.query_params.get('sort') } if query: options["query"] = query options["dataview_pk"] = dataview_pk if export_type == Export.GOOGLE_SHEETS_EXPORT: options["google_credentials"] = _get_google_credential(request).to_json() try: if export_type == Export.EXTERNAL_EXPORT: options["token"] = request.GET.get("token") options["data_id"] = request.GET.get("data_id") options["meta"] = request.GET.get("meta") export = generate_external_export( export_type, xform.user.username, xform.id_string, None, options, xform=xform, ) elif export_type == Export.OSM_EXPORT: export = generate_osm_export( export_type, xform.user.username, xform.id_string, None, options, xform=xform, ) elif export_type == Export.ZIP_EXPORT: export = generate_attachments_zip_export( export_type, xform.user.username, xform.id_string, None, options, xform=xform, ) elif export_type == Export.KML_EXPORT: export = generate_kml_export( export_type, xform.user.username, xform.id_string, None, options, xform=xform, ) elif export_type == Export.GEOJSON_EXPORT: export = generate_geojson_export( export_type, xform.user.username, xform.id_string, metadata, None, options, xform=xform, ) else: options.update(parse_request_export_options(request.query_params)) export = generate_export(export_type, xform, None, options) audit = {"xform": xform.id_string, "export_type": export_type} log.audit_log( log.Actions.EXPORT_CREATED, request.user, xform.user, _("Created %(export_type)s export on '%(id_string)s'.") % {"id_string": xform.id_string, "export_type": export_type.upper()}, audit, request, ) except NoRecordsFoundError as e: raise Http404(_("No records found to export")) from e except J2XException as e: # j2x exception return async_status(FAILED, str(e)) except SPSSIOError as e: raise exceptions.ParseError(str(e)) from e else: return export def log_export(request, xform, export_type): """ Logs audit logs of export requests. """ # log download as well audit = {"xform": xform.id_string, "export_type": export_type} log.audit_log( log.Actions.EXPORT_DOWNLOADED, request.user, xform.user, _(f"Downloaded {export_type.upper()} export on '{xform.id_string}'."), audit, request, ) def external_export_response(export): """ Redirects to export_url of XLSReports successful export. In case of a failure, returns a 400 HTTP JSON response with the error message. """ if isinstance(export, Export) and export.internal_status == Export.SUCCESSFUL: return HttpResponseRedirect(export.export_url) http_status = status.HTTP_400_BAD_REQUEST return Response(json.dumps(export), http_status, content_type="application/json") def _generate_filename(request, xform, remove_group_name=False, dataview_pk=False): if request.GET.get("raw"): filename = None else: # append group name removed flag otherwise use the form id_string if remove_group_name: filename = f"{xform.id_string}-{GROUPNAME_REMOVED_FLAG}" elif dataview_pk: filename = f"{xform.id_string}-{DATAVIEW_EXPORT}" else: filename = xform.id_string return filename def _set_start_end_params(request, query): # check for start and end params if "start" in request.GET or "end" in request.GET: query = json.loads(query) if isinstance(query, six.string_types) else query query[SUBMISSION_TIME] = {} try: if request.GET.get("start"): query[SUBMISSION_TIME]["$gte"] = _format_date_for_mongo( request.GET["start"] ) if request.GET.get("end"): query[SUBMISSION_TIME]["$lte"] = _format_date_for_mongo( request.GET["end"] ) except ValueError as e: raise exceptions.ParseError( _("Dates must be in the format YY_MM_DD_hh_mm_ss") ) from e else: query = json.dumps(query) return query def _get_extension_from_export_type(export_type): extension = export_type if export_type == Export.XLSX_EXPORT: extension = "xlsx" elif export_type
raise exceptions.ParseError(export.error_message)
conditional_block
api_export_tools.py
's xform :return True or False """ return bool(set(hxl_columns).intersection(set(dv_columns))) def _get_export_type(export_type): if export_type not in EXPORT_EXT or ( export_type == Export.GOOGLE_SHEETS_EXPORT and not getattr(settings, "GOOGLE_EXPORT", False) ): raise exceptions.ParseError( _(f"'{export_type}' format not known or not implemented!") ) return EXPORT_EXT[export_type] # pylint: disable=too-many-arguments, too-many-locals, too-many-branches def custom_response_handler( # noqa: C0901 request, xform, query, export_type, token=None, meta=None, dataview=False, filename=None, metadata=None, ): """ Returns a HTTP response with export file for download. """ export_type = _get_export_type(export_type) if ( export_type in EXTERNAL_EXPORT_TYPES and (token is not None) or (meta is not None) ): export_type = Export.EXTERNAL_EXPORT options = parse_request_export_options(request.query_params) dataview_pk = hasattr(dataview, "pk") and dataview.pk options["dataview_pk"] = dataview_pk options["host"] = request.get_host() if dataview: columns_with_hxl = get_columns_with_hxl(xform.survey.get("children")) if columns_with_hxl: options["include_hxl"] = include_hxl_row( dataview.columns, list(columns_with_hxl) ) try: query = filter_queryset_xform_meta_perms_sql(xform, request.user, query) except NoRecordsPermission: return Response( data=json.dumps({"details": _("You don't have permission")}), status=status.HTTP_403_FORBIDDEN, content_type="application/json", ) if query: options["query"] = query remove_group_name = options.get("remove_group_name") export_id = request.query_params.get("export_id") if export_id: export = get_object_or_404(Export, id=export_id, xform=xform) else: if export_type == Export.GOOGLE_SHEETS_EXPORT: return Response( data=json.dumps( {"details": _("Sheets export only supported in async mode")} ), status=status.HTTP_403_FORBIDDEN, content_type="application/json", ) # check if we need to re-generate, # we always re-generate if a filter is specified def _new_export(): return _generate_new_export( request, xform, query, export_type, dataview_pk=dataview_pk, metadata=metadata, ) if should_create_new_export(xform, export_type, options, request=request): export = _new_export() else: export = newest_export_for(xform, export_type, options) if not export.filename and not export.error_message: export = _new_export() log_export(request, xform, export_type) if export_type == Export.EXTERNAL_EXPORT: return external_export_response(export) if export.filename is None and export.error_message: raise exceptions.ParseError(export.error_message) # get extension from file_path, exporter could modify to # xlsx if it exceeds limits _path, ext = os.path.splitext(export.filename) ext = ext[1:] show_date = True if filename is None and export.status == Export.SUCCESSFUL: filename = _generate_filename( request, xform, remove_group_name, dataview_pk=dataview_pk ) else: show_date = False response = response_with_mimetype_and_name( Export.EXPORT_MIMES[ext], filename, extension=ext, show_date=show_date, file_path=export.filepath, ) return response def _generate_new_export( # noqa: C0901 request, xform, query, export_type, dataview_pk=False, metadata=None ): query = _set_start_end_params(request, query) extension = _get_extension_from_export_type(export_type) options = { "extension": extension, "username": xform.user.username, "id_string": xform.id_string, "host": request.get_host(), "sort": request.query_params.get('sort') } if query: options["query"] = query options["dataview_pk"] = dataview_pk if export_type == Export.GOOGLE_SHEETS_EXPORT: options["google_credentials"] = _get_google_credential(request).to_json() try: if export_type == Export.EXTERNAL_EXPORT: options["token"] = request.GET.get("token") options["data_id"] = request.GET.get("data_id") options["meta"] = request.GET.get("meta") export = generate_external_export( export_type, xform.user.username, xform.id_string, None, options, xform=xform, ) elif export_type == Export.OSM_EXPORT: export = generate_osm_export( export_type, xform.user.username, xform.id_string, None, options, xform=xform, ) elif export_type == Export.ZIP_EXPORT: export = generate_attachments_zip_export( export_type, xform.user.username, xform.id_string, None, options, xform=xform, ) elif export_type == Export.KML_EXPORT: export = generate_kml_export( export_type, xform.user.username, xform.id_string, None, options, xform=xform, ) elif export_type == Export.GEOJSON_EXPORT: export = generate_geojson_export( export_type, xform.user.username, xform.id_string, metadata, None, options, xform=xform, ) else: options.update(parse_request_export_options(request.query_params)) export = generate_export(export_type, xform, None, options) audit = {"xform": xform.id_string, "export_type": export_type} log.audit_log( log.Actions.EXPORT_CREATED, request.user, xform.user, _("Created %(export_type)s export on '%(id_string)s'.") % {"id_string": xform.id_string, "export_type": export_type.upper()}, audit, request, ) except NoRecordsFoundError as e: raise Http404(_("No records found to export")) from e except J2XException as e: # j2x exception return async_status(FAILED, str(e)) except SPSSIOError as e: raise exceptions.ParseError(str(e)) from e else: return export def log_export(request, xform, export_type): """ Logs audit logs of export requests. """ # log download as well audit = {"xform": xform.id_string, "export_type": export_type} log.audit_log( log.Actions.EXPORT_DOWNLOADED, request.user, xform.user, _(f"Downloaded {export_type.upper()} export on '{xform.id_string}'."), audit, request, ) def external_export_response(export):
def _generate_filename(request, xform, remove_group_name=False, dataview_pk=False): if request.GET.get("raw"): filename = None else: # append group name removed flag otherwise use the form id_string if remove_group_name: filename = f"{xform.id_string}-{GROUPNAME_REMOVED_FLAG}" elif dataview_pk: filename = f"{xform.id_string}-{DATAVIEW_EXPORT}" else: filename = xform.id_string return filename def _set_start_end_params(request, query): # check for start and end params if "start" in request.GET or "end" in request.GET: query = json.loads(query) if isinstance(query, six.string_types) else query query[SUBMISSION_TIME] = {} try: if request.GET.get("start"): query[SUBMISSION_TIME]["$gte"] = _format_date_for_mongo( request.GET["start"] ) if request.GET.get("end"): query[SUBMISSION_TIME]["$lte"] = _format_date_for_mongo( request.GET["end"] ) except ValueError as e: raise exceptions.ParseError( _("Dates must be in the format YY_MM_DD_hh_mm_ss") ) from e else: query = json.dumps(query) return query def _get_extension_from_export_type(export_type): extension = export_type if export_type == Export.XLSX_EXPORT: extension = "xlsx" elif export_type
""" Redirects to export_url of XLSReports successful export. In case of a failure, returns a 400 HTTP JSON response with the error message. """ if isinstance(export, Export) and export.internal_status == Export.SUCCESSFUL: return HttpResponseRedirect(export.export_url) http_status = status.HTTP_400_BAD_REQUEST return Response(json.dumps(export), http_status, content_type="application/json")
identifier_body
api_export_tools.py
_columns))) def _get_export_type(export_type): if export_type not in EXPORT_EXT or ( export_type == Export.GOOGLE_SHEETS_EXPORT and not getattr(settings, "GOOGLE_EXPORT", False) ): raise exceptions.ParseError( _(f"'{export_type}' format not known or not implemented!") ) return EXPORT_EXT[export_type] # pylint: disable=too-many-arguments, too-many-locals, too-many-branches def custom_response_handler( # noqa: C0901 request, xform, query, export_type, token=None, meta=None, dataview=False, filename=None, metadata=None, ): """ Returns a HTTP response with export file for download. """ export_type = _get_export_type(export_type) if ( export_type in EXTERNAL_EXPORT_TYPES and (token is not None) or (meta is not None) ): export_type = Export.EXTERNAL_EXPORT options = parse_request_export_options(request.query_params) dataview_pk = hasattr(dataview, "pk") and dataview.pk options["dataview_pk"] = dataview_pk options["host"] = request.get_host() if dataview: columns_with_hxl = get_columns_with_hxl(xform.survey.get("children")) if columns_with_hxl: options["include_hxl"] = include_hxl_row( dataview.columns, list(columns_with_hxl) ) try: query = filter_queryset_xform_meta_perms_sql(xform, request.user, query) except NoRecordsPermission: return Response( data=json.dumps({"details": _("You don't have permission")}), status=status.HTTP_403_FORBIDDEN, content_type="application/json", ) if query: options["query"] = query remove_group_name = options.get("remove_group_name") export_id = request.query_params.get("export_id") if export_id: export = get_object_or_404(Export, id=export_id, xform=xform) else: if export_type == Export.GOOGLE_SHEETS_EXPORT: return Response( data=json.dumps( {"details": _("Sheets export only supported in async mode")} ), status=status.HTTP_403_FORBIDDEN, content_type="application/json", ) # check if we need to re-generate, # we always re-generate if a filter is specified def _new_export(): return _generate_new_export( request, xform, query, export_type, dataview_pk=dataview_pk, metadata=metadata, ) if should_create_new_export(xform, export_type, options, request=request): export = _new_export() else: export = newest_export_for(xform, export_type, options) if not export.filename and not export.error_message: export = _new_export() log_export(request, xform, export_type) if export_type == Export.EXTERNAL_EXPORT: return external_export_response(export) if export.filename is None and export.error_message: raise exceptions.ParseError(export.error_message) # get extension from file_path, exporter could modify to # xlsx if it exceeds limits _path, ext = os.path.splitext(export.filename) ext = ext[1:] show_date = True if filename is None and export.status == Export.SUCCESSFUL: filename = _generate_filename( request, xform, remove_group_name, dataview_pk=dataview_pk ) else: show_date = False response = response_with_mimetype_and_name( Export.EXPORT_MIMES[ext], filename, extension=ext, show_date=show_date, file_path=export.filepath, ) return response def _generate_new_export( # noqa: C0901 request, xform, query, export_type, dataview_pk=False, metadata=None ): query = _set_start_end_params(request, query) extension = _get_extension_from_export_type(export_type) options = { "extension": extension, "username": xform.user.username, "id_string": xform.id_string, "host": request.get_host(), "sort": request.query_params.get('sort') } if query: options["query"] = query options["dataview_pk"] = dataview_pk if export_type == Export.GOOGLE_SHEETS_EXPORT: options["google_credentials"] = _get_google_credential(request).to_json() try: if export_type == Export.EXTERNAL_EXPORT: options["token"] = request.GET.get("token") options["data_id"] = request.GET.get("data_id") options["meta"] = request.GET.get("meta") export = generate_external_export( export_type, xform.user.username, xform.id_string, None, options, xform=xform, ) elif export_type == Export.OSM_EXPORT: export = generate_osm_export( export_type, xform.user.username, xform.id_string, None, options, xform=xform, ) elif export_type == Export.ZIP_EXPORT: export = generate_attachments_zip_export( export_type, xform.user.username, xform.id_string, None, options, xform=xform, ) elif export_type == Export.KML_EXPORT: export = generate_kml_export( export_type, xform.user.username, xform.id_string, None, options, xform=xform, ) elif export_type == Export.GEOJSON_EXPORT: export = generate_geojson_export( export_type, xform.user.username, xform.id_string, metadata, None, options, xform=xform, ) else: options.update(parse_request_export_options(request.query_params)) export = generate_export(export_type, xform, None, options) audit = {"xform": xform.id_string, "export_type": export_type} log.audit_log( log.Actions.EXPORT_CREATED, request.user, xform.user, _("Created %(export_type)s export on '%(id_string)s'.") % {"id_string": xform.id_string, "export_type": export_type.upper()}, audit, request, ) except NoRecordsFoundError as e: raise Http404(_("No records found to export")) from e except J2XException as e: # j2x exception return async_status(FAILED, str(e)) except SPSSIOError as e: raise exceptions.ParseError(str(e)) from e else: return export def log_export(request, xform, export_type): """ Logs audit logs of export requests. """ # log download as well audit = {"xform": xform.id_string, "export_type": export_type} log.audit_log( log.Actions.EXPORT_DOWNLOADED, request.user, xform.user, _(f"Downloaded {export_type.upper()} export on '{xform.id_string}'."), audit, request, ) def external_export_response(export): """ Redirects to export_url of XLSReports successful export. In case of a failure, returns a 400 HTTP JSON response with the error message. """ if isinstance(export, Export) and export.internal_status == Export.SUCCESSFUL: return HttpResponseRedirect(export.export_url) http_status = status.HTTP_400_BAD_REQUEST return Response(json.dumps(export), http_status, content_type="application/json") def _generate_filename(request, xform, remove_group_name=False, dataview_pk=False): if request.GET.get("raw"): filename = None else: # append group name removed flag otherwise use the form id_string if remove_group_name: filename = f"{xform.id_string}-{GROUPNAME_REMOVED_FLAG}" elif dataview_pk: filename = f"{xform.id_string}-{DATAVIEW_EXPORT}" else: filename = xform.id_string return filename def _set_start_end_params(request, query): # check for start and end params if "start" in request.GET or "end" in request.GET: query = json.loads(query) if isinstance(query, six.string_types) else query query[SUBMISSION_TIME] = {} try: if request.GET.get("start"): query[SUBMISSION_TIME]["$gte"] = _format_date_for_mongo( request.GET["start"] ) if request.GET.get("end"): query[SUBMISSION_TIME]["$lte"] = _format_date_for_mongo( request.GET["end"] ) except ValueError as e: raise exceptions.ParseError( _("Dates must be in the format YY_MM_DD_hh_mm_ss") ) from e else: query = json.dumps(query) return query def _get_extension_from_export_type(export_type): extension = export_type if export_type == Export.XLSX_EXPORT: extension = "xlsx"
elif export_type in [Export.CSV_ZIP_EXPORT, Export.SAV_ZIP_EXPORT]: extension = "zip" return extension
random_line_split
api_export_tools.py
iew's xform :return True or False """ return bool(set(hxl_columns).intersection(set(dv_columns))) def _get_export_type(export_type): if export_type not in EXPORT_EXT or ( export_type == Export.GOOGLE_SHEETS_EXPORT and not getattr(settings, "GOOGLE_EXPORT", False) ): raise exceptions.ParseError( _(f"'{export_type}' format not known or not implemented!") ) return EXPORT_EXT[export_type] # pylint: disable=too-many-arguments, too-many-locals, too-many-branches def custom_response_handler( # noqa: C0901 request, xform, query, export_type, token=None, meta=None, dataview=False, filename=None, metadata=None, ): """ Returns a HTTP response with export file for download. """ export_type = _get_export_type(export_type) if ( export_type in EXTERNAL_EXPORT_TYPES and (token is not None) or (meta is not None) ): export_type = Export.EXTERNAL_EXPORT options = parse_request_export_options(request.query_params) dataview_pk = hasattr(dataview, "pk") and dataview.pk options["dataview_pk"] = dataview_pk options["host"] = request.get_host() if dataview: columns_with_hxl = get_columns_with_hxl(xform.survey.get("children")) if columns_with_hxl: options["include_hxl"] = include_hxl_row( dataview.columns, list(columns_with_hxl) ) try: query = filter_queryset_xform_meta_perms_sql(xform, request.user, query) except NoRecordsPermission: return Response( data=json.dumps({"details": _("You don't have permission")}), status=status.HTTP_403_FORBIDDEN, content_type="application/json", ) if query: options["query"] = query remove_group_name = options.get("remove_group_name") export_id = request.query_params.get("export_id") if export_id: export = get_object_or_404(Export, id=export_id, xform=xform) else: if export_type == Export.GOOGLE_SHEETS_EXPORT: return Response( data=json.dumps( {"details": _("Sheets export only supported in async mode")} ), status=status.HTTP_403_FORBIDDEN, content_type="application/json", ) # check if we need to re-generate, # we always re-generate if a filter is specified def _new_export(): return _generate_new_export( request, xform, query, export_type, dataview_pk=dataview_pk, metadata=metadata, ) if should_create_new_export(xform, export_type, options, request=request): export = _new_export() else: export = newest_export_for(xform, export_type, options) if not export.filename and not export.error_message: export = _new_export() log_export(request, xform, export_type) if export_type == Export.EXTERNAL_EXPORT: return external_export_response(export) if export.filename is None and export.error_message: raise exceptions.ParseError(export.error_message) # get extension from file_path, exporter could modify to # xlsx if it exceeds limits _path, ext = os.path.splitext(export.filename) ext = ext[1:] show_date = True if filename is None and export.status == Export.SUCCESSFUL: filename = _generate_filename( request, xform, remove_group_name, dataview_pk=dataview_pk ) else: show_date = False response = response_with_mimetype_and_name( Export.EXPORT_MIMES[ext], filename, extension=ext, show_date=show_date, file_path=export.filepath, ) return response def
( # noqa: C0901 request, xform, query, export_type, dataview_pk=False, metadata=None ): query = _set_start_end_params(request, query) extension = _get_extension_from_export_type(export_type) options = { "extension": extension, "username": xform.user.username, "id_string": xform.id_string, "host": request.get_host(), "sort": request.query_params.get('sort') } if query: options["query"] = query options["dataview_pk"] = dataview_pk if export_type == Export.GOOGLE_SHEETS_EXPORT: options["google_credentials"] = _get_google_credential(request).to_json() try: if export_type == Export.EXTERNAL_EXPORT: options["token"] = request.GET.get("token") options["data_id"] = request.GET.get("data_id") options["meta"] = request.GET.get("meta") export = generate_external_export( export_type, xform.user.username, xform.id_string, None, options, xform=xform, ) elif export_type == Export.OSM_EXPORT: export = generate_osm_export( export_type, xform.user.username, xform.id_string, None, options, xform=xform, ) elif export_type == Export.ZIP_EXPORT: export = generate_attachments_zip_export( export_type, xform.user.username, xform.id_string, None, options, xform=xform, ) elif export_type == Export.KML_EXPORT: export = generate_kml_export( export_type, xform.user.username, xform.id_string, None, options, xform=xform, ) elif export_type == Export.GEOJSON_EXPORT: export = generate_geojson_export( export_type, xform.user.username, xform.id_string, metadata, None, options, xform=xform, ) else: options.update(parse_request_export_options(request.query_params)) export = generate_export(export_type, xform, None, options) audit = {"xform": xform.id_string, "export_type": export_type} log.audit_log( log.Actions.EXPORT_CREATED, request.user, xform.user, _("Created %(export_type)s export on '%(id_string)s'.") % {"id_string": xform.id_string, "export_type": export_type.upper()}, audit, request, ) except NoRecordsFoundError as e: raise Http404(_("No records found to export")) from e except J2XException as e: # j2x exception return async_status(FAILED, str(e)) except SPSSIOError as e: raise exceptions.ParseError(str(e)) from e else: return export def log_export(request, xform, export_type): """ Logs audit logs of export requests. """ # log download as well audit = {"xform": xform.id_string, "export_type": export_type} log.audit_log( log.Actions.EXPORT_DOWNLOADED, request.user, xform.user, _(f"Downloaded {export_type.upper()} export on '{xform.id_string}'."), audit, request, ) def external_export_response(export): """ Redirects to export_url of XLSReports successful export. In case of a failure, returns a 400 HTTP JSON response with the error message. """ if isinstance(export, Export) and export.internal_status == Export.SUCCESSFUL: return HttpResponseRedirect(export.export_url) http_status = status.HTTP_400_BAD_REQUEST return Response(json.dumps(export), http_status, content_type="application/json") def _generate_filename(request, xform, remove_group_name=False, dataview_pk=False): if request.GET.get("raw"): filename = None else: # append group name removed flag otherwise use the form id_string if remove_group_name: filename = f"{xform.id_string}-{GROUPNAME_REMOVED_FLAG}" elif dataview_pk: filename = f"{xform.id_string}-{DATAVIEW_EXPORT}" else: filename = xform.id_string return filename def _set_start_end_params(request, query): # check for start and end params if "start" in request.GET or "end" in request.GET: query = json.loads(query) if isinstance(query, six.string_types) else query query[SUBMISSION_TIME] = {} try: if request.GET.get("start"): query[SUBMISSION_TIME]["$gte"] = _format_date_for_mongo( request.GET["start"] ) if request.GET.get("end"): query[SUBMISSION_TIME]["$lte"] = _format_date_for_mongo( request.GET["end"] ) except ValueError as e: raise exceptions.ParseError( _("Dates must be in the format YY_MM_DD_hh_mm_ss") ) from e else: query = json.dumps(query) return query def _get_extension_from_export_type(export_type): extension = export_type if export_type == Export.XLSX_EXPORT: extension = "xlsx" elif export_type
_generate_new_export
identifier_name
custom_widget.rs
the widget color depending on the current interaction. fn color(&self, color: Color) -> Color { match *self { /// The base color as defined in the Style struct, or a default provided /// by the current Theme if the Style has no color. Interaction::Normal => color, /// The Color object (from Elmesque) can calculate a highlighted version /// of itself. We don't have to use it, though. We could specify any color /// we want. Interaction::Highlighted => color.highlighted(), /// Ditto for clicked. Interaction::Clicked => color.clicked(), } } } /// Check the current interaction with the button. Takes into account whether the mouse is /// over the button and the previous interaction state. fn get_new_interaction(is_over: bool, prev: Interaction, mouse: Mouse) -> Interaction { use conrod::MouseButtonPosition::{Down, Up}; use self::Interaction::{Normal, Highlighted, Clicked}; match (is_over, prev, mouse.left.position) { // LMB is down over the button. But the button wasn't Highlighted last // update. This means the user clicked somewhere outside the button and // moved over the button holding LMB down. We do nothing in this case. (true, Normal, Down) => Normal, // LMB is down over the button. The button was either Highlighted or Clicked // last update. If it was highlighted before, that means the user clicked // just now, and we transition to the Clicked state. If it was clicked // before, that means the user is still holding LMB down from a previous // click, in which case the state remains Clicked. (true, _, Down) => Clicked, // LMB is up. The mouse is hovering over the button. Regardless of what the // state was last update, the state should definitely be Highlighted now. (true, _, Up) => Highlighted, // LMB is down, the mouse is not over the button, but the previous state was // Clicked. That means the user clicked the button and then moved the mouse // outside the button while holding LMB down. The button stays Clicked. (false, Clicked, Down) => Clicked, // If none of the above applies, then nothing interesting is happening with // this button. _ => Normal, } } /// Return whether or not a given point is over a circle at a given point on a /// Cartesian plane. We use this to determine whether the mouse is over the button. pub fn is_over_circ(circ_center: Point, mouse_point: Point, dim: Dimensions) -> bool { // Offset vector from the center of the circle to the mouse. let offset = ::vecmath::vec2_sub(mouse_point, circ_center); // If the length of the offset vector is less than or equal to the circle's // radius, then the mouse is inside the circle. We assume that dim is a square // bounding box around the circle, thus 2 * radius == dim[0] == dim[1]. ::vecmath::vec2_len(offset) <= dim[0] / 2.0 } impl<'a, F> CircularButton<'a, F> { /// Create a button context to be built upon. pub fn new() -> CircularButton<'a, F> { CircularButton { common: CommonBuilder::new(), maybe_react: None, maybe_label: None, style: Style::new(), enabled: true, } } /// Set the reaction for the Button. The reaction will be triggered upon release /// of the button. Like other Conrod configs, this returns self for chainability. pub fn react(mut self, reaction: F) -> Self { self.maybe_react = Some(reaction); self } /// If true, will allow user inputs. If false, will disallow user inputs. Like /// other Conrod configs, this returns self for chainability. Allow dead code /// because we never call this in the example. #[allow(dead_code)] pub fn enabled(mut self, flag: bool) -> Self { self.enabled = flag; self } } /// A custom Conrod widget must implement the Widget trait. See the **Widget** trait /// documentation for more details. impl<'a, F> Widget for CircularButton<'a, F> where F: FnMut() { /// The State struct that we defined above. type State = State; /// The Style struct that we defined above. type Style = Style; fn common(&self) -> &CommonBuilder { &self.common } fn common_mut(&mut self) -> &mut CommonBuilder { &mut self.common } fn unique_kind(&self) -> &'static str { KIND } fn init_state(&self) -> State { State { interaction: Interaction::Normal, circle_idx: IndexSlot::new(), text_idx: IndexSlot::new(), } } fn style(&self) -> Style { self.style.clone() } /// Default width of the widget. /// /// This method is optional. ///
// // Defaults can come from several places. Here, we define how certain defaults take // precedence over others. // // Most commonly, defaults are to be retrieved from the `Theme`, however in some cases // some other logic may need to be considered. default_x_dimension(self, ui).unwrap_or(Dimension::Absolute(64.0)) } /// Default height of the widget. /// /// This method is optional. /// /// The default implementation is the same as below, but unwraps to an absolute scalar of /// `0.0` instead of `64.0`. fn default_y_dimension<C: CharacterCache>(&self, ui: &Ui<C>) -> Dimension { default_y_dimension(self, ui).unwrap_or(Dimension::Absolute(64.0)) } /// Update the state of the button. The state may or may not have changed since /// the last update. (E.g. it may have changed because the user moused over the /// button.) If the state has changed, return the new state. Else, return None. fn update<C: CharacterCache>(mut self, args: UpdateArgs<Self, C>) { let UpdateArgs { idx, state, rect, mut ui, style, .. } = args; let (xy, dim) = rect.xy_dim(); let maybe_mouse = ui.input().maybe_mouse.map(|mouse| mouse.relative_to(xy)); // Check whether or not a new interaction has occurred. let new_interaction = match (self.enabled, maybe_mouse) { (false, _) | (true, None) => Interaction::Normal, (true, Some(mouse)) => { // Conrod does us a favor by transforming mouse.xy into this widget's // local coordinate system. Because mouse.xy is in local coords, // we must also pass the circle center in local coords. Thus we pass // [0.0, 0.0] as the center. // // See above where we define is_over_circ. let is_over = is_over_circ([0.0, 0.0], mouse.xy, dim); // See above where we define get_new_interaction. get_new_interaction(is_over, state.view().interaction, mouse) }, }; // If the mouse was released over the button, react. state.interaction is the // button's state as of a moment ago. new_interaction is the updated state as // of right now. So this if statement is saying: If the button was clicked a // moment ago, and it's now highlighted, then the button has been activated. if let (Interaction::Clicked, Interaction::Highlighted) = (state.view().interaction, new_interaction) { // Recall that our CircularButton struct includes maybe_react, which // stores either a reaction function or None. If maybe_react is Some, call // the function. if let Some(ref mut react) = self.maybe_react { react(); } } // Here we check to see whether or not our button should capture the mouse. // // Widgets can "capture" user input. If the button captures the mouse, then mouse // events will only be seen by the button. Other widgets will not see mouse events // until the button uncaptures the mouse. match (state.view().interaction, new_interaction) { // If the user has pressed the button we capture the mouse. (Interaction::Highlighted, Interaction::Clicked) => { ui.capture_mouse(); }, // If the user releases the button, we uncapture the mouse. (Interaction::
/// The default implementation is the same as below, but unwraps to an absolute scalar of /// `0.0` instead of `64.0`. fn default_x_dimension<C: CharacterCache>(&self, ui: &Ui<C>) -> Dimension { // If no width was given via the `Sizeable` (a trait implemented for all widgets) // methods, some default width must be chosen.
random_line_split
custom_widget.rs
widget color depending on the current interaction. fn color(&self, color: Color) -> Color { match *self { /// The base color as defined in the Style struct, or a default provided /// by the current Theme if the Style has no color. Interaction::Normal => color, /// The Color object (from Elmesque) can calculate a highlighted version /// of itself. We don't have to use it, though. We could specify any color /// we want. Interaction::Highlighted => color.highlighted(), /// Ditto for clicked. Interaction::Clicked => color.clicked(), } } } /// Check the current interaction with the button. Takes into account whether the mouse is /// over the button and the previous interaction state. fn get_new_interaction(is_over: bool, prev: Interaction, mouse: Mouse) -> Interaction { use conrod::MouseButtonPosition::{Down, Up}; use self::Interaction::{Normal, Highlighted, Clicked}; match (is_over, prev, mouse.left.position) { // LMB is down over the button. But the button wasn't Highlighted last // update. This means the user clicked somewhere outside the button and // moved over the button holding LMB down. We do nothing in this case. (true, Normal, Down) => Normal, // LMB is down over the button. The button was either Highlighted or Clicked // last update. If it was highlighted before, that means the user clicked // just now, and we transition to the Clicked state. If it was clicked // before, that means the user is still holding LMB down from a previous // click, in which case the state remains Clicked. (true, _, Down) => Clicked, // LMB is up. The mouse is hovering over the button. Regardless of what the // state was last update, the state should definitely be Highlighted now. (true, _, Up) => Highlighted, // LMB is down, the mouse is not over the button, but the previous state was // Clicked. That means the user clicked the button and then moved the mouse // outside the button while holding LMB down. The button stays Clicked. (false, Clicked, Down) => Clicked, // If none of the above applies, then nothing interesting is happening with // this button. _ => Normal, } } /// Return whether or not a given point is over a circle at a given point on a /// Cartesian plane. We use this to determine whether the mouse is over the button. pub fn is_over_circ(circ_center: Point, mouse_point: Point, dim: Dimensions) -> bool { // Offset vector from the center of the circle to the mouse. let offset = ::vecmath::vec2_sub(mouse_point, circ_center); // If the length of the offset vector is less than or equal to the circle's // radius, then the mouse is inside the circle. We assume that dim is a square // bounding box around the circle, thus 2 * radius == dim[0] == dim[1]. ::vecmath::vec2_len(offset) <= dim[0] / 2.0 } impl<'a, F> CircularButton<'a, F> { /// Create a button context to be built upon. pub fn new() -> CircularButton<'a, F> { CircularButton { common: CommonBuilder::new(), maybe_react: None, maybe_label: None, style: Style::new(), enabled: true, } } /// Set the reaction for the Button. The reaction will be triggered upon release /// of the button. Like other Conrod configs, this returns self for chainability. pub fn react(mut self, reaction: F) -> Self { self.maybe_react = Some(reaction); self } /// If true, will allow user inputs. If false, will disallow user inputs. Like /// other Conrod configs, this returns self for chainability. Allow dead code /// because we never call this in the example. #[allow(dead_code)] pub fn enabled(mut self, flag: bool) -> Self { self.enabled = flag; self } } /// A custom Conrod widget must implement the Widget trait. See the **Widget** trait /// documentation for more details. impl<'a, F> Widget for CircularButton<'a, F> where F: FnMut() { /// The State struct that we defined above. type State = State; /// The Style struct that we defined above. type Style = Style; fn common(&self) -> &CommonBuilder { &self.common } fn common_mut(&mut self) -> &mut CommonBuilder { &mut self.common } fn unique_kind(&self) -> &'static str { KIND } fn init_state(&self) -> State { State { interaction: Interaction::Normal, circle_idx: IndexSlot::new(), text_idx: IndexSlot::new(), } } fn style(&self) -> Style { self.style.clone() } /// Default width of the widget. /// /// This method is optional. /// /// The default implementation is the same as below, but unwraps to an absolute scalar of /// `0.0` instead of `64.0`. fn default_x_dimension<C: CharacterCache>(&self, ui: &Ui<C>) -> Dimension { // If no width was given via the `Sizeable` (a trait implemented for all widgets) // methods, some default width must be chosen. // // Defaults can come from several places. Here, we define how certain defaults take // precedence over others. // // Most commonly, defaults are to be retrieved from the `Theme`, however in some cases // some other logic may need to be considered. default_x_dimension(self, ui).unwrap_or(Dimension::Absolute(64.0)) } /// Default height of the widget. /// /// This method is optional. /// /// The default implementation is the same as below, but unwraps to an absolute scalar of /// `0.0` instead of `64.0`. fn default_y_dimension<C: CharacterCache>(&self, ui: &Ui<C>) -> Dimension { default_y_dimension(self, ui).unwrap_or(Dimension::Absolute(64.0)) } /// Update the state of the button. The state may or may not have changed since /// the last update. (E.g. it may have changed because the user moused over the /// button.) If the state has changed, return the new state. Else, return None. fn update<C: CharacterCache>(mut self, args: UpdateArgs<Self, C>) { let UpdateArgs { idx, state, rect, mut ui, style, .. } = args; let (xy, dim) = rect.xy_dim(); let maybe_mouse = ui.input().maybe_mouse.map(|mouse| mouse.relative_to(xy)); // Check whether or not a new interaction has occurred. let new_interaction = match (self.enabled, maybe_mouse) { (false, _) | (true, None) => Interaction::Normal, (true, Some(mouse)) => { // Conrod does us a favor by transforming mouse.xy into this widget's // local coordinate system. Because mouse.xy is in local coords, // we must also pass the circle center in local coords. Thus we pass // [0.0, 0.0] as the center. // // See above where we define is_over_circ. let is_over = is_over_circ([0.0, 0.0], mouse.xy, dim); // See above where we define get_new_interaction. get_new_interaction(is_over, state.view().interaction, mouse) }, }; // If the mouse was released over the button, react. state.interaction is the // button's state as of a moment ago. new_interaction is the updated state as // of right now. So this if statement is saying: If the button was clicked a // moment ago, and it's now highlighted, then the button has been activated. if let (Interaction::Clicked, Interaction::Highlighted) = (state.view().interaction, new_interaction) { // Recall that our CircularButton struct includes maybe_react, which // stores either a reaction function or None. If maybe_react is Some, call // the function. if let Some(ref mut react) = self.maybe_react
} // Here we check to see whether or not our button should capture the mouse. // // Widgets can "capture" user input. If the button captures the mouse, then mouse // events will only be seen by the button. Other widgets will not see mouse events // until the button uncaptures the mouse. match (state.view().interaction, new_interaction) { // If the user has pressed the button we capture the mouse. (Interaction::Highlighted, Interaction::Clicked) => { ui.capture_mouse(); }, // If the user releases the button, we uncapture the mouse. (Interaction
{ react(); }
conditional_block
custom_widget.rs
widget color depending on the current interaction. fn color(&self, color: Color) -> Color { match *self { /// The base color as defined in the Style struct, or a default provided /// by the current Theme if the Style has no color. Interaction::Normal => color, /// The Color object (from Elmesque) can calculate a highlighted version /// of itself. We don't have to use it, though. We could specify any color /// we want. Interaction::Highlighted => color.highlighted(), /// Ditto for clicked. Interaction::Clicked => color.clicked(), } } } /// Check the current interaction with the button. Takes into account whether the mouse is /// over the button and the previous interaction state. fn get_new_interaction(is_over: bool, prev: Interaction, mouse: Mouse) -> Interaction { use conrod::MouseButtonPosition::{Down, Up}; use self::Interaction::{Normal, Highlighted, Clicked}; match (is_over, prev, mouse.left.position) { // LMB is down over the button. But the button wasn't Highlighted last // update. This means the user clicked somewhere outside the button and // moved over the button holding LMB down. We do nothing in this case. (true, Normal, Down) => Normal, // LMB is down over the button. The button was either Highlighted or Clicked // last update. If it was highlighted before, that means the user clicked // just now, and we transition to the Clicked state. If it was clicked // before, that means the user is still holding LMB down from a previous // click, in which case the state remains Clicked. (true, _, Down) => Clicked, // LMB is up. The mouse is hovering over the button. Regardless of what the // state was last update, the state should definitely be Highlighted now. (true, _, Up) => Highlighted, // LMB is down, the mouse is not over the button, but the previous state was // Clicked. That means the user clicked the button and then moved the mouse // outside the button while holding LMB down. The button stays Clicked. (false, Clicked, Down) => Clicked, // If none of the above applies, then nothing interesting is happening with // this button. _ => Normal, } } /// Return whether or not a given point is over a circle at a given point on a /// Cartesian plane. We use this to determine whether the mouse is over the button. pub fn is_over_circ(circ_center: Point, mouse_point: Point, dim: Dimensions) -> bool { // Offset vector from the center of the circle to the mouse. let offset = ::vecmath::vec2_sub(mouse_point, circ_center); // If the length of the offset vector is less than or equal to the circle's // radius, then the mouse is inside the circle. We assume that dim is a square // bounding box around the circle, thus 2 * radius == dim[0] == dim[1]. ::vecmath::vec2_len(offset) <= dim[0] / 2.0 } impl<'a, F> CircularButton<'a, F> { /// Create a button context to be built upon. pub fn new() -> CircularButton<'a, F> { CircularButton { common: CommonBuilder::new(), maybe_react: None, maybe_label: None, style: Style::new(), enabled: true, } } /// Set the reaction for the Button. The reaction will be triggered upon release /// of the button. Like other Conrod configs, this returns self for chainability. pub fn react(mut self, reaction: F) -> Self { self.maybe_react = Some(reaction); self } /// If true, will allow user inputs. If false, will disallow user inputs. Like /// other Conrod configs, this returns self for chainability. Allow dead code /// because we never call this in the example. #[allow(dead_code)] pub fn enabled(mut self, flag: bool) -> Self { self.enabled = flag; self } } /// A custom Conrod widget must implement the Widget trait. See the **Widget** trait /// documentation for more details. impl<'a, F> Widget for CircularButton<'a, F> where F: FnMut() { /// The State struct that we defined above. type State = State; /// The Style struct that we defined above. type Style = Style; fn
(&self) -> &CommonBuilder { &self.common } fn common_mut(&mut self) -> &mut CommonBuilder { &mut self.common } fn unique_kind(&self) -> &'static str { KIND } fn init_state(&self) -> State { State { interaction: Interaction::Normal, circle_idx: IndexSlot::new(), text_idx: IndexSlot::new(), } } fn style(&self) -> Style { self.style.clone() } /// Default width of the widget. /// /// This method is optional. /// /// The default implementation is the same as below, but unwraps to an absolute scalar of /// `0.0` instead of `64.0`. fn default_x_dimension<C: CharacterCache>(&self, ui: &Ui<C>) -> Dimension { // If no width was given via the `Sizeable` (a trait implemented for all widgets) // methods, some default width must be chosen. // // Defaults can come from several places. Here, we define how certain defaults take // precedence over others. // // Most commonly, defaults are to be retrieved from the `Theme`, however in some cases // some other logic may need to be considered. default_x_dimension(self, ui).unwrap_or(Dimension::Absolute(64.0)) } /// Default height of the widget. /// /// This method is optional. /// /// The default implementation is the same as below, but unwraps to an absolute scalar of /// `0.0` instead of `64.0`. fn default_y_dimension<C: CharacterCache>(&self, ui: &Ui<C>) -> Dimension { default_y_dimension(self, ui).unwrap_or(Dimension::Absolute(64.0)) } /// Update the state of the button. The state may or may not have changed since /// the last update. (E.g. it may have changed because the user moused over the /// button.) If the state has changed, return the new state. Else, return None. fn update<C: CharacterCache>(mut self, args: UpdateArgs<Self, C>) { let UpdateArgs { idx, state, rect, mut ui, style, .. } = args; let (xy, dim) = rect.xy_dim(); let maybe_mouse = ui.input().maybe_mouse.map(|mouse| mouse.relative_to(xy)); // Check whether or not a new interaction has occurred. let new_interaction = match (self.enabled, maybe_mouse) { (false, _) | (true, None) => Interaction::Normal, (true, Some(mouse)) => { // Conrod does us a favor by transforming mouse.xy into this widget's // local coordinate system. Because mouse.xy is in local coords, // we must also pass the circle center in local coords. Thus we pass // [0.0, 0.0] as the center. // // See above where we define is_over_circ. let is_over = is_over_circ([0.0, 0.0], mouse.xy, dim); // See above where we define get_new_interaction. get_new_interaction(is_over, state.view().interaction, mouse) }, }; // If the mouse was released over the button, react. state.interaction is the // button's state as of a moment ago. new_interaction is the updated state as // of right now. So this if statement is saying: If the button was clicked a // moment ago, and it's now highlighted, then the button has been activated. if let (Interaction::Clicked, Interaction::Highlighted) = (state.view().interaction, new_interaction) { // Recall that our CircularButton struct includes maybe_react, which // stores either a reaction function or None. If maybe_react is Some, call // the function. if let Some(ref mut react) = self.maybe_react { react(); } } // Here we check to see whether or not our button should capture the mouse. // // Widgets can "capture" user input. If the button captures the mouse, then mouse // events will only be seen by the button. Other widgets will not see mouse events // until the button uncaptures the mouse. match (state.view().interaction, new_interaction) { // If the user has pressed the button we capture the mouse. (Interaction::Highlighted, Interaction::Clicked) => { ui.capture_mouse(); }, // If the user releases the button, we uncapture the mouse. (Interaction::
common
identifier_name
handlers.go
"ASC" var queues []string var status []string if len(queuesStr) > 0 { queues = strings.Split(queuesStr, ",") } if len(statusStr) > 0 { status = strings.Split(statusStr, ",") } page, err := strconv.Atoi(c.QueryParam("page")) if err != nil { // if err, set default page = 0 } foundJobs, err := s.Storage.Find(&jobs.Options{ Search: search, Limit: defaultQueryLimit, Offset: page * defaultQueryLimit, Queues: queues, SortBy: column, SortAsc: sort, Status: status, }) if err != nil { log.Error(err) c.JSON(http.StatusInternalServerError, err) return err } c.JSON(http.StatusOK, foundJobs) return nil } func (s *Server) GetStatus(c echo.Context) error { span := opentracing.StartSpan("API.GetStatus") defer span.Finish() query := c.Param("id") job, err := s.Storage.GetStatus(query) if err != nil { log.Error(err) c.JSON(http.StatusInternalServerError, err) return err } if job == nil { c.JSON(http.StatusNotFound, job) return nil } else { c.JSON(http.StatusOK, job) return nil } } // FindOne is a request handler, returns a job matching the query parameter 'q' func (s *Server) FindOne(c echo.Context) error { span := opentracing.StartSpan("API.FindOne") defer span.Finish() query := c.Param("id") job, err := s.Storage.FindOne(query) if err != nil { log.Error(err) c.JSON(http.StatusInternalServerError, err) return err } c.JSON(http.StatusOK, job) return nil } // KillMany is a request handler, kills a job matching the post parameter 'id' (AWS task ID) func (s *Server) KillMany(c echo.Context) error { span := opentracing.StartSpan("API.KillMany") defer span.Finish() obj, err := BodyToKillTask(c) if err != nil { c.JSON(http.StatusBadRequest, "{\"error\": \"Cannot deserialize\"}") } values := obj.IDs results := make(map[string]string) for _, value := range values { err := s.Killer.KillOne(value, "terminated from UI", s.Storage) if err != nil { results[value] = err.Error() } results[value] = "OK" } c.JSON(http.StatusOK, results) return nil } func (s *Server) FetchLogs(c echo.Context) error { span := opentracing.StartSpan("API.FetchLogs") defer span.Finish() const LOG_GROUP_NAME = "/aws/batch/job" format := c.QueryParam("format") if format != "text" { c.JSON(http.StatusBadRequest, "Only 'text' format is supported. Add format=text to your query.") return nil } c.Response().Header().Set(echo.HeaderContentType, echo.MIMETextPlain) id := c.Param("id") job, err := s.Storage.FindOne(id) if err != nil { log.Error(err) c.JSON(http.StatusInternalServerError, err) return err } svc := awsclients.CloudWatchLogs oldStyleLogs := func() (*string, error) { // AWS Batch seems to cap name strings to 50 characters for cloudwatch. truncated_job_name := job.Name if len(job.Name) > 50 { truncated_job_name = job.Name[:50] } s := truncated_job_name + "/" + id + "/" return &s, nil } newStyleLogs := func() (*string, error) { log_stream_name := job.LogStreamName if log_stream_name == nil || len(*log_stream_name) == 0 { return nil, nil } return log_stream_name, nil } logSources := [...]func() (*string, error){oldStyleLogs, newStyleLogs} var logStreams *cloudwatchlogs.DescribeLogStreamsOutput for _, log_source := range logSources { var name *string name, err = log_source() if err != nil { continue } // No error but no logs either if name == nil { continue } logStreams, err = svc.DescribeLogStreams(&cloudwatchlogs.DescribeLogStreamsInput{ LogGroupName: aws.String(LOG_GROUP_NAME), LogStreamNamePrefix: aws.String(*name), }) if err != nil || len(logStreams.LogStreams) <= 0 { continue } break } if err != nil { log.Error(err) c.String(http.StatusInternalServerError, err.Error()) return err } c.Response().WriteHeader(http.StatusOK) if logStreams == nil || len(logStreams.LogStreams) <= 0 { // Write empty logs. c.Response().Flush() return nil } startFromHead := true var previousToken *string var nextToken *string lines_pushed := 0 for { var logEvents *cloudwatchlogs.GetLogEventsOutput var err2 error previousToken = nextToken if nextToken != nil { logEvents, err2 = svc.GetLogEvents(&cloudwatchlogs.GetLogEventsInput{ LogGroupName: aws.String(LOG_GROUP_NAME), LogStreamName: logStreams.LogStreams[0].LogStreamName, StartFromHead: &startFromHead, NextToken: nextToken, }) } else { logEvents, err2 = svc.GetLogEvents(&cloudwatchlogs.GetLogEventsInput{ LogGroupName: aws.String(LOG_GROUP_NAME), LogStreamName: logStreams.LogStreams[0].LogStreamName, StartFromHead: &startFromHead, }) } if err2 != nil { return err2 } nextToken = logEvents.NextForwardToken events := logEvents.Events for _, event := range events { _, err2 = c.Response().Write([]byte(*event.Message + "\n")) if err2 != nil { return err2 } lines_pushed += 1 if lines_pushed >= 1000 { c.Response().Flush() } } if nextToken == nil || (previousToken != nil && *previousToken == *nextToken) { break } } return nil } // KillOne is a request handler, kills a job matching the post parameter 'id' (AWS task ID) func (s *Server) KillOne(c echo.Context) error { span := opentracing.StartSpan("API.KillOne") defer span.Finish() task := new(KillTaskID) if err := c.Bind(task); err != nil { return err } err := s.Killer.KillOne(task.ID, "terminated from UI", s.Storage) if err != nil { log.Error(err) c.JSON(http.StatusInternalServerError, err) return err } c.JSON(http.StatusOK, task.ID) return nil } func (s *Server) ListActiveJobQueues(c echo.Context) error { span := opentracing.StartSpan("API.ListActiveJobQueues") defer span.Finish() active_job_queues, err := s.Storage.ListActiveJobQueues() if err != nil { log.Error(err) c.JSON(http.StatusInternalServerError, err) return err } c.JSON(http.StatusOK, active_job_queues) return nil } func (s *Server) ListAllJobQueues(c echo.Context) error { span := opentracing.StartSpan("API.ListAllJobQueues") defer span.Finish() // This function gets *all* job queues, even those not registered to // Batchiepatchie. Therefore, we must ask AWS about all the job // queues. (as opposed to looking in our data store). svc := awsclients.Batch result := make([]string, 0) var next_token *string for { var input *batch.DescribeJobQueuesInput if next_token != nil { input = &batch.DescribeJobQueuesInput{NextToken: next_token} } else { input = &batch.DescribeJobQueuesInput{} } job_queues, err := svc.DescribeJobQueues(input) if err != nil { c.JSON(http.StatusInternalServerError, err) return err } for _, job_queue := range job_queues.JobQueues { name := job_queue.JobQueueName result = append(result, *name) } if input.NextToken != nil { next_token = input.NextToken } else { break } } c.JSON(http.StatusOK, result) return nil } func (s *Server) ActivateJobQueue(c echo.Context) error { span := opentracing.StartSpan("API.ActivateJobQueue") defer span.Finish() job_queue_name := c.Param("name") err := s.Storage.ActivateJobQueue(job_queue_name) if err != nil { c.JSON(http.StatusInternalServerError, err) return err } else { c.String(http.StatusOK, "[]") return nil } } func (s *Server)
DeactivateJobQueue
identifier_name
handlers.go
inish() c.QueryParams() search := c.QueryParam("q") queuesStr := c.QueryParam("queue") statusStr := c.QueryParam("status") column := c.QueryParam("sortColumn") sort := strings.ToUpper(c.QueryParam("sortDirection")) == "ASC" var queues []string var status []string if len(queuesStr) > 0 { queues = strings.Split(queuesStr, ",") } if len(statusStr) > 0 { status = strings.Split(statusStr, ",") } page, err := strconv.Atoi(c.QueryParam("page")) if err != nil { // if err, set default page = 0 } foundJobs, err := s.Storage.Find(&jobs.Options{ Search: search, Limit: defaultQueryLimit, Offset: page * defaultQueryLimit, Queues: queues, SortBy: column, SortAsc: sort, Status: status, }) if err != nil { log.Error(err) c.JSON(http.StatusInternalServerError, err) return err } c.JSON(http.StatusOK, foundJobs) return nil } func (s *Server) GetStatus(c echo.Context) error { span := opentracing.StartSpan("API.GetStatus") defer span.Finish() query := c.Param("id") job, err := s.Storage.GetStatus(query) if err != nil { log.Error(err) c.JSON(http.StatusInternalServerError, err) return err } if job == nil { c.JSON(http.StatusNotFound, job) return nil } else { c.JSON(http.StatusOK, job) return nil } } // FindOne is a request handler, returns a job matching the query parameter 'q' func (s *Server) FindOne(c echo.Context) error { span := opentracing.StartSpan("API.FindOne") defer span.Finish() query := c.Param("id") job, err := s.Storage.FindOne(query) if err != nil { log.Error(err) c.JSON(http.StatusInternalServerError, err) return err } c.JSON(http.StatusOK, job) return nil } // KillMany is a request handler, kills a job matching the post parameter 'id' (AWS task ID) func (s *Server) KillMany(c echo.Context) error { span := opentracing.StartSpan("API.KillMany") defer span.Finish() obj, err := BodyToKillTask(c) if err != nil { c.JSON(http.StatusBadRequest, "{\"error\": \"Cannot deserialize\"}") } values := obj.IDs results := make(map[string]string) for _, value := range values { err := s.Killer.KillOne(value, "terminated from UI", s.Storage) if err != nil { results[value] = err.Error() } results[value] = "OK" } c.JSON(http.StatusOK, results) return nil } func (s *Server) FetchLogs(c echo.Context) error { span := opentracing.StartSpan("API.FetchLogs") defer span.Finish() const LOG_GROUP_NAME = "/aws/batch/job" format := c.QueryParam("format") if format != "text" { c.JSON(http.StatusBadRequest, "Only 'text' format is supported. Add format=text to your query.") return nil } c.Response().Header().Set(echo.HeaderContentType, echo.MIMETextPlain) id := c.Param("id") job, err := s.Storage.FindOne(id) if err != nil { log.Error(err) c.JSON(http.StatusInternalServerError, err) return err } svc := awsclients.CloudWatchLogs oldStyleLogs := func() (*string, error) { // AWS Batch seems to cap name strings to 50 characters for cloudwatch. truncated_job_name := job.Name if len(job.Name) > 50 { truncated_job_name = job.Name[:50] } s := truncated_job_name + "/" + id + "/" return &s, nil } newStyleLogs := func() (*string, error) { log_stream_name := job.LogStreamName if log_stream_name == nil || len(*log_stream_name) == 0 { return nil, nil } return log_stream_name, nil } logSources := [...]func() (*string, error){oldStyleLogs, newStyleLogs} var logStreams *cloudwatchlogs.DescribeLogStreamsOutput for _, log_source := range logSources { var name *string name, err = log_source() if err != nil { continue } // No error but no logs either if name == nil { continue } logStreams, err = svc.DescribeLogStreams(&cloudwatchlogs.DescribeLogStreamsInput{ LogGroupName: aws.String(LOG_GROUP_NAME), LogStreamNamePrefix: aws.String(*name), }) if err != nil || len(logStreams.LogStreams) <= 0 { continue } break } if err != nil { log.Error(err) c.String(http.StatusInternalServerError, err.Error()) return err } c.Response().WriteHeader(http.StatusOK) if logStreams == nil || len(logStreams.LogStreams) <= 0 { // Write empty logs. c.Response().Flush() return nil } startFromHead := true var previousToken *string var nextToken *string lines_pushed := 0 for { var logEvents *cloudwatchlogs.GetLogEventsOutput var err2 error previousToken = nextToken if nextToken != nil { logEvents, err2 = svc.GetLogEvents(&cloudwatchlogs.GetLogEventsInput{ LogGroupName: aws.String(LOG_GROUP_NAME), LogStreamName: logStreams.LogStreams[0].LogStreamName, StartFromHead: &startFromHead, NextToken: nextToken, }) } else { logEvents, err2 = svc.GetLogEvents(&cloudwatchlogs.GetLogEventsInput{ LogGroupName: aws.String(LOG_GROUP_NAME), LogStreamName: logStreams.LogStreams[0].LogStreamName, StartFromHead: &startFromHead, }) } if err2 != nil { return err2 } nextToken = logEvents.NextForwardToken events := logEvents.Events for _, event := range events { _, err2 = c.Response().Write([]byte(*event.Message + "\n")) if err2 != nil { return err2 } lines_pushed += 1 if lines_pushed >= 1000 { c.Response().Flush() } } if nextToken == nil || (previousToken != nil && *previousToken == *nextToken) { break } } return nil } // KillOne is a request handler, kills a job matching the post parameter 'id' (AWS task ID) func (s *Server) KillOne(c echo.Context) error { span := opentracing.StartSpan("API.KillOne") defer span.Finish() task := new(KillTaskID) if err := c.Bind(task); err != nil { return err } err := s.Killer.KillOne(task.ID, "terminated from UI", s.Storage) if err != nil { log.Error(err) c.JSON(http.StatusInternalServerError, err) return err } c.JSON(http.StatusOK, task.ID) return nil } func (s *Server) ListActiveJobQueues(c echo.Context) error { span := opentracing.StartSpan("API.ListActiveJobQueues") defer span.Finish() active_job_queues, err := s.Storage.ListActiveJobQueues() if err != nil { log.Error(err) c.JSON(http.StatusInternalServerError, err) return err } c.JSON(http.StatusOK, active_job_queues) return nil } func (s *Server) ListAllJobQueues(c echo.Context) error
if err != nil { c.JSON(http.StatusInternalServerError, err) return err } for _, job_queue := range job_queues.JobQueues { name := job_queue.JobQueueName result = append(result, *name) } if input.NextToken != nil { next_token = input.NextToken } else { break } } c.JSON(http.StatusOK, result) return nil } func (s *Server) ActivateJobQueue(c echo.Context) error { span := opentracing.StartSpan("API.ActivateJobQueue") defer span.Finish() job_queue_name := c.Param("name") err := s.Storage.Act
{ span := opentracing.StartSpan("API.ListAllJobQueues") defer span.Finish() // This function gets *all* job queues, even those not registered to // Batchiepatchie. Therefore, we must ask AWS about all the job // queues. (as opposed to looking in our data store). svc := awsclients.Batch result := make([]string, 0) var next_token *string for { var input *batch.DescribeJobQueuesInput if next_token != nil { input = &batch.DescribeJobQueuesInput{NextToken: next_token} } else { input = &batch.DescribeJobQueuesInput{} } job_queues, err := svc.DescribeJobQueues(input)
identifier_body
handlers.go
inish() c.QueryParams() search := c.QueryParam("q") queuesStr := c.QueryParam("queue") statusStr := c.QueryParam("status") column := c.QueryParam("sortColumn") sort := strings.ToUpper(c.QueryParam("sortDirection")) == "ASC" var queues []string var status []string if len(queuesStr) > 0 { queues = strings.Split(queuesStr, ",") } if len(statusStr) > 0 { status = strings.Split(statusStr, ",") } page, err := strconv.Atoi(c.QueryParam("page")) if err != nil { // if err, set default page = 0 } foundJobs, err := s.Storage.Find(&jobs.Options{ Search: search, Limit: defaultQueryLimit, Offset: page * defaultQueryLimit, Queues: queues, SortBy: column, SortAsc: sort, Status: status, }) if err != nil { log.Error(err) c.JSON(http.StatusInternalServerError, err) return err } c.JSON(http.StatusOK, foundJobs) return nil } func (s *Server) GetStatus(c echo.Context) error { span := opentracing.StartSpan("API.GetStatus") defer span.Finish() query := c.Param("id") job, err := s.Storage.GetStatus(query) if err != nil { log.Error(err) c.JSON(http.StatusInternalServerError, err) return err } if job == nil { c.JSON(http.StatusNotFound, job) return nil } else { c.JSON(http.StatusOK, job) return nil } } // FindOne is a request handler, returns a job matching the query parameter 'q' func (s *Server) FindOne(c echo.Context) error { span := opentracing.StartSpan("API.FindOne") defer span.Finish() query := c.Param("id") job, err := s.Storage.FindOne(query) if err != nil { log.Error(err) c.JSON(http.StatusInternalServerError, err) return err } c.JSON(http.StatusOK, job) return nil } // KillMany is a request handler, kills a job matching the post parameter 'id' (AWS task ID) func (s *Server) KillMany(c echo.Context) error { span := opentracing.StartSpan("API.KillMany") defer span.Finish() obj, err := BodyToKillTask(c) if err != nil { c.JSON(http.StatusBadRequest, "{\"error\": \"Cannot deserialize\"}") } values := obj.IDs results := make(map[string]string) for _, value := range values { err := s.Killer.KillOne(value, "terminated from UI", s.Storage) if err != nil { results[value] = err.Error() } results[value] = "OK" } c.JSON(http.StatusOK, results) return nil } func (s *Server) FetchLogs(c echo.Context) error { span := opentracing.StartSpan("API.FetchLogs") defer span.Finish() const LOG_GROUP_NAME = "/aws/batch/job" format := c.QueryParam("format") if format != "text" { c.JSON(http.StatusBadRequest, "Only 'text' format is supported. Add format=text to your query.") return nil } c.Response().Header().Set(echo.HeaderContentType, echo.MIMETextPlain) id := c.Param("id") job, err := s.Storage.FindOne(id) if err != nil { log.Error(err) c.JSON(http.StatusInternalServerError, err) return err } svc := awsclients.CloudWatchLogs oldStyleLogs := func() (*string, error) { // AWS Batch seems to cap name strings to 50 characters for cloudwatch. truncated_job_name := job.Name if len(job.Name) > 50 { truncated_job_name = job.Name[:50] } s := truncated_job_name + "/" + id + "/" return &s, nil } newStyleLogs := func() (*string, error) { log_stream_name := job.LogStreamName if log_stream_name == nil || len(*log_stream_name) == 0 { return nil, nil } return log_stream_name, nil } logSources := [...]func() (*string, error){oldStyleLogs, newStyleLogs} var logStreams *cloudwatchlogs.DescribeLogStreamsOutput for _, log_source := range logSources { var name *string name, err = log_source() if err != nil { continue } // No error but no logs either if name == nil { continue } logStreams, err = svc.DescribeLogStreams(&cloudwatchlogs.DescribeLogStreamsInput{ LogGroupName: aws.String(LOG_GROUP_NAME), LogStreamNamePrefix: aws.String(*name), }) if err != nil || len(logStreams.LogStreams) <= 0 { continue } break } if err != nil { log.Error(err) c.String(http.StatusInternalServerError, err.Error()) return err } c.Response().WriteHeader(http.StatusOK) if logStreams == nil || len(logStreams.LogStreams) <= 0 { // Write empty logs. c.Response().Flush() return nil } startFromHead := true var previousToken *string var nextToken *string lines_pushed := 0 for { var logEvents *cloudwatchlogs.GetLogEventsOutput var err2 error previousToken = nextToken if nextToken != nil { logEvents, err2 = svc.GetLogEvents(&cloudwatchlogs.GetLogEventsInput{ LogGroupName: aws.String(LOG_GROUP_NAME), LogStreamName: logStreams.LogStreams[0].LogStreamName, StartFromHead: &startFromHead, NextToken: nextToken, }) } else { logEvents, err2 = svc.GetLogEvents(&cloudwatchlogs.GetLogEventsInput{ LogGroupName: aws.String(LOG_GROUP_NAME), LogStreamName: logStreams.LogStreams[0].LogStreamName, StartFromHead: &startFromHead, }) } if err2 != nil { return err2 } nextToken = logEvents.NextForwardToken events := logEvents.Events for _, event := range events { _, err2 = c.Response().Write([]byte(*event.Message + "\n")) if err2 != nil { return err2 } lines_pushed += 1 if lines_pushed >= 1000 { c.Response().Flush() } } if nextToken == nil || (previousToken != nil && *previousToken == *nextToken) { break } } return nil } // KillOne is a request handler, kills a job matching the post parameter 'id' (AWS task ID) func (s *Server) KillOne(c echo.Context) error { span := opentracing.StartSpan("API.KillOne") defer span.Finish() task := new(KillTaskID) if err := c.Bind(task); err != nil { return err } err := s.Killer.KillOne(task.ID, "terminated from UI", s.Storage) if err != nil { log.Error(err) c.JSON(http.StatusInternalServerError, err) return err } c.JSON(http.StatusOK, task.ID) return nil } func (s *Server) ListActiveJobQueues(c echo.Context) error { span := opentracing.StartSpan("API.ListActiveJobQueues") defer span.Finish() active_job_queues, err := s.Storage.ListActiveJobQueues() if err != nil { log.Error(err) c.JSON(http.StatusInternalServerError, err) return err } c.JSON(http.StatusOK, active_job_queues) return nil } func (s *Server) ListAllJobQueues(c echo.Context) error { span := opentracing.StartSpan("API.ListAllJobQueues") defer span.Finish() // This function gets *all* job queues, even those not registered to // Batchiepatchie. Therefore, we must ask AWS about all the job // queues. (as opposed to looking in our data store). svc := awsclients.Batch result := make([]string, 0) var next_token *string for { var input *batch.DescribeJobQueuesInput if next_token != nil { input = &batch.DescribeJobQueuesInput{NextToken: next_token} } else { input = &batch.DescribeJobQueuesInput{} } job_queues, err := svc.DescribeJobQueues(input) if err != nil { c.JSON(http.StatusInternalServerError, err) return err } for _, job_queue := range job_queues.JobQueues { name := job_queue.JobQueueName result = append(result, *name) } if input.NextToken != nil { next_token = input.NextToken } else { break } } c.JSON(http.StatusOK, result) return nil
span := opentracing.StartSpan("API.ActivateJobQueue") defer span.Finish() job_queue_name := c.Param("name") err := s.Storage.ActivateJob
} func (s *Server) ActivateJobQueue(c echo.Context) error {
random_line_split
handlers.go
() c.QueryParams() search := c.QueryParam("q") queuesStr := c.QueryParam("queue") statusStr := c.QueryParam("status") column := c.QueryParam("sortColumn") sort := strings.ToUpper(c.QueryParam("sortDirection")) == "ASC" var queues []string var status []string if len(queuesStr) > 0 { queues = strings.Split(queuesStr, ",") } if len(statusStr) > 0 { status = strings.Split(statusStr, ",") } page, err := strconv.Atoi(c.QueryParam("page")) if err != nil { // if err, set default page = 0 } foundJobs, err := s.Storage.Find(&jobs.Options{ Search: search, Limit: defaultQueryLimit, Offset: page * defaultQueryLimit, Queues: queues, SortBy: column, SortAsc: sort, Status: status, }) if err != nil { log.Error(err) c.JSON(http.StatusInternalServerError, err) return err } c.JSON(http.StatusOK, foundJobs) return nil } func (s *Server) GetStatus(c echo.Context) error { span := opentracing.StartSpan("API.GetStatus") defer span.Finish() query := c.Param("id") job, err := s.Storage.GetStatus(query) if err != nil { log.Error(err) c.JSON(http.StatusInternalServerError, err) return err } if job == nil { c.JSON(http.StatusNotFound, job) return nil } else { c.JSON(http.StatusOK, job) return nil } } // FindOne is a request handler, returns a job matching the query parameter 'q' func (s *Server) FindOne(c echo.Context) error { span := opentracing.StartSpan("API.FindOne") defer span.Finish() query := c.Param("id") job, err := s.Storage.FindOne(query) if err != nil { log.Error(err) c.JSON(http.StatusInternalServerError, err) return err } c.JSON(http.StatusOK, job) return nil } // KillMany is a request handler, kills a job matching the post parameter 'id' (AWS task ID) func (s *Server) KillMany(c echo.Context) error { span := opentracing.StartSpan("API.KillMany") defer span.Finish() obj, err := BodyToKillTask(c) if err != nil { c.JSON(http.StatusBadRequest, "{\"error\": \"Cannot deserialize\"}") } values := obj.IDs results := make(map[string]string) for _, value := range values { err := s.Killer.KillOne(value, "terminated from UI", s.Storage) if err != nil { results[value] = err.Error() } results[value] = "OK" } c.JSON(http.StatusOK, results) return nil } func (s *Server) FetchLogs(c echo.Context) error { span := opentracing.StartSpan("API.FetchLogs") defer span.Finish() const LOG_GROUP_NAME = "/aws/batch/job" format := c.QueryParam("format") if format != "text" { c.JSON(http.StatusBadRequest, "Only 'text' format is supported. Add format=text to your query.") return nil } c.Response().Header().Set(echo.HeaderContentType, echo.MIMETextPlain) id := c.Param("id") job, err := s.Storage.FindOne(id) if err != nil { log.Error(err) c.JSON(http.StatusInternalServerError, err) return err } svc := awsclients.CloudWatchLogs oldStyleLogs := func() (*string, error) { // AWS Batch seems to cap name strings to 50 characters for cloudwatch. truncated_job_name := job.Name if len(job.Name) > 50 { truncated_job_name = job.Name[:50] } s := truncated_job_name + "/" + id + "/" return &s, nil } newStyleLogs := func() (*string, error) { log_stream_name := job.LogStreamName if log_stream_name == nil || len(*log_stream_name) == 0 { return nil, nil } return log_stream_name, nil } logSources := [...]func() (*string, error){oldStyleLogs, newStyleLogs} var logStreams *cloudwatchlogs.DescribeLogStreamsOutput for _, log_source := range logSources { var name *string name, err = log_source() if err != nil { continue } // No error but no logs either if name == nil { continue } logStreams, err = svc.DescribeLogStreams(&cloudwatchlogs.DescribeLogStreamsInput{ LogGroupName: aws.String(LOG_GROUP_NAME), LogStreamNamePrefix: aws.String(*name), }) if err != nil || len(logStreams.LogStreams) <= 0 { continue } break } if err != nil { log.Error(err) c.String(http.StatusInternalServerError, err.Error()) return err } c.Response().WriteHeader(http.StatusOK) if logStreams == nil || len(logStreams.LogStreams) <= 0 { // Write empty logs. c.Response().Flush() return nil } startFromHead := true var previousToken *string var nextToken *string lines_pushed := 0 for { var logEvents *cloudwatchlogs.GetLogEventsOutput var err2 error previousToken = nextToken if nextToken != nil { logEvents, err2 = svc.GetLogEvents(&cloudwatchlogs.GetLogEventsInput{ LogGroupName: aws.String(LOG_GROUP_NAME), LogStreamName: logStreams.LogStreams[0].LogStreamName, StartFromHead: &startFromHead, NextToken: nextToken, }) } else { logEvents, err2 = svc.GetLogEvents(&cloudwatchlogs.GetLogEventsInput{ LogGroupName: aws.String(LOG_GROUP_NAME), LogStreamName: logStreams.LogStreams[0].LogStreamName, StartFromHead: &startFromHead, }) } if err2 != nil { return err2 } nextToken = logEvents.NextForwardToken events := logEvents.Events for _, event := range events { _, err2 = c.Response().Write([]byte(*event.Message + "\n")) if err2 != nil { return err2 } lines_pushed += 1 if lines_pushed >= 1000 { c.Response().Flush() } } if nextToken == nil || (previousToken != nil && *previousToken == *nextToken) { break } } return nil } // KillOne is a request handler, kills a job matching the post parameter 'id' (AWS task ID) func (s *Server) KillOne(c echo.Context) error { span := opentracing.StartSpan("API.KillOne") defer span.Finish() task := new(KillTaskID) if err := c.Bind(task); err != nil { return err } err := s.Killer.KillOne(task.ID, "terminated from UI", s.Storage) if err != nil
c.JSON(http.StatusOK, task.ID) return nil } func (s *Server) ListActiveJobQueues(c echo.Context) error { span := opentracing.StartSpan("API.ListActiveJobQueues") defer span.Finish() active_job_queues, err := s.Storage.ListActiveJobQueues() if err != nil { log.Error(err) c.JSON(http.StatusInternalServerError, err) return err } c.JSON(http.StatusOK, active_job_queues) return nil } func (s *Server) ListAllJobQueues(c echo.Context) error { span := opentracing.StartSpan("API.ListAllJobQueues") defer span.Finish() // This function gets *all* job queues, even those not registered to // Batchiepatchie. Therefore, we must ask AWS about all the job // queues. (as opposed to looking in our data store). svc := awsclients.Batch result := make([]string, 0) var next_token *string for { var input *batch.DescribeJobQueuesInput if next_token != nil { input = &batch.DescribeJobQueuesInput{NextToken: next_token} } else { input = &batch.DescribeJobQueuesInput{} } job_queues, err := svc.DescribeJobQueues(input) if err != nil { c.JSON(http.StatusInternalServerError, err) return err } for _, job_queue := range job_queues.JobQueues { name := job_queue.JobQueueName result = append(result, *name) } if input.NextToken != nil { next_token = input.NextToken } else { break } } c.JSON(http.StatusOK, result) return nil } func (s *Server) ActivateJobQueue(c echo.Context) error { span := opentracing.StartSpan("API.ActivateJobQueue") defer span.Finish() job_queue_name := c.Param("name") err := s.Storage.Act
{ log.Error(err) c.JSON(http.StatusInternalServerError, err) return err }
conditional_block
wire.rs
: ToDname + ?Sized>( &mut self, name: &N, ) -> Result<(), Self::AppendError> { name.compose(self) } fn can_compress(&self) -> bool { false } } #[cfg(feature = "std")] impl Composer for std::vec::Vec<u8> {} impl<const N: usize> Composer for octseq::array::Array<N> {} #[cfg(feature = "bytes")] impl Composer for bytes::BytesMut {} #[cfg(feature = "smallvec")] impl<A: smallvec::Array<Item = u8>> Composer for smallvec::SmallVec<A> {} //------------ Compose ------------------------------------------------------- /// An extension trait to add composing to foreign types. /// /// This trait can be used to add the `compose` method to a foreign type. For /// local types, the method should be added directly to the type instead. /// /// The trait can only be used for types that have a fixed-size wire /// representation. pub trait Compose { /// The length in octets of the wire representation of a value. /// /// Because all wire format lengths are limited to 16 bit, this is a /// `u16` rather than a `usize`. const COMPOSE_LEN: u16 = 0; /// Appends the wire format representation of the value to the target. fn compose<Target: OctetsBuilder + ?Sized>( &self, target: &mut Target, ) -> Result<(), Target::AppendError>; } impl<'a, T: Compose + ?Sized> Compose for &'a T { const COMPOSE_LEN: u16 = T::COMPOSE_LEN; fn compose<Target: OctetsBuilder + ?Sized>( &self, target: &mut Target, ) -> Result<(), Target::AppendError> { (*self).compose(target) } } impl Compose for i8 { const COMPOSE_LEN: u16 = 1; fn compose<Target: OctetsBuilder + ?Sized>( &self, target: &mut Target, ) -> Result<(), Target::AppendError> { target.append_slice(&[*self as u8]) } } impl Compose for u8 { const COMPOSE_LEN: u16 = 1; fn compose<Target: OctetsBuilder + ?Sized>( &self, target: &mut Target, ) -> Result<(), Target::AppendError> { target.append_slice(&[*self]) } } macro_rules! compose_to_be_bytes { ( $type:ident ) => { impl Compose for $type { const COMPOSE_LEN: u16 = ($type::BITS >> 3) as u16; fn compose<Target: OctetsBuilder + ?Sized>( &self, target: &mut Target, ) -> Result<(), Target::AppendError> { target.append_slice(&self.to_be_bytes()) } } }; } compose_to_be_bytes!(i16); compose_to_be_bytes!(u16); compose_to_be_bytes!(i32); compose_to_be_bytes!(u32); compose_to_be_bytes!(i64); compose_to_be_bytes!(u64); compose_to_be_bytes!(i128); compose_to_be_bytes!(u128); impl Compose for Ipv4Addr { const COMPOSE_LEN: u16 = 4; fn compose<Target: OctetsBuilder + ?Sized>( &self, target: &mut Target, ) -> Result<(), Target::AppendError> { target.append_slice(&self.octets()) } } impl Compose for Ipv6Addr { const COMPOSE_LEN: u16 = 16; fn compose<Target: OctetsBuilder + ?Sized>( &self, target: &mut Target, ) -> Result<(), Target::AppendError> { target.append_slice(&self.octets()) } } // No impl for [u8; const N: usize] because we can’t guarantee a correct // COMPOSE_LEN -- it may be longer than a u16 can hold. //------------ Parse ------------------------------------------------------ /// An extension trait to add parsing to foreign types. /// /// This trait can be used to add the `parse` method to a foreign type. For /// local types, the method should be added directly to the type instead. pub trait Parse<'a, Octs: ?Sized>: Sized { /// Extracts a value from the beginning of `parser`. /// /// If parsing fails and an error is returned, the parser’s position /// should be considered to be undefined. If it is supposed to be reused /// in this case, you should store the position before attempting to parse /// and seek to that position again before continuing. fn parse(parser: &mut Parser<'a, Octs>) -> Result<Self, ParseError>; } impl<'a, Octs: AsRef<[u8]> + ?Sized> Parse<'a, Octs> for i8 { fn parse(parser: &mut Parser<'a, Octs>) -> Result<Self, ParseError> { parser.parse_i8().map_err(Into::into) } } impl<'a, Octs: AsRef<[u8]> + ?Sized> Parse<'a, Octs> for u8 { fn parse(parser: &mut Parser<'a, Octs>) -> Result<Self, ParseError> { parser.parse_u8().map_err(Into::into) } } impl<'a, Octs: AsRef<[u8]> + ?Sized> Parse<'a, Octs> for i16 { fn parse(parser: &mut Parser<'a, Octs>) -> Result<Self, ParseError> { parser.parse_i16_be().map_err(Into::into) } } impl<'a, Octs: AsRef<[u8]> + ?Sized> Parse<'a, Octs> for u16 { fn parse(parser: &mut Parser<'a, Octs>) -> Result<Self, ParseError> {
pl<'a, Octs: AsRef<[u8]> + ?Sized> Parse<'a, Octs> for i32 { fn parse(parser: &mut Parser<'a, Octs>) -> Result<Self, ParseError> { parser.parse_i32_be().map_err(Into::into) } } impl<'a, Octs: AsRef<[u8]> + ?Sized> Parse<'a, Octs> for u32 { fn parse(parser: &mut Parser<'a, Octs>) -> Result<Self, ParseError> { parser.parse_u32_be().map_err(Into::into) } } impl<'a, Octs: AsRef<[u8]> + ?Sized> Parse<'a, Octs> for u64 { fn parse(parser: &mut Parser<'a, Octs>) -> Result<Self, ParseError> { parser.parse_u64_be().map_err(Into::into) } } impl<'a, Octs: AsRef<[u8]> + ?Sized> Parse<'a, Octs> for i64 { fn parse(parser: &mut Parser<'a, Octs>) -> Result<Self, ParseError> { parser.parse_i64_be().map_err(Into::into) } } impl<'a, Octs: AsRef<[u8]> + ?Sized> Parse<'a, Octs> for Ipv4Addr { fn parse(parser: &mut Parser<'a, Octs>) -> Result<Self, ParseError> { Ok(Self::new( u8::parse(parser)?, u8::parse(parser)?, u8::parse(parser)?, u8::parse(parser)?, )) } } impl<'a, Octs: AsRef<[u8]> + ?Sized> Parse<'a, Octs> for Ipv6Addr { fn parse(parser: &mut Parser<'a, Octs>) -> Result<Self, ParseError> { let mut buf = [0u8; 16]; parser.parse_buf(&mut buf)?; Ok(buf.into()) } } impl<'a, Octs: AsRef<[u8]> + ?Sized, const N: usize> Parse<'a, Octs> for [u8; N] { fn parse(parser: &mut Parser<'a, Octs>) -> Result<Self, ParseError> { let mut res = [0u8; N]; parser.parse_buf(&mut res)?; Ok(res) } } //============ Helpful Function ============================================== /// Parses something from a `Vec<u8>`. /// /// The actual parsing happens in the provided closure. Returns an error if /// the closure returns an error or if there is unparsed data left over after /// the closure returns. Otherwise returns whatever the closure returned. #[cfg(feature = "std")] pub fn parse_slice<F, T>(data: &[u8], op: F) -> Result<T, ParseError> where F: FnOnce(&mut Parser<[u8]>) -> Result<T, ParseError>, { let mut parser = Parser::from_ref(data); let res = op(&mut parser)?; if parser.remaining() > 0 { Err(ParseError::form_error("trailing data"))
parser.parse_u16_be().map_err(Into::into) } } im
identifier_body
wire.rs
, ) -> Result<(), Target::AppendError>; } impl<'a, T: Compose + ?Sized> Compose for &'a T { const COMPOSE_LEN: u16 = T::COMPOSE_LEN; fn compose<Target: OctetsBuilder + ?Sized>( &self, target: &mut Target, ) -> Result<(), Target::AppendError> { (*self).compose(target) } } impl Compose for i8 { const COMPOSE_LEN: u16 = 1; fn compose<Target: OctetsBuilder + ?Sized>( &self, target: &mut Target, ) -> Result<(), Target::AppendError> { target.append_slice(&[*self as u8]) } } impl Compose for u8 { const COMPOSE_LEN: u16 = 1; fn compose<Target: OctetsBuilder + ?Sized>( &self, target: &mut Target, ) -> Result<(), Target::AppendError> { target.append_slice(&[*self]) } } macro_rules! compose_to_be_bytes { ( $type:ident ) => { impl Compose for $type { const COMPOSE_LEN: u16 = ($type::BITS >> 3) as u16; fn compose<Target: OctetsBuilder + ?Sized>( &self, target: &mut Target, ) -> Result<(), Target::AppendError> { target.append_slice(&self.to_be_bytes()) } } }; } compose_to_be_bytes!(i16); compose_to_be_bytes!(u16); compose_to_be_bytes!(i32); compose_to_be_bytes!(u32); compose_to_be_bytes!(i64); compose_to_be_bytes!(u64); compose_to_be_bytes!(i128); compose_to_be_bytes!(u128); impl Compose for Ipv4Addr { const COMPOSE_LEN: u16 = 4; fn compose<Target: OctetsBuilder + ?Sized>( &self, target: &mut Target, ) -> Result<(), Target::AppendError> { target.append_slice(&self.octets()) } } impl Compose for Ipv6Addr { const COMPOSE_LEN: u16 = 16; fn compose<Target: OctetsBuilder + ?Sized>( &self, target: &mut Target, ) -> Result<(), Target::AppendError> { target.append_slice(&self.octets()) } } // No impl for [u8; const N: usize] because we can’t guarantee a correct // COMPOSE_LEN -- it may be longer than a u16 can hold. //------------ Parse ------------------------------------------------------ /// An extension trait to add parsing to foreign types. /// /// This trait can be used to add the `parse` method to a foreign type. For /// local types, the method should be added directly to the type instead. pub trait Parse<'a, Octs: ?Sized>: Sized { /// Extracts a value from the beginning of `parser`. /// /// If parsing fails and an error is returned, the parser’s position /// should be considered to be undefined. If it is supposed to be reused /// in this case, you should store the position before attempting to parse /// and seek to that position again before continuing. fn parse(parser: &mut Parser<'a, Octs>) -> Result<Self, ParseError>; } impl<'a, Octs: AsRef<[u8]> + ?Sized> Parse<'a, Octs> for i8 { fn parse(parser: &mut Parser<'a, Octs>) -> Result<Self, ParseError> { parser.parse_i8().map_err(Into::into) } } impl<'a, Octs: AsRef<[u8]> + ?Sized> Parse<'a, Octs> for u8 { fn parse(parser: &mut Parser<'a, Octs>) -> Result<Self, ParseError> { parser.parse_u8().map_err(Into::into) } } impl<'a, Octs: AsRef<[u8]> + ?Sized> Parse<'a, Octs> for i16 { fn parse(parser: &mut Parser<'a, Octs>) -> Result<Self, ParseError> { parser.parse_i16_be().map_err(Into::into) } } impl<'a, Octs: AsRef<[u8]> + ?Sized> Parse<'a, Octs> for u16 { fn parse(parser: &mut Parser<'a, Octs>) -> Result<Self, ParseError> { parser.parse_u16_be().map_err(Into::into) } } impl<'a, Octs: AsRef<[u8]> + ?Sized> Parse<'a, Octs> for i32 { fn parse(parser: &mut Parser<'a, Octs>) -> Result<Self, ParseError> { parser.parse_i32_be().map_err(Into::into) } } impl<'a, Octs: AsRef<[u8]> + ?Sized> Parse<'a, Octs> for u32 { fn parse(parser: &mut Parser<'a, Octs>) -> Result<Self, ParseError> { parser.parse_u32_be().map_err(Into::into) } } impl<'a, Octs: AsRef<[u8]> + ?Sized> Parse<'a, Octs> for u64 { fn parse(parser: &mut Parser<'a, Octs>) -> Result<Self, ParseError> { parser.parse_u64_be().map_err(Into::into) } } impl<'a, Octs: AsRef<[u8]> + ?Sized> Parse<'a, Octs> for i64 { fn parse(parser: &mut Parser<'a, Octs>) -> Result<Self, ParseError> { parser.parse_i64_be().map_err(Into::into) } } impl<'a, Octs: AsRef<[u8]> + ?Sized> Parse<'a, Octs> for Ipv4Addr { fn parse(parser: &mut Parser<'a, Octs>) -> Result<Self, ParseError> { Ok(Self::new( u8::parse(parser)?, u8::parse(parser)?, u8::parse(parser)?, u8::parse(parser)?, )) } } impl<'a, Octs: AsRef<[u8]> + ?Sized> Parse<'a, Octs> for Ipv6Addr { fn parse(parser: &mut Parser<'a, Octs>) -> Result<Self, ParseError> { let mut buf = [0u8; 16]; parser.parse_buf(&mut buf)?; Ok(buf.into()) } } impl<'a, Octs: AsRef<[u8]> + ?Sized, const N: usize> Parse<'a, Octs> for [u8; N] { fn parse(parser: &mut Parser<'a, Octs>) -> Result<Self, ParseError> { let mut res = [0u8; N]; parser.parse_buf(&mut res)?; Ok(res) } } //============ Helpful Function ============================================== /// Parses something from a `Vec<u8>`. /// /// The actual parsing happens in the provided closure. Returns an error if /// the closure returns an error or if there is unparsed data left over after /// the closure returns. Otherwise returns whatever the closure returned. #[cfg(feature = "std")] pub fn parse_slice<F, T>(data: &[u8], op: F) -> Result<T, ParseError> where F: FnOnce(&mut Parser<[u8]>) -> Result<T, ParseError>, { let mut parser = Parser::from_ref(data); let res = op(&mut parser)?; if parser.remaining() > 0 { Err(ParseError::form_error("trailing data")) } else { Ok(res) } } /// Composes something into a `Vec<u8>`. /// /// The actual composing happens in the provided closure. /// This function is mostly useful in testing so you can construct this vec /// directly inside an asserting. #[cfg(feature = "std")] pub fn compose_vec( op: impl FnOnce( &mut std::vec::Vec<u8>, ) -> Result<(), core::convert::Infallible>, ) -> std::vec::Vec<u8> { let mut res = std::vec::Vec::new(); octseq::builder::infallible(op(&mut res)); res } //============ Error Types =================================================== //------------ ParseError ---------------------------------------------------- /// An error happened while parsing data. #[derive(Clone, Copy, Debug, Eq, PartialEq)] pub enum ParseError { /// An attempt was made to go beyond the end of the parser. ShortInput, /// A formatting error occurred. Form(FormError), } impl ParseError { /// Creates a new parse error as a form error with the given message. pub fn form_error(msg: &'static str) -> Self { FormError::new(msg).into() } } //--- From impl From<ShortInput> for ParseError { fn from(_: ShortInput) -> Self { ParseError::ShortInput } }
impl From<FormError> for ParseError { fn from(err: FormError) -> Self { ParseError::Form(err)
random_line_split
wire.rs
: ToDname + ?Sized>( &mut self, name: &N, ) -> Result<(), Self::AppendError> { name.compose(self) } fn ca
self) -> bool { false } } #[cfg(feature = "std")] impl Composer for std::vec::Vec<u8> {} impl<const N: usize> Composer for octseq::array::Array<N> {} #[cfg(feature = "bytes")] impl Composer for bytes::BytesMut {} #[cfg(feature = "smallvec")] impl<A: smallvec::Array<Item = u8>> Composer for smallvec::SmallVec<A> {} //------------ Compose ------------------------------------------------------- /// An extension trait to add composing to foreign types. /// /// This trait can be used to add the `compose` method to a foreign type. For /// local types, the method should be added directly to the type instead. /// /// The trait can only be used for types that have a fixed-size wire /// representation. pub trait Compose { /// The length in octets of the wire representation of a value. /// /// Because all wire format lengths are limited to 16 bit, this is a /// `u16` rather than a `usize`. const COMPOSE_LEN: u16 = 0; /// Appends the wire format representation of the value to the target. fn compose<Target: OctetsBuilder + ?Sized>( &self, target: &mut Target, ) -> Result<(), Target::AppendError>; } impl<'a, T: Compose + ?Sized> Compose for &'a T { const COMPOSE_LEN: u16 = T::COMPOSE_LEN; fn compose<Target: OctetsBuilder + ?Sized>( &self, target: &mut Target, ) -> Result<(), Target::AppendError> { (*self).compose(target) } } impl Compose for i8 { const COMPOSE_LEN: u16 = 1; fn compose<Target: OctetsBuilder + ?Sized>( &self, target: &mut Target, ) -> Result<(), Target::AppendError> { target.append_slice(&[*self as u8]) } } impl Compose for u8 { const COMPOSE_LEN: u16 = 1; fn compose<Target: OctetsBuilder + ?Sized>( &self, target: &mut Target, ) -> Result<(), Target::AppendError> { target.append_slice(&[*self]) } } macro_rules! compose_to_be_bytes { ( $type:ident ) => { impl Compose for $type { const COMPOSE_LEN: u16 = ($type::BITS >> 3) as u16; fn compose<Target: OctetsBuilder + ?Sized>( &self, target: &mut Target, ) -> Result<(), Target::AppendError> { target.append_slice(&self.to_be_bytes()) } } }; } compose_to_be_bytes!(i16); compose_to_be_bytes!(u16); compose_to_be_bytes!(i32); compose_to_be_bytes!(u32); compose_to_be_bytes!(i64); compose_to_be_bytes!(u64); compose_to_be_bytes!(i128); compose_to_be_bytes!(u128); impl Compose for Ipv4Addr { const COMPOSE_LEN: u16 = 4; fn compose<Target: OctetsBuilder + ?Sized>( &self, target: &mut Target, ) -> Result<(), Target::AppendError> { target.append_slice(&self.octets()) } } impl Compose for Ipv6Addr { const COMPOSE_LEN: u16 = 16; fn compose<Target: OctetsBuilder + ?Sized>( &self, target: &mut Target, ) -> Result<(), Target::AppendError> { target.append_slice(&self.octets()) } } // No impl for [u8; const N: usize] because we can’t guarantee a correct // COMPOSE_LEN -- it may be longer than a u16 can hold. //------------ Parse ------------------------------------------------------ /// An extension trait to add parsing to foreign types. /// /// This trait can be used to add the `parse` method to a foreign type. For /// local types, the method should be added directly to the type instead. pub trait Parse<'a, Octs: ?Sized>: Sized { /// Extracts a value from the beginning of `parser`. /// /// If parsing fails and an error is returned, the parser’s position /// should be considered to be undefined. If it is supposed to be reused /// in this case, you should store the position before attempting to parse /// and seek to that position again before continuing. fn parse(parser: &mut Parser<'a, Octs>) -> Result<Self, ParseError>; } impl<'a, Octs: AsRef<[u8]> + ?Sized> Parse<'a, Octs> for i8 { fn parse(parser: &mut Parser<'a, Octs>) -> Result<Self, ParseError> { parser.parse_i8().map_err(Into::into) } } impl<'a, Octs: AsRef<[u8]> + ?Sized> Parse<'a, Octs> for u8 { fn parse(parser: &mut Parser<'a, Octs>) -> Result<Self, ParseError> { parser.parse_u8().map_err(Into::into) } } impl<'a, Octs: AsRef<[u8]> + ?Sized> Parse<'a, Octs> for i16 { fn parse(parser: &mut Parser<'a, Octs>) -> Result<Self, ParseError> { parser.parse_i16_be().map_err(Into::into) } } impl<'a, Octs: AsRef<[u8]> + ?Sized> Parse<'a, Octs> for u16 { fn parse(parser: &mut Parser<'a, Octs>) -> Result<Self, ParseError> { parser.parse_u16_be().map_err(Into::into) } } impl<'a, Octs: AsRef<[u8]> + ?Sized> Parse<'a, Octs> for i32 { fn parse(parser: &mut Parser<'a, Octs>) -> Result<Self, ParseError> { parser.parse_i32_be().map_err(Into::into) } } impl<'a, Octs: AsRef<[u8]> + ?Sized> Parse<'a, Octs> for u32 { fn parse(parser: &mut Parser<'a, Octs>) -> Result<Self, ParseError> { parser.parse_u32_be().map_err(Into::into) } } impl<'a, Octs: AsRef<[u8]> + ?Sized> Parse<'a, Octs> for u64 { fn parse(parser: &mut Parser<'a, Octs>) -> Result<Self, ParseError> { parser.parse_u64_be().map_err(Into::into) } } impl<'a, Octs: AsRef<[u8]> + ?Sized> Parse<'a, Octs> for i64 { fn parse(parser: &mut Parser<'a, Octs>) -> Result<Self, ParseError> { parser.parse_i64_be().map_err(Into::into) } } impl<'a, Octs: AsRef<[u8]> + ?Sized> Parse<'a, Octs> for Ipv4Addr { fn parse(parser: &mut Parser<'a, Octs>) -> Result<Self, ParseError> { Ok(Self::new( u8::parse(parser)?, u8::parse(parser)?, u8::parse(parser)?, u8::parse(parser)?, )) } } impl<'a, Octs: AsRef<[u8]> + ?Sized> Parse<'a, Octs> for Ipv6Addr { fn parse(parser: &mut Parser<'a, Octs>) -> Result<Self, ParseError> { let mut buf = [0u8; 16]; parser.parse_buf(&mut buf)?; Ok(buf.into()) } } impl<'a, Octs: AsRef<[u8]> + ?Sized, const N: usize> Parse<'a, Octs> for [u8; N] { fn parse(parser: &mut Parser<'a, Octs>) -> Result<Self, ParseError> { let mut res = [0u8; N]; parser.parse_buf(&mut res)?; Ok(res) } } //============ Helpful Function ============================================== /// Parses something from a `Vec<u8>`. /// /// The actual parsing happens in the provided closure. Returns an error if /// the closure returns an error or if there is unparsed data left over after /// the closure returns. Otherwise returns whatever the closure returned. #[cfg(feature = "std")] pub fn parse_slice<F, T>(data: &[u8], op: F) -> Result<T, ParseError> where F: FnOnce(&mut Parser<[u8]>) -> Result<T, ParseError>, { let mut parser = Parser::from_ref(data); let res = op(&mut parser)?; if parser.remaining() > 0 { Err(ParseError::form_error("trailing data")) }
n_compress(&
identifier_name
add_manifest_deployitem.go
) *cobra.Command { opts := &addManifestDeployItemOptions{} cmd := &cobra.Command{ Use: addManifestDeployItemUse, Example: addManifestDeployItemExample, Short: addManifestDeployItemShort, Args: cobra.ExactArgs(1), Run: func(cmd *cobra.Command, args []string) { if err := opts.Complete(args); err != nil { fmt.Println(err.Error()) os.Exit(1) } if err := opts.run(ctx, logger.Log); err != nil { fmt.Println(err.Error()) os.Exit(1) } fmt.Printf("Deploy item added") fmt.Printf(" \n- deploy item definition in blueprint folder in file %s created", util.ExecutionFileName(opts.deployItemName)) fmt.Printf(" \n- file reference to deploy item definition added to blueprint") fmt.Printf(" \n- import definitions added to blueprint") }, } opts.AddFlags(cmd.Flags()) return cmd } func (o *addManifestDeployItemOptions) Complete(args []string) error { o.deployItemName = args[0] if err := o.parseParameterDefinitions(); err != nil { return err } return o.validate() } func (o *addManifestDeployItemOptions) AddFlags(fs *pflag.FlagSet) { fs.StringVar(&o.componentPath, "component-directory", ".", "path to component directory (optional, default is current directory)") o.files = fs.StringArray( "manifest-file", []string{}, "manifest file containing one kubernetes resource") o.importParams = fs.StringArray( "import-param", []string{}, "import parameter as name:integer|string|boolean, e.g. replicas:integer") fs.StringVar(&o.updateStrategy, "update-strategy", "update", "update stategy") fs.StringVar(&o.policy, "policy", "manage", "policy") fs.StringVar(&o.clusterParam, "cluster-param", "targetCluster", "import parameter name for the target resource containing the access data of the target cluster") } func (o *addManifestDeployItemOptions) parseParameterDefinitions() (err error) { p := components.ParameterDefinitionParser{} o.importDefinitions, err = p.ParseImportDefinitions(o.importParams) if err != nil { return err } o.replacement = map[string]string{} for paramName := range o.importDefinitions { o.replacement[paramName] = string(uuid.NewUUID()) } return nil } func (o *addManifestDeployItemOptions) validate() error { if !identityKeyValidationRegexp.Match([]byte(o.deployItemName)) { return fmt.Errorf("the deploy item name must consist of lower case alphanumeric characters, '-', '_' " + "or '+', and must start and end with an alphanumeric character") } if o.clusterParam == "" { return fmt.Errorf("cluster-param is missing") } if o.files == nil || len(*(o.files)) == 0 { return fmt.Errorf("no manifest files specified") } for _, path := range *(o.files) { fileInfo, err := os.Stat(path) if err != nil { if os.IsNotExist(err) { return fmt.Errorf("manifest file %s does not exist", path) } return err } if fileInfo.IsDir() { return fmt.Errorf("manifest file %s is a directory", path) } } err := o.checkIfDeployItemNotAlreadyAdded() if err != nil { return err } return nil } func (o *addManifestDeployItemOptions) run(ctx context.Context, log logr.Logger) error { err := o.createExecutionFile() if err != nil { return err } blueprintPath := util.BlueprintDirectoryPath(o.componentPath) blueprint, err := blueprints.NewBlueprintReader(blueprintPath).Read() if err != nil { return err } blueprintBuilder := blueprints.NewBlueprintBuilder(blueprint) if blueprintBuilder.ExistsDeployExecution(o.deployItemName) { return fmt.Errorf("The blueprint already contains a deploy item %s\n", o.deployItemName) } blueprintBuilder.AddDeployExecution(o.deployItemName) blueprintBuilder.AddImportForTarget(o.clusterParam) blueprintBuilder.AddImportsFromMap(o.importDefinitions) return blueprints.NewBlueprintWriter(blueprintPath).Write(blueprint) } func (o *addManifestDeployItemOptions) checkIfDeployItemNotAlreadyAdded() error { _, err := os.Stat(util.ExecutionFilePath(o.componentPath, o.deployItemName)) if err != nil { if os.IsNotExist(err) { return nil } return err } return fmt.Errorf("Deploy item was already added. The corresponding deploy execution file %s already exists\n", util.ExecutionFilePath(o.componentPath, o.deployItemName)) } // parseImportDefinition creates a new ImportDefinition from a given parameter definition string. // The parameter definition string must have the format "name:type", for example "replicas:integer". // The supported types are: string, boolean, integer func (o *addManifestDeployItemOptions) parseImportDefinition(paramDef string) (*v1alpha1.ImportDefinition, error) { p := components.ParameterDefinitionParser{} fieldValueDef, err := p.ParseFieldValueDefinition(paramDef) if err != nil { return nil, err } required := true return &v1alpha1.ImportDefinition{ FieldValueDefinition: *fieldValueDef, Required: &required, }, nil } func (o *addManifestDeployItemOptions) createExecutionFile() error { manifests, err := o.getManifests() if err != nil { return err } f, err := os.Create(util.ExecutionFilePath(o.componentPath, o.deployItemName)) if err != nil { return err } defer f.Close() err = o.writeExecution(f) if err != nil { return err } _, err = f.WriteString(manifests) return err } const manifestExecutionTemplate = `deployItems: - name: {{.DeployItemName}} type: landscaper.gardener.cloud/kubernetes-manifest target: name: {{.TargetNameExpression}} namespace: {{.TargetNamespaceExpression}} config: apiVersion: manifest.deployer.landscaper.gardener.cloud/v1alpha2 kind: ProviderConfiguration updateStrategy: {{.UpdateStrategy}} ` func (o *addManifestDeployItemOptions) writeExecution(f io.Writer) error { t, err := template.New("").Parse(manifestExecutionTemplate) if err != nil { return err } data := struct { DeployItemName string TargetNameExpression string TargetNamespaceExpression string UpdateStrategy string }{ DeployItemName: o.deployItemName, TargetNameExpression: blueprints.GetTargetNameExpression(o.clusterParam), TargetNamespaceExpression: blueprints.GetTargetNamespaceExpression(o.clusterParam), UpdateStrategy: o.updateStrategy, } err = t.Execute(f, data) if err != nil { return err } return nil } func (o *addManifestDeployItemOptions) getManifests() (string, error) { data, err := o.getManifestsYaml() if err != nil { return "", err } stringData := string(data) stringData = indentLines(stringData, 4) return stringData, nil } func indentLines(data string, n int) string { indent := strings.Repeat(" ", n) return indent + strings.ReplaceAll(data, "\n", "\n"+indent) } func (o *addManifestDeployItemOptions) getManifestsYaml() ([]byte, error) { manifests, err := o.readManifests() if err != nil { return nil, err } m := map[string][]managedresource.Manifest{ "manifests": manifests, } data, err := yaml.Marshal(m) if err != nil { return nil, err } data = o.replaceUUIDsByImportTemplates(data) return data, nil } func (o *addManifestDeployItemOptions) readManifests() ([]managedresource.Manifest, error) { manifests := []managedresource.Manifest{} if o.files == nil { return manifests, nil } for _, filename := range *o.files { m, err := o.readManifest(filename) if err != nil { return manifests, err } manifests = append(manifests, *m) } return manifests, nil } func (o *addManifestDeployItemOptions) readManifest(filename string) (*managedresource.Manifest, error) { yamlData, err := ioutil.ReadFile(filename) if err != nil { return nil, err } var m interface{} err = yaml.Unmarshal(yamlData, &m) if err != nil { return nil, err } m = o.replaceParamsByUUIDs(m) // render to string uuidData, err := json.Marshal(m) if err != nil { return nil, err } m2 := &managedresource.Manifest{ Policy: managedresource.ManifestPolicy(o.policy), Manifest: &runtime.RawExtension{ Raw: uuidData, }, } return m2, nil } func (o *addManifestDeployItemOptions)
replaceParamsByUUIDs
identifier_name
add_manifest_deployitem.go
cli/pkg/logger" "github.com/gardener/landscapercli/pkg/util" ) const addManifestDeployItemUse = `deployitem \ [deployitem name] \ ` const addManifestDeployItemExample = ` landscaper-cli component add manifest deployitem \ nginx \ --component-directory ~/myComponent \ --manifest-file ./deployment.yaml \ --manifest-file ./service.yaml \ --import-param replicas:integer --cluster-param target-cluster ` const addManifestDeployItemShort = ` Command to add a deploy item skeleton to the blueprint of a component` //var identityKeyValidationRegexp = regexp.MustCompile("^[a-z0-9]([-_+a-z0-9]*[a-z0-9])?$") type addManifestDeployItemOptions struct { componentPath string deployItemName string // names of manifest files files *[]string // import parameter definitions in the format "name:type" importParams *[]string // parsed import parameter definitions importDefinitions map[string]*v1alpha1.ImportDefinition // a map that assigns with each import parameter name a uuid replacement map[string]string updateStrategy string policy string clusterParam string } // NewCreateCommand creates a new blueprint command to create a blueprint func NewAddManifestDeployItemCommand(ctx context.Context) *cobra.Command { opts := &addManifestDeployItemOptions{} cmd := &cobra.Command{ Use: addManifestDeployItemUse, Example: addManifestDeployItemExample, Short: addManifestDeployItemShort, Args: cobra.ExactArgs(1), Run: func(cmd *cobra.Command, args []string) { if err := opts.Complete(args); err != nil { fmt.Println(err.Error()) os.Exit(1) } if err := opts.run(ctx, logger.Log); err != nil { fmt.Println(err.Error()) os.Exit(1) } fmt.Printf("Deploy item added") fmt.Printf(" \n- deploy item definition in blueprint folder in file %s created", util.ExecutionFileName(opts.deployItemName)) fmt.Printf(" \n- file reference to deploy item definition added to blueprint") fmt.Printf(" \n- import definitions added to blueprint") }, } opts.AddFlags(cmd.Flags()) return cmd } func (o *addManifestDeployItemOptions) Complete(args []string) error { o.deployItemName = args[0] if err := o.parseParameterDefinitions(); err != nil { return err } return o.validate() } func (o *addManifestDeployItemOptions) AddFlags(fs *pflag.FlagSet) { fs.StringVar(&o.componentPath, "component-directory", ".", "path to component directory (optional, default is current directory)") o.files = fs.StringArray( "manifest-file", []string{}, "manifest file containing one kubernetes resource") o.importParams = fs.StringArray( "import-param", []string{}, "import parameter as name:integer|string|boolean, e.g. replicas:integer") fs.StringVar(&o.updateStrategy, "update-strategy", "update", "update stategy") fs.StringVar(&o.policy, "policy", "manage", "policy") fs.StringVar(&o.clusterParam, "cluster-param", "targetCluster", "import parameter name for the target resource containing the access data of the target cluster") } func (o *addManifestDeployItemOptions) parseParameterDefinitions() (err error) { p := components.ParameterDefinitionParser{} o.importDefinitions, err = p.ParseImportDefinitions(o.importParams) if err != nil { return err } o.replacement = map[string]string{} for paramName := range o.importDefinitions { o.replacement[paramName] = string(uuid.NewUUID()) } return nil } func (o *addManifestDeployItemOptions) validate() error
return err } if fileInfo.IsDir() { return fmt.Errorf("manifest file %s is a directory", path) } } err := o.checkIfDeployItemNotAlreadyAdded() if err != nil { return err } return nil } func (o *addManifestDeployItemOptions) run(ctx context.Context, log logr.Logger) error { err := o.createExecutionFile() if err != nil { return err } blueprintPath := util.BlueprintDirectoryPath(o.componentPath) blueprint, err := blueprints.NewBlueprintReader(blueprintPath).Read() if err != nil { return err } blueprintBuilder := blueprints.NewBlueprintBuilder(blueprint) if blueprintBuilder.ExistsDeployExecution(o.deployItemName) { return fmt.Errorf("The blueprint already contains a deploy item %s\n", o.deployItemName) } blueprintBuilder.AddDeployExecution(o.deployItemName) blueprintBuilder.AddImportForTarget(o.clusterParam) blueprintBuilder.AddImportsFromMap(o.importDefinitions) return blueprints.NewBlueprintWriter(blueprintPath).Write(blueprint) } func (o *addManifestDeployItemOptions) checkIfDeployItemNotAlreadyAdded() error { _, err := os.Stat(util.ExecutionFilePath(o.componentPath, o.deployItemName)) if err != nil { if os.IsNotExist(err) { return nil } return err } return fmt.Errorf("Deploy item was already added. The corresponding deploy execution file %s already exists\n", util.ExecutionFilePath(o.componentPath, o.deployItemName)) } // parseImportDefinition creates a new ImportDefinition from a given parameter definition string. // The parameter definition string must have the format "name:type", for example "replicas:integer". // The supported types are: string, boolean, integer func (o *addManifestDeployItemOptions) parseImportDefinition(paramDef string) (*v1alpha1.ImportDefinition, error) { p := components.ParameterDefinitionParser{} fieldValueDef, err := p.ParseFieldValueDefinition(paramDef) if err != nil { return nil, err } required := true return &v1alpha1.ImportDefinition{ FieldValueDefinition: *fieldValueDef, Required: &required, }, nil } func (o *addManifestDeployItemOptions) createExecutionFile() error { manifests, err := o.getManifests() if err != nil { return err } f, err := os.Create(util.ExecutionFilePath(o.componentPath, o.deployItemName)) if err != nil { return err } defer f.Close() err = o.writeExecution(f) if err != nil { return err } _, err = f.WriteString(manifests) return err } const manifestExecutionTemplate = `deployItems: - name: {{.DeployItemName}} type: landscaper.gardener.cloud/kubernetes-manifest target: name: {{.TargetNameExpression}} namespace: {{.TargetNamespaceExpression}} config: apiVersion: manifest.deployer.landscaper.gardener.cloud/v1alpha2 kind: ProviderConfiguration updateStrategy: {{.UpdateStrategy}} ` func (o *addManifestDeployItemOptions) writeExecution(f io.Writer) error { t, err := template.New("").Parse(manifestExecutionTemplate) if err != nil { return err } data := struct { DeployItemName string TargetNameExpression string TargetNamespaceExpression string UpdateStrategy string }{ DeployItemName: o.deployItemName, TargetNameExpression: blueprints.GetTargetNameExpression(o.clusterParam), TargetNamespaceExpression: blueprints.GetTargetNamespaceExpression(o.clusterParam), UpdateStrategy: o.updateStrategy, } err = t.Execute(f, data) if err != nil { return err } return nil } func (o *addManifestDeployItemOptions) getManifests() (string, error) { data, err := o.getManifestsYaml() if err != nil { return "", err } stringData := string(data) stringData = indentLines(stringData, 4) return stringData, nil } func indentLines(data string, n int) string { indent := strings.Repeat(" ", n) return indent + strings.ReplaceAll(data, "\n", "\n"+indent) } func (o *addManifestDeployItemOptions) getManifestsYaml() ([]byte, error) { manifests, err := o.readManifests() if err != nil { return nil, err } m := map[string][]managedresource.Manifest{ "manifests": manifests, } data, err := yaml.Marshal(m) if err != nil { return nil, err } data = o.replaceUUIDsByImport
{ if !identityKeyValidationRegexp.Match([]byte(o.deployItemName)) { return fmt.Errorf("the deploy item name must consist of lower case alphanumeric characters, '-', '_' " + "or '+', and must start and end with an alphanumeric character") } if o.clusterParam == "" { return fmt.Errorf("cluster-param is missing") } if o.files == nil || len(*(o.files)) == 0 { return fmt.Errorf("no manifest files specified") } for _, path := range *(o.files) { fileInfo, err := os.Stat(path) if err != nil { if os.IsNotExist(err) { return fmt.Errorf("manifest file %s does not exist", path) }
identifier_body
add_manifest_deployitem.go
Example = ` landscaper-cli component add manifest deployitem \ nginx \ --component-directory ~/myComponent \ --manifest-file ./deployment.yaml \ --manifest-file ./service.yaml \ --import-param replicas:integer --cluster-param target-cluster ` const addManifestDeployItemShort = ` Command to add a deploy item skeleton to the blueprint of a component` //var identityKeyValidationRegexp = regexp.MustCompile("^[a-z0-9]([-_+a-z0-9]*[a-z0-9])?$") type addManifestDeployItemOptions struct { componentPath string deployItemName string // names of manifest files files *[]string // import parameter definitions in the format "name:type" importParams *[]string // parsed import parameter definitions importDefinitions map[string]*v1alpha1.ImportDefinition // a map that assigns with each import parameter name a uuid replacement map[string]string updateStrategy string policy string clusterParam string } // NewCreateCommand creates a new blueprint command to create a blueprint func NewAddManifestDeployItemCommand(ctx context.Context) *cobra.Command { opts := &addManifestDeployItemOptions{} cmd := &cobra.Command{ Use: addManifestDeployItemUse, Example: addManifestDeployItemExample, Short: addManifestDeployItemShort, Args: cobra.ExactArgs(1), Run: func(cmd *cobra.Command, args []string) { if err := opts.Complete(args); err != nil { fmt.Println(err.Error()) os.Exit(1) } if err := opts.run(ctx, logger.Log); err != nil { fmt.Println(err.Error()) os.Exit(1) } fmt.Printf("Deploy item added") fmt.Printf(" \n- deploy item definition in blueprint folder in file %s created", util.ExecutionFileName(opts.deployItemName)) fmt.Printf(" \n- file reference to deploy item definition added to blueprint") fmt.Printf(" \n- import definitions added to blueprint") }, } opts.AddFlags(cmd.Flags()) return cmd } func (o *addManifestDeployItemOptions) Complete(args []string) error { o.deployItemName = args[0] if err := o.parseParameterDefinitions(); err != nil { return err } return o.validate() } func (o *addManifestDeployItemOptions) AddFlags(fs *pflag.FlagSet) { fs.StringVar(&o.componentPath, "component-directory", ".", "path to component directory (optional, default is current directory)") o.files = fs.StringArray( "manifest-file", []string{}, "manifest file containing one kubernetes resource") o.importParams = fs.StringArray( "import-param", []string{}, "import parameter as name:integer|string|boolean, e.g. replicas:integer") fs.StringVar(&o.updateStrategy, "update-strategy", "update", "update stategy") fs.StringVar(&o.policy, "policy", "manage", "policy") fs.StringVar(&o.clusterParam, "cluster-param", "targetCluster", "import parameter name for the target resource containing the access data of the target cluster") } func (o *addManifestDeployItemOptions) parseParameterDefinitions() (err error) { p := components.ParameterDefinitionParser{} o.importDefinitions, err = p.ParseImportDefinitions(o.importParams) if err != nil { return err } o.replacement = map[string]string{} for paramName := range o.importDefinitions { o.replacement[paramName] = string(uuid.NewUUID()) } return nil } func (o *addManifestDeployItemOptions) validate() error { if !identityKeyValidationRegexp.Match([]byte(o.deployItemName)) { return fmt.Errorf("the deploy item name must consist of lower case alphanumeric characters, '-', '_' " + "or '+', and must start and end with an alphanumeric character") } if o.clusterParam == "" { return fmt.Errorf("cluster-param is missing") } if o.files == nil || len(*(o.files)) == 0 { return fmt.Errorf("no manifest files specified") } for _, path := range *(o.files) { fileInfo, err := os.Stat(path) if err != nil { if os.IsNotExist(err) { return fmt.Errorf("manifest file %s does not exist", path) } return err } if fileInfo.IsDir() { return fmt.Errorf("manifest file %s is a directory", path) } } err := o.checkIfDeployItemNotAlreadyAdded() if err != nil { return err } return nil } func (o *addManifestDeployItemOptions) run(ctx context.Context, log logr.Logger) error { err := o.createExecutionFile() if err != nil { return err } blueprintPath := util.BlueprintDirectoryPath(o.componentPath) blueprint, err := blueprints.NewBlueprintReader(blueprintPath).Read() if err != nil { return err } blueprintBuilder := blueprints.NewBlueprintBuilder(blueprint) if blueprintBuilder.ExistsDeployExecution(o.deployItemName) { return fmt.Errorf("The blueprint already contains a deploy item %s\n", o.deployItemName) } blueprintBuilder.AddDeployExecution(o.deployItemName) blueprintBuilder.AddImportForTarget(o.clusterParam) blueprintBuilder.AddImportsFromMap(o.importDefinitions) return blueprints.NewBlueprintWriter(blueprintPath).Write(blueprint) } func (o *addManifestDeployItemOptions) checkIfDeployItemNotAlreadyAdded() error { _, err := os.Stat(util.ExecutionFilePath(o.componentPath, o.deployItemName)) if err != nil { if os.IsNotExist(err) { return nil } return err } return fmt.Errorf("Deploy item was already added. The corresponding deploy execution file %s already exists\n", util.ExecutionFilePath(o.componentPath, o.deployItemName)) } // parseImportDefinition creates a new ImportDefinition from a given parameter definition string. // The parameter definition string must have the format "name:type", for example "replicas:integer". // The supported types are: string, boolean, integer func (o *addManifestDeployItemOptions) parseImportDefinition(paramDef string) (*v1alpha1.ImportDefinition, error) { p := components.ParameterDefinitionParser{} fieldValueDef, err := p.ParseFieldValueDefinition(paramDef) if err != nil { return nil, err } required := true return &v1alpha1.ImportDefinition{ FieldValueDefinition: *fieldValueDef, Required: &required, }, nil } func (o *addManifestDeployItemOptions) createExecutionFile() error { manifests, err := o.getManifests() if err != nil { return err } f, err := os.Create(util.ExecutionFilePath(o.componentPath, o.deployItemName)) if err != nil { return err } defer f.Close() err = o.writeExecution(f) if err != nil { return err } _, err = f.WriteString(manifests) return err } const manifestExecutionTemplate = `deployItems: - name: {{.DeployItemName}} type: landscaper.gardener.cloud/kubernetes-manifest target: name: {{.TargetNameExpression}} namespace: {{.TargetNamespaceExpression}} config: apiVersion: manifest.deployer.landscaper.gardener.cloud/v1alpha2 kind: ProviderConfiguration updateStrategy: {{.UpdateStrategy}} ` func (o *addManifestDeployItemOptions) writeExecution(f io.Writer) error { t, err := template.New("").Parse(manifestExecutionTemplate) if err != nil { return err } data := struct { DeployItemName string TargetNameExpression string TargetNamespaceExpression string UpdateStrategy string }{ DeployItemName: o.deployItemName, TargetNameExpression: blueprints.GetTargetNameExpression(o.clusterParam), TargetNamespaceExpression: blueprints.GetTargetNamespaceExpression(o.clusterParam), UpdateStrategy: o.updateStrategy, } err = t.Execute(f, data) if err != nil { return err } return nil } func (o *addManifestDeployItemOptions) getManifests() (string, error) { data, err := o.getManifestsYaml() if err != nil { return "", err } stringData := string(data) stringData = indentLines(stringData, 4) return stringData, nil } func indentLines(data string, n int) string { indent := strings.Repeat(" ", n) return indent + strings.ReplaceAll(data, "\n", "\n"+indent) } func (o *addManifestDeployItemOptions) getManifestsYaml() ([]byte, error) { manifests, err := o.readManifests() if err != nil { return nil, err } m := map[string][]managedresource.Manifest{ "manifests": manifests, } data, err := yaml.Marshal(m) if err != nil { return nil, err } data = o.replaceUUIDsByImportTemplates(data) return data, nil }
func (o *addManifestDeployItemOptions) readManifests() ([]managedresource.Manifest, error) { manifests := []managedresource.Manifest{} if o.files == nil {
random_line_split
add_manifest_deployitem.go
cli/pkg/logger" "github.com/gardener/landscapercli/pkg/util" ) const addManifestDeployItemUse = `deployitem \ [deployitem name] \ ` const addManifestDeployItemExample = ` landscaper-cli component add manifest deployitem \ nginx \ --component-directory ~/myComponent \ --manifest-file ./deployment.yaml \ --manifest-file ./service.yaml \ --import-param replicas:integer --cluster-param target-cluster ` const addManifestDeployItemShort = ` Command to add a deploy item skeleton to the blueprint of a component` //var identityKeyValidationRegexp = regexp.MustCompile("^[a-z0-9]([-_+a-z0-9]*[a-z0-9])?$") type addManifestDeployItemOptions struct { componentPath string deployItemName string // names of manifest files files *[]string // import parameter definitions in the format "name:type" importParams *[]string // parsed import parameter definitions importDefinitions map[string]*v1alpha1.ImportDefinition // a map that assigns with each import parameter name a uuid replacement map[string]string updateStrategy string policy string clusterParam string } // NewCreateCommand creates a new blueprint command to create a blueprint func NewAddManifestDeployItemCommand(ctx context.Context) *cobra.Command { opts := &addManifestDeployItemOptions{} cmd := &cobra.Command{ Use: addManifestDeployItemUse, Example: addManifestDeployItemExample, Short: addManifestDeployItemShort, Args: cobra.ExactArgs(1), Run: func(cmd *cobra.Command, args []string) { if err := opts.Complete(args); err != nil { fmt.Println(err.Error()) os.Exit(1) } if err := opts.run(ctx, logger.Log); err != nil { fmt.Println(err.Error()) os.Exit(1) } fmt.Printf("Deploy item added") fmt.Printf(" \n- deploy item definition in blueprint folder in file %s created", util.ExecutionFileName(opts.deployItemName)) fmt.Printf(" \n- file reference to deploy item definition added to blueprint") fmt.Printf(" \n- import definitions added to blueprint") }, } opts.AddFlags(cmd.Flags()) return cmd } func (o *addManifestDeployItemOptions) Complete(args []string) error { o.deployItemName = args[0] if err := o.parseParameterDefinitions(); err != nil { return err } return o.validate() } func (o *addManifestDeployItemOptions) AddFlags(fs *pflag.FlagSet) { fs.StringVar(&o.componentPath, "component-directory", ".", "path to component directory (optional, default is current directory)") o.files = fs.StringArray( "manifest-file", []string{}, "manifest file containing one kubernetes resource") o.importParams = fs.StringArray( "import-param", []string{}, "import parameter as name:integer|string|boolean, e.g. replicas:integer") fs.StringVar(&o.updateStrategy, "update-strategy", "update", "update stategy") fs.StringVar(&o.policy, "policy", "manage", "policy") fs.StringVar(&o.clusterParam, "cluster-param", "targetCluster", "import parameter name for the target resource containing the access data of the target cluster") } func (o *addManifestDeployItemOptions) parseParameterDefinitions() (err error) { p := components.ParameterDefinitionParser{} o.importDefinitions, err = p.ParseImportDefinitions(o.importParams) if err != nil { return err } o.replacement = map[string]string{} for paramName := range o.importDefinitions { o.replacement[paramName] = string(uuid.NewUUID()) } return nil } func (o *addManifestDeployItemOptions) validate() error { if !identityKeyValidationRegexp.Match([]byte(o.deployItemName)) { return fmt.Errorf("the deploy item name must consist of lower case alphanumeric characters, '-', '_' " + "or '+', and must start and end with an alphanumeric character") } if o.clusterParam == "" { return fmt.Errorf("cluster-param is missing") } if o.files == nil || len(*(o.files)) == 0 { return fmt.Errorf("no manifest files specified") } for _, path := range *(o.files)
err := o.checkIfDeployItemNotAlreadyAdded() if err != nil { return err } return nil } func (o *addManifestDeployItemOptions) run(ctx context.Context, log logr.Logger) error { err := o.createExecutionFile() if err != nil { return err } blueprintPath := util.BlueprintDirectoryPath(o.componentPath) blueprint, err := blueprints.NewBlueprintReader(blueprintPath).Read() if err != nil { return err } blueprintBuilder := blueprints.NewBlueprintBuilder(blueprint) if blueprintBuilder.ExistsDeployExecution(o.deployItemName) { return fmt.Errorf("The blueprint already contains a deploy item %s\n", o.deployItemName) } blueprintBuilder.AddDeployExecution(o.deployItemName) blueprintBuilder.AddImportForTarget(o.clusterParam) blueprintBuilder.AddImportsFromMap(o.importDefinitions) return blueprints.NewBlueprintWriter(blueprintPath).Write(blueprint) } func (o *addManifestDeployItemOptions) checkIfDeployItemNotAlreadyAdded() error { _, err := os.Stat(util.ExecutionFilePath(o.componentPath, o.deployItemName)) if err != nil { if os.IsNotExist(err) { return nil } return err } return fmt.Errorf("Deploy item was already added. The corresponding deploy execution file %s already exists\n", util.ExecutionFilePath(o.componentPath, o.deployItemName)) } // parseImportDefinition creates a new ImportDefinition from a given parameter definition string. // The parameter definition string must have the format "name:type", for example "replicas:integer". // The supported types are: string, boolean, integer func (o *addManifestDeployItemOptions) parseImportDefinition(paramDef string) (*v1alpha1.ImportDefinition, error) { p := components.ParameterDefinitionParser{} fieldValueDef, err := p.ParseFieldValueDefinition(paramDef) if err != nil { return nil, err } required := true return &v1alpha1.ImportDefinition{ FieldValueDefinition: *fieldValueDef, Required: &required, }, nil } func (o *addManifestDeployItemOptions) createExecutionFile() error { manifests, err := o.getManifests() if err != nil { return err } f, err := os.Create(util.ExecutionFilePath(o.componentPath, o.deployItemName)) if err != nil { return err } defer f.Close() err = o.writeExecution(f) if err != nil { return err } _, err = f.WriteString(manifests) return err } const manifestExecutionTemplate = `deployItems: - name: {{.DeployItemName}} type: landscaper.gardener.cloud/kubernetes-manifest target: name: {{.TargetNameExpression}} namespace: {{.TargetNamespaceExpression}} config: apiVersion: manifest.deployer.landscaper.gardener.cloud/v1alpha2 kind: ProviderConfiguration updateStrategy: {{.UpdateStrategy}} ` func (o *addManifestDeployItemOptions) writeExecution(f io.Writer) error { t, err := template.New("").Parse(manifestExecutionTemplate) if err != nil { return err } data := struct { DeployItemName string TargetNameExpression string TargetNamespaceExpression string UpdateStrategy string }{ DeployItemName: o.deployItemName, TargetNameExpression: blueprints.GetTargetNameExpression(o.clusterParam), TargetNamespaceExpression: blueprints.GetTargetNamespaceExpression(o.clusterParam), UpdateStrategy: o.updateStrategy, } err = t.Execute(f, data) if err != nil { return err } return nil } func (o *addManifestDeployItemOptions) getManifests() (string, error) { data, err := o.getManifestsYaml() if err != nil { return "", err } stringData := string(data) stringData = indentLines(stringData, 4) return stringData, nil } func indentLines(data string, n int) string { indent := strings.Repeat(" ", n) return indent + strings.ReplaceAll(data, "\n", "\n"+indent) } func (o *addManifestDeployItemOptions) getManifestsYaml() ([]byte, error) { manifests, err := o.readManifests() if err != nil { return nil, err } m := map[string][]managedresource.Manifest{ "manifests": manifests, } data, err := yaml.Marshal(m) if err != nil { return nil, err } data = o.replaceUUIDsBy
{ fileInfo, err := os.Stat(path) if err != nil { if os.IsNotExist(err) { return fmt.Errorf("manifest file %s does not exist", path) } return err } if fileInfo.IsDir() { return fmt.Errorf("manifest file %s is a directory", path) } }
conditional_block
WBSTree.js
NodeSelectForCopy:function(obj){}, oncreateCopyTask:function(obj){}, //cuando se crea una tarea usando copiar, @param obj es la tarea resultado de la copia onNodeSelectForCut:function(obj){}, onNodePaste:function(obj){}, onNodeCut:function(obj){}, onCancelClipoard:function(obj){}, onNoderequestEdit:function(obj){return true;}, //se aplica a todos los nodos, cuando se envía una petición de edición, si retorna true se procede con la edicion onNodeEdit:function(obj){} //se lanza este evento luego que un nodo fue editado gráficamente. } }; for(var i in dataObj) { if(this.config[i] != undefined) { if(i == "listeners") { for(var j in dataObj[i]) this.config[i][j] = dataObj[i][j]; } else this.config[i] = dataObj[i]; } } //posicion inicial para algoritmo eco------------------- this.rootYOffset = -80; this.rootXOffset = 0; //---------------------------------------------------------- this.algorithm = this.config.algorithm; this.id = this.config.id; this.title = this.config.title; this.container = this.config.container; this.className = "WBSTree"; this.Dibujado = false; //si el arbol ya esta dibujado sera true this.items = this.config.items; this.nodos = []; this.grupo= this.config.group; this.nodoSeleccionado = this.config.nodoSeleccionado; this.svgcontainer = this.config.svgcontainer; this.listeners = this.config.listeners; this.clipboard = -1; this.TareaPreCortada = -1; this.TareaPreCopiada = -1; if(this.algorithm=='ecotree'){ this.maxLevelHeight = []; //array que indica la altura maxima de cada nivel this.maxLevelWidth = []; //array que indica el ancho maximo de cada nivel this.previousLevelNode = []; this.nodoDerechaMax = ''; //nodo que se encuentra mas a la derecha this.nodoizquierdaMax = ''; //nodo que se encuentra mas a la izquierda } } //--------------------------------------------------------------------------------------------------------------------------------------------- Components.WBSTree.prototype._setNeighbors = function(node, level) { //funcion utilizada en el algoritmo eco, pero util para otras cosas mas. node.leftNeighbor = this.previousLevelNode[level]; if(node.leftNeighbor != null) node.leftNeighbor.rightNeighbor = node; this.previousLevelNode[level] = node; } Components.WBSTree.prototype._setLevelHeight = function (node, level) { /* usado en el algoritmo eco Node: nodo que esta siendo procesado level: nivel en el que se encuentra el nodo. - si el valor maxLevelHeight[level] no esta aun definido se le asigna la altura del nodo, - si el valor ya estaba definido se busca el mayor y este queda como valor. */ if (this.maxLevelHeight[level] == null) this.maxLevelHeight[level] = 0; if(this.maxLevelHeight[level] < node.Alto) this.maxLevelHeight[level] = node.Alto; } Components.WBSTree.prototype._setLevelWidth = function (node, level) { /* usado en el algoritmo eco Node: nodo que esta siendo procesado level: nivel en el que se encuentra el nodo. - si el valor maxLevelWidth[level] no esta aun definido se le asigna el mayor ancho , - si el valor ya estaba definido se busca el mayor y este queda como valor. */ if (this.maxLevelWidth[level] == null) this.maxLevelWidth[level] = 0; if(this.maxLevelWidth[level] < node.AnchoCajita) this.maxLevelWidth[level] = node.AnchoCajita; } //---------------------------------------------------------------------------------------------------------------------------------------------- Components.WBSTree.prototype.create = function() { this.root = this.svgcontainer.root; //El elemento padre - root de todas las imagenes SVG en el inspector es el elemento <svg> this.screenGrid = this.svgcontainer.screenGrid; this.svgContend = this.svgcontainer.svgContend; if (this.config.clipboard){ this.clipboard = new Clipboard(); this.clipboard.tree=this; } if(this.grupo != false) this.svgcontainer.svg.group({id: this.grupo}); //this.grupo=this.svgcontainer.svg.group({id: this.grupo}); this.MakeItems(); switch(this.algorithm){ case 'basicWilliam': this.LayoutAlgoritm= new LayoutWilliam(this); break; case 'ecotree': this.LayoutAlgoritm=new ECOTree(this); break } this.calcTree(); this.drawTree(); this.RecalcSize(); return this ; } //-------------------------------------------------------------------------------------------------------- Components.WBSTree.prototype.addNodo = function(nodo) { /*construir un nuevo nodo, con los datos enviados en la variable nodo. //el nodo se agrega a la derecha de sus hermanos, mas no se dibuja ni calcula aun. @param nodo tiene que tener esta estructura { type: 'WBSNode', id: id propio, diferente a los ids de los demas nodos, idp:id del padre, Descripcion:'Descripcion del contenido del nodo', } */ if(nodo.type != "WBSNode"){ return false; } if(nodo.id == -1){ return false; } nodo.container = this.root; nodo.screenGrid = this.screenGrid; nodo.tree = this; if(nodo.idp == null){ nodo.idp=-1; this.padres.push(nodo.id); if(position=='derecha'){ this.nodos[idp].childsId.push(nodo.id); } } else{ //agregar al nodo padre el nodo hijo como child if ( this.nodos[nodo.idp].childsId.indexOf(nodo.id)==-1) {this.nodos[nodo.idp].childsId.push(nodo.id);} this.nodos[nodo.idp].Status="Maximizado"; //sin esta linea el nodo pienza que esta minimizado //this.items[nodo.idp].childsId.push(nodo.id); } var element = Components.create('WBSNode', nodo); this.nodos[nodo.id] = element; this.nodos[nodo.id] = element; //los que fueron detectados como padres son los hijos del nodo -1 } //-------------------------------------------------------------------------------------------------------- Components.WBSTree.prototype.MakeItems = function() { //construir todos los nodos a partir de los datos ingresados por el usuario. this.padres = []; for(var i in this.items) { var item = this.items[i]; if(item.type != "WBSNode") continue; item.container = this.root; item.screenGrid = this.screenGrid; item.tree = this; if(item.idp == null){ item.idp=-1; this.padres.push(item.id); } var element = Components.create('WBSNode', item); this.nodos[element.id] = element; } //fake item var fake = { type: 'WBSNode', id:-1, idp:null, Level:0, Descripcion:'', tipoObjeto:'WBSPARENT', childsId:this.padres, container: this.root, screenGrid : this.screenGrid, tree: this } var element = Components.create('WBSNode', fake); this.nodos[element.id] = element; //los que fueron detectados como padres son los hijos del nodo -1 for(i in this.padres){ this.nodos[element.id].childs[i]=this.nodos[this.padres[i]]; } } //----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- Components.WBSTree.prototype._getNodeSize = function (node) { switch(this.config.iRootOrientation) { case ECOTree.RO_TOP: case ECOTree.RO_BOTTOM: return node.AnchoCajita; case ECOTree.RO_RIGHT: case ECOTree.RO_LEFT: return node.Alto; } return 0; } //----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- Components.WBSTree.prototype.cambiarLayout=function(algoritmo){ if (this.algorithm == algoritmo){return false;} else{
this.algorithm ='basicWilliam'; return true; break; case 'ecotree': this.LayoutAlgoritm= new ECOTree(this); this.algorithm ='ecotree'; return true; break; } } } //----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- Components.WBSTree.prototype.prepareForSave=function(){ var resultado=new Array(); for(var i in this.nodos){ if(this.nodos[i].id == -1){ continue; } var NodoItem = { type:"WBSNode", id:this.nodos[i].id, Descripcion:this.nodos[i].Descripcion, Background:this.nodos[i].Background, st
switch(algoritmo){ case 'basicWilliam': this.LayoutAlgoritm=new LayoutWilliam(this);
random_line_split
request.rs
: &'b Bucket, path: &'b str, command: Command<'b>) -> Request<'b> { Request { bucket, path, command, datetime: Utc::now(), async: false, } } fn url(&self) -> Url { let mut url_str = match self.command { Command::GetBucketLocation => { format!("{}://{}", self.bucket.scheme(), self.bucket.self_host()) } _ => format!("{}://{}", self.bucket.scheme(), self.bucket.host()), }; match self.command { Command::GetBucketLocation => {} _ => { url_str.push_str("/"); url_str.push_str(&self.bucket.name()); } } if !self.path.starts_with('/') { url_str.push_str("/"); } match self.command { Command::GetBucketLocation => url_str.push_str(self.path), _ => url_str.push_str(&signing::uri_encode(self.path, false)), }; // Since every part of this URL is either pre-encoded or statically // generated, there's really no way this should fail. let mut url = Url::parse(&url_str).expect("static URL parsing"); for (key, value) in &self.bucket.extra_query { url.query_pairs_mut().append_pair(key, value); } if let Command::ListBucket { prefix, delimiter, continuation_token, } = self.command.clone() { let mut query_pairs = url.query_pairs_mut(); delimiter.map(|d| query_pairs.append_pair("delimiter", &d.clone())); query_pairs.append_pair("prefix", &prefix); query_pairs.append_pair("list-type", "2"); if let Some(token) = continuation_token { query_pairs.append_pair("continuation-token", &token); } } match self.command { Command::PutObjectTagging { .. } | Command::GetObjectTagging | Command::DeleteObjectTagging => { url.query_pairs_mut().append_pair("tagging", ""); } _ => {} } // println!("{}", url); url } fn content_length(&self) -> usize { match self.command { Command::PutObject { content, .. } => content.len(), Command::PutObjectTagging { tags } => tags.len(), _ => 0, } } fn content_type(&self) -> String { match self.command { Command::PutObject { content_type, .. } => content_type.into(), _ => "text/plain".into(), } } fn sha256(&self) -> String { match self.command { Command::PutObject { content, .. } => { let mut sha = Sha256::default(); sha.input(content); hex::encode(sha.result().as_slice()) } Command::PutObjectTagging { tags } => { let mut sha = Sha256::default(); sha.input(tags.as_bytes()); hex::encode(sha.result().as_slice()) } _ => EMPTY_PAYLOAD_SHA.into(), } } fn long_date(&self) -> String { self.datetime.format(LONG_DATE).to_string() } fn canonical_request(&self, headers: &HeaderMap) -> String { signing::canonical_request(
} fn string_to_sign(&self, request: &str) -> String { signing::string_to_sign(&self.datetime, &self.bucket.region(), request) } fn signing_key(&self) -> S3Result<Vec<u8>> { Ok(signing::signing_key( &self.datetime, &self.bucket.secret_key(), &self.bucket.region(), "s3", )?) } fn authorization(&self, headers: &HeaderMap) -> S3Result<String> { let canonical_request = self.canonical_request(headers); let string_to_sign = self.string_to_sign(&canonical_request); let mut hmac = signing::HmacSha256::new_varkey(&self.signing_key()?)?; hmac.input(string_to_sign.as_bytes()); let signature = hex::encode(hmac.result().code()); let signed_header = signing::signed_header_string(headers); Ok(signing::authorization_header( &self.bucket.access_key(), &self.datetime, &self.bucket.region(), &signed_header, &signature, )) } fn headers(&self) -> S3Result<HeaderMap> { // Generate this once, but it's used in more than one place. let sha256 = self.sha256(); // Start with extra_headers, that way our headers replace anything with // the same name. let mut headers = self .bucket .extra_headers .iter() .map(|(k, v)| Ok((k.parse::<HeaderName>()?, v.parse::<HeaderValue>()?))) .collect::<Result<HeaderMap, S3Error>>()?; match self.command { Command::GetBucketLocation => { headers.insert(header::HOST, self.bucket.self_host().parse()?) } _ => headers.insert(header::HOST, self.bucket.host().parse()?), }; headers.insert( header::CONTENT_LENGTH, self.content_length().to_string().parse()?, ); headers.insert(header::CONTENT_TYPE, self.content_type().parse()?); headers.insert("X-Amz-Content-Sha256", sha256.parse()?); headers.insert("X-Amz-Date", self.long_date().parse()?); if let Some(token) = self.bucket.credentials().token.as_ref() { headers.insert("X-Amz-Security-Token", token.parse()?); } if let Command::PutObjectTagging { tags } = self.command { let digest = md5::compute(tags); let hash = base64::encode(digest.as_ref()); headers.insert("Content-MD5", hash.parse()?); } else if let Command::PutObject { content, .. } = self.command { let digest = md5::compute(content); let hash = base64::encode(digest.as_ref()); headers.insert("Content-MD5", hash.parse()?); } else if let Command::GetObject {} = self.command { headers.insert( header::ACCEPT, HeaderValue::from_str("application/octet-stream")?, ); // headers.insert(header::ACCEPT_CHARSET, HeaderValue::from_str("UTF-8")?); } // This must be last, as it signs the other headers let authorization = self.authorization(&headers)?; headers.insert(header::AUTHORIZATION, authorization.parse()?); // The format of RFC2822 is somewhat malleable, so including it in // signed headers can cause signature mismatches. We do include the // X-Amz-Date header, so requests are still properly limited to a date // range and can't be used again e.g. reply attacks. Adding this header // after the generation of the Authorization header leaves it out of // the signed headers. headers.insert(header::DATE, self.datetime.to_rfc2822().parse()?); Ok(headers) } pub fn response_data(&self) -> S3Result<(Vec<u8>, u16)> { let response_data = self.response_data_future().then(|result| match result { Ok((response_data, status_code)) => Ok((response_data, status_code)), Err(e) => Err(e), }); let mut runtime = Runtime::new().unwrap(); runtime.block_on(response_data) } pub fn response_data_to_writer<T: Write>(&self, writer: &mut T) -> S3Result<u16> { let status_code_future = self.response_data_to_writer_future(writer) .then(|result| match result { Ok(status_code) => Ok(status_code), Err(_) => Err(S3Error::from("ReqwestFuture")), }); let mut runtime = Runtime::new().unwrap(); runtime.block_on(status_code_future) } pub fn response_future(&self) -> impl Future<Item = Response, Error = S3Error> { let client = if cfg!(feature = "no-verify-ssl") { async::Client::builder() .danger_accept_invalid_certs(true) .danger_accept_invalid_hostnames(true) .build() .expect("Could not build dangereous client!") } else { async::Client::new() }; // Build headers let headers = self.headers().expect("Could not get headers!"); // Get owned content to pass to reqwest let content = if let Command::PutObject { content, .. } = self.command { Vec::from(content) } else if let Command::PutObjectTagging { tags } = self.command { Vec::from(tags) } else { Vec::new() }; let request = client .request(self.command.http_verb(), self.url().as_str()) .headers(headers.to_owned()) .body(content.to_owned()); request.send().map_err(S3Error::from) } pub fn response_data_future(&self) -> impl Future<Item = (Vec<u8>, u16), Error = S3Error> { self.response_future() .and_then(|response| { // println!("{:?}", response.headers()); let status
self.command.http_verb().as_str(), &self.url(), headers, &self.sha256(), )
random_line_split
request.rs
: &'b Bucket, path: &'b str, command: Command<'b>) -> Request<'b> { Request { bucket, path, command, datetime: Utc::now(), async: false, } } fn url(&self) -> Url { let mut url_str = match self.command { Command::GetBucketLocation => { format!("{}://{}", self.bucket.scheme(), self.bucket.self_host()) } _ => format!("{}://{}", self.bucket.scheme(), self.bucket.host()), }; match self.command { Command::GetBucketLocation => {} _ => { url_str.push_str("/"); url_str.push_str(&self.bucket.name()); } } if !self.path.starts_with('/') { url_str.push_str("/"); } match self.command { Command::GetBucketLocation => url_str.push_str(self.path), _ => url_str.push_str(&signing::uri_encode(self.path, false)), }; // Since every part of this URL is either pre-encoded or statically // generated, there's really no way this should fail. let mut url = Url::parse(&url_str).expect("static URL parsing"); for (key, value) in &self.bucket.extra_query { url.query_pairs_mut().append_pair(key, value); } if let Command::ListBucket { prefix, delimiter, continuation_token, } = self.command.clone() { let mut query_pairs = url.query_pairs_mut(); delimiter.map(|d| query_pairs.append_pair("delimiter", &d.clone())); query_pairs.append_pair("prefix", &prefix); query_pairs.append_pair("list-type", "2"); if let Some(token) = continuation_token { query_pairs.append_pair("continuation-token", &token); } } match self.command { Command::PutObjectTagging { .. } | Command::GetObjectTagging | Command::DeleteObjectTagging => { url.query_pairs_mut().append_pair("tagging", ""); } _ => {} } // println!("{}", url); url } fn content_length(&self) -> usize { match self.command { Command::PutObject { content, .. } => content.len(), Command::PutObjectTagging { tags } => tags.len(), _ => 0, } } fn content_type(&self) -> String { match self.command { Command::PutObject { content_type, .. } => content_type.into(), _ => "text/plain".into(), } } fn sha256(&self) -> String { match self.command { Command::PutObject { content, .. } => { let mut sha = Sha256::default(); sha.input(content); hex::encode(sha.result().as_slice()) } Command::PutObjectTagging { tags } => { let mut sha = Sha256::default(); sha.input(tags.as_bytes()); hex::encode(sha.result().as_slice()) } _ => EMPTY_PAYLOAD_SHA.into(), } } fn long_date(&self) -> String { self.datetime.format(LONG_DATE).to_string() } fn canonical_request(&self, headers: &HeaderMap) -> String { signing::canonical_request( self.command.http_verb().as_str(), &self.url(), headers, &self.sha256(), ) } fn string_to_sign(&self, request: &str) -> String { signing::string_to_sign(&self.datetime, &self.bucket.region(), request) } fn signing_key(&self) -> S3Result<Vec<u8>> { Ok(signing::signing_key( &self.datetime, &self.bucket.secret_key(), &self.bucket.region(), "s3", )?) } fn authorization(&self, headers: &HeaderMap) -> S3Result<String> { let canonical_request = self.canonical_request(headers); let string_to_sign = self.string_to_sign(&canonical_request); let mut hmac = signing::HmacSha256::new_varkey(&self.signing_key()?)?; hmac.input(string_to_sign.as_bytes()); let signature = hex::encode(hmac.result().code()); let signed_header = signing::signed_header_string(headers); Ok(signing::authorization_header( &self.bucket.access_key(), &self.datetime, &self.bucket.region(), &signed_header, &signature, )) } fn headers(&self) -> S3Result<HeaderMap> { // Generate this once, but it's used in more than one place. let sha256 = self.sha256(); // Start with extra_headers, that way our headers replace anything with // the same name. let mut headers = self .bucket .extra_headers .iter() .map(|(k, v)| Ok((k.parse::<HeaderName>()?, v.parse::<HeaderValue>()?))) .collect::<Result<HeaderMap, S3Error>>()?; match self.command { Command::GetBucketLocation => { headers.insert(header::HOST, self.bucket.self_host().parse()?) } _ => headers.insert(header::HOST, self.bucket.host().parse()?), }; headers.insert( header::CONTENT_LENGTH, self.content_length().to_string().parse()?, ); headers.insert(header::CONTENT_TYPE, self.content_type().parse()?); headers.insert("X-Amz-Content-Sha256", sha256.parse()?); headers.insert("X-Amz-Date", self.long_date().parse()?); if let Some(token) = self.bucket.credentials().token.as_ref() { headers.insert("X-Amz-Security-Token", token.parse()?); } if let Command::PutObjectTagging { tags } = self.command { let digest = md5::compute(tags); let hash = base64::encode(digest.as_ref()); headers.insert("Content-MD5", hash.parse()?); } else if let Command::PutObject { content, .. } = self.command { let digest = md5::compute(content); let hash = base64::encode(digest.as_ref()); headers.insert("Content-MD5", hash.parse()?); } else if let Command::GetObject {} = self.command { headers.insert( header::ACCEPT, HeaderValue::from_str("application/octet-stream")?, ); // headers.insert(header::ACCEPT_CHARSET, HeaderValue::from_str("UTF-8")?); } // This must be last, as it signs the other headers let authorization = self.authorization(&headers)?; headers.insert(header::AUTHORIZATION, authorization.parse()?); // The format of RFC2822 is somewhat malleable, so including it in // signed headers can cause signature mismatches. We do include the // X-Amz-Date header, so requests are still properly limited to a date // range and can't be used again e.g. reply attacks. Adding this header // after the generation of the Authorization header leaves it out of // the signed headers. headers.insert(header::DATE, self.datetime.to_rfc2822().parse()?); Ok(headers) } pub fn
(&self) -> S3Result<(Vec<u8>, u16)> { let response_data = self.response_data_future().then(|result| match result { Ok((response_data, status_code)) => Ok((response_data, status_code)), Err(e) => Err(e), }); let mut runtime = Runtime::new().unwrap(); runtime.block_on(response_data) } pub fn response_data_to_writer<T: Write>(&self, writer: &mut T) -> S3Result<u16> { let status_code_future = self.response_data_to_writer_future(writer) .then(|result| match result { Ok(status_code) => Ok(status_code), Err(_) => Err(S3Error::from("ReqwestFuture")), }); let mut runtime = Runtime::new().unwrap(); runtime.block_on(status_code_future) } pub fn response_future(&self) -> impl Future<Item = Response, Error = S3Error> { let client = if cfg!(feature = "no-verify-ssl") { async::Client::builder() .danger_accept_invalid_certs(true) .danger_accept_invalid_hostnames(true) .build() .expect("Could not build dangereous client!") } else { async::Client::new() }; // Build headers let headers = self.headers().expect("Could not get headers!"); // Get owned content to pass to reqwest let content = if let Command::PutObject { content, .. } = self.command { Vec::from(content) } else if let Command::PutObjectTagging { tags } = self.command { Vec::from(tags) } else { Vec::new() }; let request = client .request(self.command.http_verb(), self.url().as_str()) .headers(headers.to_owned()) .body(content.to_owned()); request.send().map_err(S3Error::from) } pub fn response_data_future(&self) -> impl Future<Item = (Vec<u8>, u16), Error = S3Error> { self.response_future() .and_then(|response| { // println!("{:?}", response.headers()); let
response_data
identifier_name
request.rs
: &'b Bucket, path: &'b str, command: Command<'b>) -> Request<'b> { Request { bucket, path, command, datetime: Utc::now(), async: false, } } fn url(&self) -> Url { let mut url_str = match self.command { Command::GetBucketLocation => { format!("{}://{}", self.bucket.scheme(), self.bucket.self_host()) } _ => format!("{}://{}", self.bucket.scheme(), self.bucket.host()), }; match self.command { Command::GetBucketLocation => {} _ => { url_str.push_str("/"); url_str.push_str(&self.bucket.name()); } } if !self.path.starts_with('/') { url_str.push_str("/"); } match self.command { Command::GetBucketLocation => url_str.push_str(self.path), _ => url_str.push_str(&signing::uri_encode(self.path, false)), }; // Since every part of this URL is either pre-encoded or statically // generated, there's really no way this should fail. let mut url = Url::parse(&url_str).expect("static URL parsing"); for (key, value) in &self.bucket.extra_query { url.query_pairs_mut().append_pair(key, value); } if let Command::ListBucket { prefix, delimiter, continuation_token, } = self.command.clone() { let mut query_pairs = url.query_pairs_mut(); delimiter.map(|d| query_pairs.append_pair("delimiter", &d.clone())); query_pairs.append_pair("prefix", &prefix); query_pairs.append_pair("list-type", "2"); if let Some(token) = continuation_token { query_pairs.append_pair("continuation-token", &token); } } match self.command { Command::PutObjectTagging { .. } | Command::GetObjectTagging | Command::DeleteObjectTagging => { url.query_pairs_mut().append_pair("tagging", ""); } _ => {} } // println!("{}", url); url } fn content_length(&self) -> usize { match self.command { Command::PutObject { content, .. } => content.len(), Command::PutObjectTagging { tags } => tags.len(), _ => 0, } } fn content_type(&self) -> String { match self.command { Command::PutObject { content_type, .. } => content_type.into(), _ => "text/plain".into(), } } fn sha256(&self) -> String { match self.command { Command::PutObject { content, .. } => { let mut sha = Sha256::default(); sha.input(content); hex::encode(sha.result().as_slice()) } Command::PutObjectTagging { tags } => { let mut sha = Sha256::default(); sha.input(tags.as_bytes()); hex::encode(sha.result().as_slice()) } _ => EMPTY_PAYLOAD_SHA.into(), } } fn long_date(&self) -> String { self.datetime.format(LONG_DATE).to_string() } fn canonical_request(&self, headers: &HeaderMap) -> String { signing::canonical_request( self.command.http_verb().as_str(), &self.url(), headers, &self.sha256(), ) } fn string_to_sign(&self, request: &str) -> String { signing::string_to_sign(&self.datetime, &self.bucket.region(), request) } fn signing_key(&self) -> S3Result<Vec<u8>>
fn authorization(&self, headers: &HeaderMap) -> S3Result<String> { let canonical_request = self.canonical_request(headers); let string_to_sign = self.string_to_sign(&canonical_request); let mut hmac = signing::HmacSha256::new_varkey(&self.signing_key()?)?; hmac.input(string_to_sign.as_bytes()); let signature = hex::encode(hmac.result().code()); let signed_header = signing::signed_header_string(headers); Ok(signing::authorization_header( &self.bucket.access_key(), &self.datetime, &self.bucket.region(), &signed_header, &signature, )) } fn headers(&self) -> S3Result<HeaderMap> { // Generate this once, but it's used in more than one place. let sha256 = self.sha256(); // Start with extra_headers, that way our headers replace anything with // the same name. let mut headers = self .bucket .extra_headers .iter() .map(|(k, v)| Ok((k.parse::<HeaderName>()?, v.parse::<HeaderValue>()?))) .collect::<Result<HeaderMap, S3Error>>()?; match self.command { Command::GetBucketLocation => { headers.insert(header::HOST, self.bucket.self_host().parse()?) } _ => headers.insert(header::HOST, self.bucket.host().parse()?), }; headers.insert( header::CONTENT_LENGTH, self.content_length().to_string().parse()?, ); headers.insert(header::CONTENT_TYPE, self.content_type().parse()?); headers.insert("X-Amz-Content-Sha256", sha256.parse()?); headers.insert("X-Amz-Date", self.long_date().parse()?); if let Some(token) = self.bucket.credentials().token.as_ref() { headers.insert("X-Amz-Security-Token", token.parse()?); } if let Command::PutObjectTagging { tags } = self.command { let digest = md5::compute(tags); let hash = base64::encode(digest.as_ref()); headers.insert("Content-MD5", hash.parse()?); } else if let Command::PutObject { content, .. } = self.command { let digest = md5::compute(content); let hash = base64::encode(digest.as_ref()); headers.insert("Content-MD5", hash.parse()?); } else if let Command::GetObject {} = self.command { headers.insert( header::ACCEPT, HeaderValue::from_str("application/octet-stream")?, ); // headers.insert(header::ACCEPT_CHARSET, HeaderValue::from_str("UTF-8")?); } // This must be last, as it signs the other headers let authorization = self.authorization(&headers)?; headers.insert(header::AUTHORIZATION, authorization.parse()?); // The format of RFC2822 is somewhat malleable, so including it in // signed headers can cause signature mismatches. We do include the // X-Amz-Date header, so requests are still properly limited to a date // range and can't be used again e.g. reply attacks. Adding this header // after the generation of the Authorization header leaves it out of // the signed headers. headers.insert(header::DATE, self.datetime.to_rfc2822().parse()?); Ok(headers) } pub fn response_data(&self) -> S3Result<(Vec<u8>, u16)> { let response_data = self.response_data_future().then(|result| match result { Ok((response_data, status_code)) => Ok((response_data, status_code)), Err(e) => Err(e), }); let mut runtime = Runtime::new().unwrap(); runtime.block_on(response_data) } pub fn response_data_to_writer<T: Write>(&self, writer: &mut T) -> S3Result<u16> { let status_code_future = self.response_data_to_writer_future(writer) .then(|result| match result { Ok(status_code) => Ok(status_code), Err(_) => Err(S3Error::from("ReqwestFuture")), }); let mut runtime = Runtime::new().unwrap(); runtime.block_on(status_code_future) } pub fn response_future(&self) -> impl Future<Item = Response, Error = S3Error> { let client = if cfg!(feature = "no-verify-ssl") { async::Client::builder() .danger_accept_invalid_certs(true) .danger_accept_invalid_hostnames(true) .build() .expect("Could not build dangereous client!") } else { async::Client::new() }; // Build headers let headers = self.headers().expect("Could not get headers!"); // Get owned content to pass to reqwest let content = if let Command::PutObject { content, .. } = self.command { Vec::from(content) } else if let Command::PutObjectTagging { tags } = self.command { Vec::from(tags) } else { Vec::new() }; let request = client .request(self.command.http_verb(), self.url().as_str()) .headers(headers.to_owned()) .body(content.to_owned()); request.send().map_err(S3Error::from) } pub fn response_data_future(&self) -> impl Future<Item = (Vec<u8>, u16), Error = S3Error> { self.response_future() .and_then(|response| { // println!("{:?}", response.headers());
{ Ok(signing::signing_key( &self.datetime, &self.bucket.secret_key(), &self.bucket.region(), "s3", )?) }
identifier_body
__init__.py
predict_des.strip().replace("-", "") return predict_des return predict_des # def detect_texts(img_path, ocr): # """Detects text in the file.""" # result = ocr.ocr(img_path, det=False, rec=True, cls=False) # for line in result: # # print(line[0]) # if line[1] > 0.7: # return line[0] # return "" def translate_text_google_cloud(target, text): """Translates text into the target language. Target must be an ISO 639-1 language code. See https://g.co/cloud/translate/v2/translate-reference#supported_languages """ if isinstance(text, six.binary_type): text = text.decode("utf-8") # Text can also be a sequence of strings, in which case this method # will return a sequence of results for each text. result = translate_client.translate(text, target_language=target) result = html.unescape(result["translatedText"]) # print(u"Text: {}".format(result["input"])) print("Translation: {}".format(result)) # print(u"Detected source language: {}".format(result["detectedSourceLanguage"])) return result def translate_text(target, text):
def generate_subtitles( source_path, output=None, dst_language=DEFAULT_DST_LANGUAGE, debug=False, cloud=False, disable_time=False, min_height=80, max_height=100, l_v=240 ): """ Given an input audio/video file, generate subtitles in the specified language and format. """ # Opens the Video file print(f"starting: using cloud {cloud}, source_path {source_path}") if not cloud: ocr = PaddleOCR(lang='ch', use_gpu=False, rec_model_dir=r"C:\autosub_models\rec", cls_model_dir=r"C:\autosub_models\cls", det_model_dir=r"C:\autosub_models\det", use_angle_cls=True, rec_char_type='ch', drop_score=0.8, det_db_box_thresh=0.3, cls=True) cap = cv2.VideoCapture(source_path) # cap.set(3, 1280) # cap.set(4, 720) # cv2.createTrackbar("L - V", "Trackbars", 0, 100, nothing) # cv2.createTrackbar("Min height", "Trackbars", 80, 100, nothing) # cv2.createTrackbar("Max Height", "Trackbars", 100, 100, nothing) fps = cap.get(cv2.CAP_PROP_FPS) print(f"fps {fps}") time_per_frame = 1 / fps i = 0 div_frame = 6 # 5 frame /s sub_idx = 1 list_srt = [] old_des = "" prev_time = 0 current_time = 0 file_name = os.path.basename(source_path) extenstion = ".srt" if not disable_time else ".txt" filesub = f"{os.path.splitext(file_name)[0]}{extenstion}" if os.path.isfile(filesub): os.remove(filesub) while (cap.isOpened()): ret, frame = cap.read() if ret == False: break # min_height = cv2.getTrackbarPos("Min height", "Trackbars") # max_height = cv2.getTrackbarPos("Max Height", "Trackbars") # if max_height < min_height: # max_height = min_height + 10 # l_v = cv2.getTrackbarPos("L - V", "Trackbars") if i % div_frame == 0: prev_time_ts = datetime.utcfromtimestamp(prev_time).strftime('%H:%M:%S,%f')[:-4] current_time_ts = datetime.utcfromtimestamp(current_time).strftime('%H:%M:%S,%f')[:-4] h, w, c = frame.shape crop_img = frame[int(h * min_height/100):int(h * max_height/100), 0:w] hsv = cv2.cvtColor(crop_img, cv2.COLOR_BGR2HSV) # define range of white color in HSV # change it according to your need ! lower_white = np.array([0, 0, 246], dtype=np.uint8) upper_white = np.array([157, 21, 255], dtype=np.uint8) # Threshold the HSV image to get only white colors mask = cv2.inRange(hsv, lower_white, upper_white) # Bitwise-AND mask and original image crop_img = cv2.bitwise_and(crop_img, crop_img, mask=mask) # crop_img = cv2.cvtColor(crop_img, cv2.COLOR_HSV2RGB) # crop_img = cv2.cvtColor(crop_img, cv2.COLOR_RGB2GRAY) description = "" if cloud: success, encoded_image = cv2.imencode('.jpg', crop_img) description = detect_texts_google_cloud(encoded_image.tobytes()) else: # dst = cv2.fastNlMeansDenoisingColored(crop_img,None,10,10,7,21) # stacked = np.hstack((dst, crop_img)) if debug: # cv2.imshow('dst', dst) cv2.imshow('crop_img', crop_img) cv2.imshow('frame', frame) cv2.waitKey(1) result = ocr.ocr(crop_img, det=False, rec=True, cls=False) for line in result: # print(current_time_ts, line) if line[1] > 0.7: description = html.unescape(line[0].strip().replace(',', '').replace('、', '').replace('.', '')) break description = "" if len(description) < 6 else description prev_des = "" ratio = fuzz.ratio(description.lower(), old_des.lower()) if len(list_srt) > 0: prev_des = list_srt[-1]['description'] print(current_time_ts, description, ratio) if (old_des != "" or description == "") and (ratio < 70) and current_time - prev_time > 0.5: list_srt.append({ "description": old_des, "translate": translate_text(dst_language, old_des), "first_time": prev_time_ts, "last_time": current_time_ts, "sub_idx": sub_idx }) # with open(f"{os.path.splitext(file_name)[0]}_raw.srt", "a", encoding="utf-8") as myfile: # myfile.write(f"{list_srt[-1]['sub_idx']}\n") # myfile.write(f"{list_srt[-1]['first_time']} --> {list_srt[-1]['last_time']}\n") # myfile.write(f"{list_srt[-1]['description']}\n") # myfile.write('\n') # myfile.close() with open(filesub, "a", encoding="utf-8") as myfile_vi: if not disable_time: myfile_vi.write(f"{list_srt[-1]['sub_idx']}\n") myfile_vi.write(f"{list_srt[-1]['first_time']} --> {list_srt[-1]['last_time']}\n") myfile_vi.write(f"{list_srt[-1]['translate']}\n") myfile_vi.write('\n') myfile_vi.close() print(f"{list_srt[-1]['sub_idx']}\n") print(f"{list_srt[-1]['first_time']} --> {list_srt[-1]['last_time']}\n") print(f"{list_srt[-1]['description']}\n") print(f"{list_srt[-1]['translate']}\n") print(f"Similarity{ratio}\n") print('\n') sub_idx += 1 prev_time = current_time if description == "": prev_time = current_time old_des = description current_time += time_per_frame * div_frame i += 1 cap.release() return output def validate(args): """ Check that the CLI arguments passed to autosub are valid. """ if args.format not in FORMATTERS: print( "Subtitle format not supported. " "Run with --list-formats to see all supported formats." ) return False if args.src_language not in LANGUAGE_CODES.keys(): print( "Source language not supported. " "Run with --list-languages to see all supported languages." ) return False if args.dst_language not in LANGUAGE_CODES.keys(): print( "Destination language not supported. " "Run with --list-languages to see all supported languages." ) return False if not args.source_path: print("Error: You need to specify a source path.") return False return True def
"""Translates text into the target language. Target must be an ISO 639-1 language code. See https://g.co/cloud/translate/v2/translate-reference#supported_languages """ return text
identifier_body
__init__.py
(content): """Detects text in the file.""" image = vision.Image(content=content) response = client.text_detection(image=image, image_context={"language_hints": ["zh"]}) texts = response.text_annotations predict_des = "" for text in texts: predict_des = text.description predict_des = predict_des.strip().replace("-", "") return predict_des return predict_des # def detect_texts(img_path, ocr): # """Detects text in the file.""" # result = ocr.ocr(img_path, det=False, rec=True, cls=False) # for line in result: # # print(line[0]) # if line[1] > 0.7: # return line[0] # return "" def translate_text_google_cloud(target, text): """Translates text into the target language. Target must be an ISO 639-1 language code. See https://g.co/cloud/translate/v2/translate-reference#supported_languages """ if isinstance(text, six.binary_type): text = text.decode("utf-8") # Text can also be a sequence of strings, in which case this method # will return a sequence of results for each text. result = translate_client.translate(text, target_language=target) result = html.unescape(result["translatedText"]) # print(u"Text: {}".format(result["input"])) print("Translation: {}".format(result)) # print(u"Detected source language: {}".format(result["detectedSourceLanguage"])) return result def translate_text(target, text): """Translates text into the target language. Target must be an ISO 639-1 language code. See https://g.co/cloud/translate/v2/translate-reference#supported_languages """ return text def generate_subtitles( source_path, output=None, dst_language=DEFAULT_DST_LANGUAGE, debug=False, cloud=False, disable_time=False, min_height=80, max_height=100, l_v=240 ): """ Given an input audio/video file, generate subtitles in the specified language and format. """ # Opens the Video file print(f"starting: using cloud {cloud}, source_path {source_path}") if not cloud: ocr = PaddleOCR(lang='ch', use_gpu=False, rec_model_dir=r"C:\autosub_models\rec", cls_model_dir=r"C:\autosub_models\cls", det_model_dir=r"C:\autosub_models\det", use_angle_cls=True, rec_char_type='ch', drop_score=0.8, det_db_box_thresh=0.3, cls=True) cap = cv2.VideoCapture(source_path) # cap.set(3, 1280) # cap.set(4, 720) # cv2.createTrackbar("L - V", "Trackbars", 0, 100, nothing) # cv2.createTrackbar("Min height", "Trackbars", 80, 100, nothing) # cv2.createTrackbar("Max Height", "Trackbars", 100, 100, nothing) fps = cap.get(cv2.CAP_PROP_FPS) print(f"fps {fps}") time_per_frame = 1 / fps i = 0 div_frame = 6 # 5 frame /s sub_idx = 1 list_srt = [] old_des = "" prev_time = 0 current_time = 0 file_name = os.path.basename(source_path) extenstion = ".srt" if not disable_time else ".txt" filesub = f"{os.path.splitext(file_name)[0]}{extenstion}" if os.path.isfile(filesub): os.remove(filesub) while (cap.isOpened()): ret, frame = cap.read() if ret == False: break # min_height = cv2.getTrackbarPos("Min height", "Trackbars") # max_height = cv2.getTrackbarPos("Max Height", "Trackbars") # if max_height < min_height: # max_height = min_height + 10 # l_v = cv2.getTrackbarPos("L - V", "Trackbars") if i % div_frame == 0: prev_time_ts = datetime.utcfromtimestamp(prev_time).strftime('%H:%M:%S,%f')[:-4] current_time_ts = datetime.utcfromtimestamp(current_time).strftime('%H:%M:%S,%f')[:-4] h, w, c = frame.shape crop_img = frame[int(h * min_height/100):int(h * max_height/100), 0:w] hsv = cv2.cvtColor(crop_img, cv2.COLOR_BGR2HSV) # define range of white color in HSV # change it according to your need ! lower_white = np.array([0, 0, 246], dtype=np.uint8) upper_white = np.array([157, 21, 255], dtype=np.uint8) # Threshold the HSV image to get only white colors mask = cv2.inRange(hsv, lower_white, upper_white) # Bitwise-AND mask and original image crop_img = cv2.bitwise_and(crop_img, crop_img, mask=mask) # crop_img = cv2.cvtColor(crop_img, cv2.COLOR_HSV2RGB) # crop_img = cv2.cvtColor(crop_img, cv2.COLOR_RGB2GRAY) description = "" if cloud: success, encoded_image = cv2.imencode('.jpg', crop_img) description = detect_texts_google_cloud(encoded_image.tobytes()) else: # dst = cv2.fastNlMeansDenoisingColored(crop_img,None,10,10,7,21) # stacked = np.hstack((dst, crop_img)) if debug: # cv2.imshow('dst', dst) cv2.imshow('crop_img', crop_img) cv2.imshow('frame', frame) cv2.waitKey(1) result = ocr.ocr(crop_img, det=False, rec=True, cls=False) for line in result: # print(current_time_ts, line) if line[1] > 0.7: description = html.unescape(line[0].strip().replace(',', '').replace('、', '').replace('.', '')) break description = "" if len(description) < 6 else description prev_des = "" ratio = fuzz.ratio(description.lower(), old_des.lower()) if len(list_srt) > 0: prev_des = list_srt[-1]['description'] print(current_time_ts, description, ratio) if (old_des != "" or description == "") and (ratio < 70) and current_time - prev_time > 0.5: list_srt.append({ "description": old_des, "translate": translate_text(dst_language, old_des), "first_time": prev_time_ts, "last_time": current_time_ts, "sub_idx": sub_idx }) # with open(f"{os.path.splitext(file_name)[0]}_raw.srt", "a", encoding="utf-8") as myfile: # myfile.write(f"{list_srt[-1]['sub_idx']}\n") # myfile.write(f"{list_srt[-1]['first_time']} --> {list_srt[-1]['last_time']}\n") # myfile.write(f"{list_srt[-1]['description']}\n") # myfile.write('\n') # myfile.close() with open(filesub, "a", encoding="utf-8") as myfile_vi: if not disable_time: myfile_vi.write(f"{list_srt[-1]['sub_idx']}\n") myfile_vi.write(f"{list_srt[-1]['first_time']} --> {list_srt[-1]['last_time']}\n") myfile_vi.write(f"{list_srt[-1]['translate']}\n") myfile_vi.write('\n') myfile_vi.close() print(f"{list_srt[-1]['sub_idx']}\n") print(f"{list_srt[-1]['first_time']} --> {list_srt[-1]['last_time']}\n") print(f"{list_srt[-1]['description']}\n") print(f"{list_srt[-1]['translate']}\n") print(f"Similarity{ratio}\n") print('\n') sub_idx += 1 prev_time = current_time if description == "": prev_time = current_time old_des = description current_time += time_per_frame * div_frame i += 1 cap.release() return output def validate(args): """ Check that the CLI arguments passed to autosub are valid. """ if args.format not in FORMATTERS: print( "Subtitle format not supported. " "Run with --list-formats to see all supported formats." ) return False if args.src_language not in LANGUAGE_CODES.keys(): print( "Source language not supported. " "Run with --list-languages to see all supported languages." ) return
detect_texts_google_cloud
identifier_name
__init__.py
utf-8") # Text can also be a sequence of strings, in which case this method # will return a sequence of results for each text. result = translate_client.translate(text, target_language=target) result = html.unescape(result["translatedText"]) # print(u"Text: {}".format(result["input"])) print("Translation: {}".format(result)) # print(u"Detected source language: {}".format(result["detectedSourceLanguage"])) return result def translate_text(target, text): """Translates text into the target language. Target must be an ISO 639-1 language code. See https://g.co/cloud/translate/v2/translate-reference#supported_languages """ return text def generate_subtitles( source_path, output=None, dst_language=DEFAULT_DST_LANGUAGE, debug=False, cloud=False, disable_time=False, min_height=80, max_height=100, l_v=240 ): """ Given an input audio/video file, generate subtitles in the specified language and format. """ # Opens the Video file print(f"starting: using cloud {cloud}, source_path {source_path}") if not cloud: ocr = PaddleOCR(lang='ch', use_gpu=False, rec_model_dir=r"C:\autosub_models\rec", cls_model_dir=r"C:\autosub_models\cls", det_model_dir=r"C:\autosub_models\det", use_angle_cls=True, rec_char_type='ch', drop_score=0.8, det_db_box_thresh=0.3, cls=True) cap = cv2.VideoCapture(source_path) # cap.set(3, 1280) # cap.set(4, 720) # cv2.createTrackbar("L - V", "Trackbars", 0, 100, nothing) # cv2.createTrackbar("Min height", "Trackbars", 80, 100, nothing) # cv2.createTrackbar("Max Height", "Trackbars", 100, 100, nothing) fps = cap.get(cv2.CAP_PROP_FPS) print(f"fps {fps}") time_per_frame = 1 / fps i = 0 div_frame = 6 # 5 frame /s sub_idx = 1 list_srt = [] old_des = "" prev_time = 0 current_time = 0 file_name = os.path.basename(source_path) extenstion = ".srt" if not disable_time else ".txt" filesub = f"{os.path.splitext(file_name)[0]}{extenstion}" if os.path.isfile(filesub): os.remove(filesub) while (cap.isOpened()): ret, frame = cap.read() if ret == False: break # min_height = cv2.getTrackbarPos("Min height", "Trackbars") # max_height = cv2.getTrackbarPos("Max Height", "Trackbars") # if max_height < min_height: # max_height = min_height + 10 # l_v = cv2.getTrackbarPos("L - V", "Trackbars") if i % div_frame == 0: prev_time_ts = datetime.utcfromtimestamp(prev_time).strftime('%H:%M:%S,%f')[:-4] current_time_ts = datetime.utcfromtimestamp(current_time).strftime('%H:%M:%S,%f')[:-4] h, w, c = frame.shape crop_img = frame[int(h * min_height/100):int(h * max_height/100), 0:w] hsv = cv2.cvtColor(crop_img, cv2.COLOR_BGR2HSV) # define range of white color in HSV # change it according to your need ! lower_white = np.array([0, 0, 246], dtype=np.uint8) upper_white = np.array([157, 21, 255], dtype=np.uint8) # Threshold the HSV image to get only white colors mask = cv2.inRange(hsv, lower_white, upper_white) # Bitwise-AND mask and original image crop_img = cv2.bitwise_and(crop_img, crop_img, mask=mask) # crop_img = cv2.cvtColor(crop_img, cv2.COLOR_HSV2RGB) # crop_img = cv2.cvtColor(crop_img, cv2.COLOR_RGB2GRAY) description = "" if cloud: success, encoded_image = cv2.imencode('.jpg', crop_img) description = detect_texts_google_cloud(encoded_image.tobytes()) else: # dst = cv2.fastNlMeansDenoisingColored(crop_img,None,10,10,7,21) # stacked = np.hstack((dst, crop_img)) if debug: # cv2.imshow('dst', dst) cv2.imshow('crop_img', crop_img) cv2.imshow('frame', frame) cv2.waitKey(1) result = ocr.ocr(crop_img, det=False, rec=True, cls=False) for line in result: # print(current_time_ts, line) if line[1] > 0.7: description = html.unescape(line[0].strip().replace(',', '').replace('、', '').replace('.', '')) break description = "" if len(description) < 6 else description prev_des = "" ratio = fuzz.ratio(description.lower(), old_des.lower()) if len(list_srt) > 0: prev_des = list_srt[-1]['description'] print(current_time_ts, description, ratio) if (old_des != "" or description == "") and (ratio < 70) and current_time - prev_time > 0.5: list_srt.append({ "description": old_des, "translate": translate_text(dst_language, old_des), "first_time": prev_time_ts, "last_time": current_time_ts, "sub_idx": sub_idx }) # with open(f"{os.path.splitext(file_name)[0]}_raw.srt", "a", encoding="utf-8") as myfile: # myfile.write(f"{list_srt[-1]['sub_idx']}\n") # myfile.write(f"{list_srt[-1]['first_time']} --> {list_srt[-1]['last_time']}\n") # myfile.write(f"{list_srt[-1]['description']}\n") # myfile.write('\n') # myfile.close() with open(filesub, "a", encoding="utf-8") as myfile_vi: if not disable_time: myfile_vi.write(f"{list_srt[-1]['sub_idx']}\n") myfile_vi.write(f"{list_srt[-1]['first_time']} --> {list_srt[-1]['last_time']}\n") myfile_vi.write(f"{list_srt[-1]['translate']}\n") myfile_vi.write('\n') myfile_vi.close() print(f"{list_srt[-1]['sub_idx']}\n") print(f"{list_srt[-1]['first_time']} --> {list_srt[-1]['last_time']}\n") print(f"{list_srt[-1]['description']}\n") print(f"{list_srt[-1]['translate']}\n") print(f"Similarity{ratio}\n") print('\n') sub_idx += 1 prev_time = current_time if description == "": prev_time = current_time old_des = description current_time += time_per_frame * div_frame i += 1 cap.release() return output def validate(args): """ Check that the CLI arguments passed to autosub are valid. """ if args.format not in FORMATTERS: print( "Subtitle format not supported. " "Run with --list-formats to see all supported formats." ) return False if args.src_language not in LANGUAGE_CODES.keys(): print( "Source language not supported. " "Run with --list-languages to see all supported languages." ) return False if args.dst_language not in LANGUAGE_CODES.keys(): print( "Destination language not supported. " "Run with --list-languages to see all supported languages." ) return False if not args.source_path: print("Error: You need to specify a source path.") return False return True def main(): """ Run autosub as a command-line program. """ parser = argparse.ArgumentParser() parser.add_argument('source_path', help="Path to the video or audio file to subtitle", nargs='?') parser.add_argument('-C', '--concurrency', help="Number of concurrent API requests to make", type=int, default=DEFAULT_CONCURRENCY) parser.add_argument('-o', '--output', help="Output path for subtitles (by default, subtitles are saved in \ the same directory and name as the source path)") parser.add_argument('-F', '--format', help="Destination subtitle format", default=DEFAULT_SUBTITLE_FORMAT) parser.add_argument('-S', '--src-language', help="Language spoken in source file", default=DEFAULT_SRC_LANGUAGE)
random_line_split
__init__.py
use_angle_cls=True, rec_char_type='ch', drop_score=0.8, det_db_box_thresh=0.3, cls=True) cap = cv2.VideoCapture(source_path) # cap.set(3, 1280) # cap.set(4, 720) # cv2.createTrackbar("L - V", "Trackbars", 0, 100, nothing) # cv2.createTrackbar("Min height", "Trackbars", 80, 100, nothing) # cv2.createTrackbar("Max Height", "Trackbars", 100, 100, nothing) fps = cap.get(cv2.CAP_PROP_FPS) print(f"fps {fps}") time_per_frame = 1 / fps i = 0 div_frame = 6 # 5 frame /s sub_idx = 1 list_srt = [] old_des = "" prev_time = 0 current_time = 0 file_name = os.path.basename(source_path) extenstion = ".srt" if not disable_time else ".txt" filesub = f"{os.path.splitext(file_name)[0]}{extenstion}" if os.path.isfile(filesub): os.remove(filesub) while (cap.isOpened()): ret, frame = cap.read() if ret == False: break # min_height = cv2.getTrackbarPos("Min height", "Trackbars") # max_height = cv2.getTrackbarPos("Max Height", "Trackbars") # if max_height < min_height: # max_height = min_height + 10 # l_v = cv2.getTrackbarPos("L - V", "Trackbars") if i % div_frame == 0: prev_time_ts = datetime.utcfromtimestamp(prev_time).strftime('%H:%M:%S,%f')[:-4] current_time_ts = datetime.utcfromtimestamp(current_time).strftime('%H:%M:%S,%f')[:-4] h, w, c = frame.shape crop_img = frame[int(h * min_height/100):int(h * max_height/100), 0:w] hsv = cv2.cvtColor(crop_img, cv2.COLOR_BGR2HSV) # define range of white color in HSV # change it according to your need ! lower_white = np.array([0, 0, 246], dtype=np.uint8) upper_white = np.array([157, 21, 255], dtype=np.uint8) # Threshold the HSV image to get only white colors mask = cv2.inRange(hsv, lower_white, upper_white) # Bitwise-AND mask and original image crop_img = cv2.bitwise_and(crop_img, crop_img, mask=mask) # crop_img = cv2.cvtColor(crop_img, cv2.COLOR_HSV2RGB) # crop_img = cv2.cvtColor(crop_img, cv2.COLOR_RGB2GRAY) description = "" if cloud: success, encoded_image = cv2.imencode('.jpg', crop_img) description = detect_texts_google_cloud(encoded_image.tobytes()) else: # dst = cv2.fastNlMeansDenoisingColored(crop_img,None,10,10,7,21) # stacked = np.hstack((dst, crop_img)) if debug: # cv2.imshow('dst', dst) cv2.imshow('crop_img', crop_img) cv2.imshow('frame', frame) cv2.waitKey(1) result = ocr.ocr(crop_img, det=False, rec=True, cls=False) for line in result: # print(current_time_ts, line) if line[1] > 0.7: description = html.unescape(line[0].strip().replace(',', '').replace('、', '').replace('.', '')) break description = "" if len(description) < 6 else description prev_des = "" ratio = fuzz.ratio(description.lower(), old_des.lower()) if len(list_srt) > 0: prev_des = list_srt[-1]['description'] print(current_time_ts, description, ratio) if (old_des != "" or description == "") and (ratio < 70) and current_time - prev_time > 0.5: list_srt.append({ "description": old_des, "translate": translate_text(dst_language, old_des), "first_time": prev_time_ts, "last_time": current_time_ts, "sub_idx": sub_idx }) # with open(f"{os.path.splitext(file_name)[0]}_raw.srt", "a", encoding="utf-8") as myfile: # myfile.write(f"{list_srt[-1]['sub_idx']}\n") # myfile.write(f"{list_srt[-1]['first_time']} --> {list_srt[-1]['last_time']}\n") # myfile.write(f"{list_srt[-1]['description']}\n") # myfile.write('\n') # myfile.close() with open(filesub, "a", encoding="utf-8") as myfile_vi: if not disable_time: myfile_vi.write(f"{list_srt[-1]['sub_idx']}\n") myfile_vi.write(f"{list_srt[-1]['first_time']} --> {list_srt[-1]['last_time']}\n") myfile_vi.write(f"{list_srt[-1]['translate']}\n") myfile_vi.write('\n') myfile_vi.close() print(f"{list_srt[-1]['sub_idx']}\n") print(f"{list_srt[-1]['first_time']} --> {list_srt[-1]['last_time']}\n") print(f"{list_srt[-1]['description']}\n") print(f"{list_srt[-1]['translate']}\n") print(f"Similarity{ratio}\n") print('\n') sub_idx += 1 prev_time = current_time if description == "": prev_time = current_time old_des = description current_time += time_per_frame * div_frame i += 1 cap.release() return output def validate(args): """ Check that the CLI arguments passed to autosub are valid. """ if args.format not in FORMATTERS: print( "Subtitle format not supported. " "Run with --list-formats to see all supported formats." ) return False if args.src_language not in LANGUAGE_CODES.keys(): print( "Source language not supported. " "Run with --list-languages to see all supported languages." ) return False if args.dst_language not in LANGUAGE_CODES.keys(): print( "Destination language not supported. " "Run with --list-languages to see all supported languages." ) return False if not args.source_path: print("Error: You need to specify a source path.") return False return True def main(): """ Run autosub as a command-line program. """ parser = argparse.ArgumentParser() parser.add_argument('source_path', help="Path to the video or audio file to subtitle", nargs='?') parser.add_argument('-C', '--concurrency', help="Number of concurrent API requests to make", type=int, default=DEFAULT_CONCURRENCY) parser.add_argument('-o', '--output', help="Output path for subtitles (by default, subtitles are saved in \ the same directory and name as the source path)") parser.add_argument('-F', '--format', help="Destination subtitle format", default=DEFAULT_SUBTITLE_FORMAT) parser.add_argument('-S', '--src-language', help="Language spoken in source file", default=DEFAULT_SRC_LANGUAGE) parser.add_argument('-D', '--dst-language', help="Desired language for the subtitles", default=DEFAULT_DST_LANGUAGE) parser.add_argument('-K', '--api-key', help="The Google Translate API key to be used. \ (Required for subtitle translation)") parser.add_argument('--list-formats', help="List all available subtitle formats", action='store_true') parser.add_argument('--list-languages', help="List all available source/destination languages", action='store_true') parser.add_argument('--min_height', help="minimum height from 0 - 100%", type=float, default=93) parser.add_argument('--max_height', help="maximum height from 0 - 100%", type=float, default=99) parser.add_argument('--l_v', help="Light sensitive", type=float, default=210) parser.add_argument('--debug', help="Allows to show cropped image on the desktop", action='store_true', default=True) parser.add_argument('--cloud', help="Use google cloud compute to extract text", action='store_true', default=False) parser.add_argument('--disable_time', help="Parse time function", action='store_true') parser.add_argument('--all', help="Render all files", action='store_true') args = parser.parse_args() if args.list_formats: print(
"List of formats:") for subtitle_format in FORMATTERS: print("{format}".format(format=subtitle_format)) return 0
conditional_block
socketclient.go
nil } func (c *Client) disconnect() error { log.Debugf("Closing socket") // cleanup msg table c.setMsgTable(make(map[string]uint16), 0) if err := c.conn.Close(); err != nil { log.Debugln("Closing socket failed:", err) return err } return nil } const ( sockCreateMsgId = 15 // hard-coded sockclnt_create message ID createMsgContext = byte(123) deleteMsgContext = byte(124) ) func (c *Client) open() error { var msgCodec = codec.DefaultCodec // Request socket client create req := &memclnt.SockclntCreate{ Name: c.clientName, } msg, err := msgCodec.EncodeMsg(req, sockCreateMsgId) if err != nil { log.Debugln("Encode error:", err) return err } // set non-0 context msg[5] = createMsgContext if err := c.writeMsg(msg); err != nil { log.Debugln("Write error: ", err) return err } msgReply, err := c.readMsgTimeout(nil, c.connectTimeout) if err != nil { log.Println("Read error:", err) return err } reply := new(memclnt.SockclntCreateReply) if err := msgCodec.DecodeMsg(msgReply, reply); err != nil { log.Println("Decoding sockclnt_create_reply failed:", err) return err } else if reply.Response != 0 { return fmt.Errorf("sockclnt_create_reply: response error (%d)", reply.Response) } log.Debugf("SockclntCreateReply: Response=%v Index=%v Count=%v", reply.Response, reply.Index, reply.Count) c.clientIndex = reply.Index msgTable := make(map[string]uint16, reply.Count) var sockDelMsgId uint16 for _, x := range reply.MessageTable { msgName := strings.Split(x.Name, "\x00")[0] name := strings.TrimSuffix(msgName, "\x13") msgTable[name] = x.Index if strings.HasPrefix(name, "sockclnt_delete_") { sockDelMsgId = x.Index } if debugMsgIds { log.Debugf(" - %4d: %q", x.Index, name) } } c.setMsgTable(msgTable, sockDelMsgId) return nil } func (c *Client) setMsgTable(msgTable map[string]uint16, sockDelMsgId uint16) { c.msgTableMu.Lock() defer c.msgTableMu.Unlock() c.msgTable = msgTable c.sockDelMsgId = sockDelMsgId } func (c *Client) GetMsgID(msgName string, msgCrc string) (uint16, error) { c.msgTableMu.RLock() defer c.msgTableMu.RUnlock() if msgID, ok := c.msgTable[msgName+"_"+msgCrc]; ok { return msgID, nil } return 0, &adapter.UnknownMsgError{ MsgName: msgName, MsgCrc: msgCrc, } } func (c *Client) SendMsg(context uint32, data []byte) error { if len(data) < 10 { return fmt.Errorf("invalid message data, length must be at least 10 bytes") } setMsgRequestHeader(data, c.clientIndex, context) if debug { log.Debugf("sendMsg (%d) context=%v client=%d: % 02X", len(data), context, c.clientIndex, data) } if err := c.writeMsg(data); err != nil { log.Debugln("writeMsg error: ", err) return err } return nil } // setMsgRequestHeader sets client index and context in the message request header // // Message request has following structure: // // type msgRequestHeader struct { // MsgID uint16 // ClientIndex uint32 // Context uint32 // } func setMsgRequestHeader(data []byte, clientIndex, context uint32) { // message ID is already set binary.BigEndian.PutUint32(data[2:6], clientIndex) binary.BigEndian.PutUint32(data[6:10], context) } func (c *Client) writeMsg(msg []byte) error { // we lock to prevent mixing multiple message writes c.writeMu.Lock() defer c.writeMu.Unlock() header, ok := c.headerPool.Get().(*[]byte) if !ok { return fmt.Errorf("failed to get header from pool") } err := writeMsgHeader(c.writer, *header, len(msg)) if err != nil { return err } c.headerPool.Put(header) if err := writeMsgData(c.writer, msg, c.writer.Size()); err != nil { return err } if err := c.writer.Flush(); err != nil { return err } log.Debugf(" -- writeMsg done") return nil } func writeMsgHeader(w io.Writer, header []byte, dataLen int) error { binary.BigEndian.PutUint32(header[8:12], uint32(dataLen)) n, err := w.Write(header) if err != nil { return err } if debug { log.Debugf(" - header sent (%d/%d): % 0X", n, len(header), header) } return nil } func writeMsgData(w io.Writer, msg []byte, writerSize int) error { for i := 0; i <= len(msg)/writerSize; i++ { x := i*writerSize + writerSize if x > len(msg) { x = len(msg) } if debug { log.Debugf(" - x=%v i=%v len=%v mod=%v", x, i, len(msg), len(msg)/writerSize) } n, err := w.Write(msg[i*writerSize : x]) if err != nil { return err } if debug { log.Debugf(" - data sent x=%d (%d/%d): % 0X", x, n, len(msg), msg) } } return nil } func (c *Client) readerLoop() { defer c.wg.Done() defer log.Debugf("reader loop done") var buf [8192]byte for { select { case <-c.quit: return default: } msg, err := c.readMsg(buf[:]) if err != nil { if isClosedError(err) { return } log.Debugf("readMsg error: %v", err) continue } msgID, context := getMsgReplyHeader(msg) if debug { log.Debugf("recvMsg (%d) msgID=%d context=%v", len(msg), msgID, context) } c.msgCallback(msgID, msg) } } // getMsgReplyHeader gets message ID and context from the message reply header // // Message reply has the following structure: // // type msgReplyHeader struct { // MsgID uint16 // Context uint32 // } func getMsgReplyHeader(msg []byte) (msgID uint16, context uint32) { msgID = binary.BigEndian.Uint16(msg[0:2]) context = binary.BigEndian.Uint32(msg[2:6]) return } func (c *Client) readMsgTimeout(buf []byte, timeout time.Duration) ([]byte, error) { // set read deadline readDeadline := time.Now().Add(timeout) if err := c.conn.SetReadDeadline(readDeadline); err != nil { return nil, err } // read message msgReply, err := c.readMsg(buf) if err != nil { return nil, err } // reset read deadline if err := c.conn.SetReadDeadline(time.Time{}); err != nil { return nil, err } return msgReply, nil } func (c *Client) readMsg(buf []byte) ([]byte, error) { log.Debug("reading msg..") header, ok := c.headerPool.Get().(*[]byte) if !ok { return nil, fmt.Errorf("failed to get header from pool") } msgLen, err := readMsgHeader(c.reader, *header) if err != nil { return nil, err } c.headerPool.Put(header) msg, err := readMsgData(c.reader, buf, msgLen) if err != nil { return nil, err } log.Debugf(" -- readMsg done (buffered: %d)", c.reader.Buffered()) return msg, nil } func readMsgHeader(r io.Reader, header []byte) (int, error)
{ n, err := io.ReadAtLeast(r, header, 16) if err != nil { return 0, err } if n == 0 { log.Debugln("zero bytes header") return 0, nil } else if n != 16 { log.Debugf("invalid header (%d bytes): % 0X", n, header[:n]) return 0, fmt.Errorf("invalid header (expected 16 bytes, got %d)", n) } dataLen := binary.BigEndian.Uint32(header[8:12]) return int(dataLen), nil }
identifier_body
socketclient.go
// SetClientName sets a client name used for identification. func (c *Client) SetClientName(name string) { c.clientName = name } // SetConnectTimeout sets timeout used during connecting. func (c *Client) SetConnectTimeout(t time.Duration) { c.connectTimeout = t } // SetDisconnectTimeout sets timeout used during disconnecting. func (c *Client) SetDisconnectTimeout(t time.Duration) { c.disconnectTimeout = t } // SetMsgCallback sets the callback for incoming messages. func (c *Client) SetMsgCallback(cb adapter.MsgCallback) { log.Debug("SetMsgCallback") c.msgCallback = cb } // WaitReady checks if the socket file exists and if it does not exist waits for // it for the duration defined by MaxWaitReady. func (c *Client) WaitReady() error { socketDir, _ := filepath.Split(c.socketPath) dirChain := strings.Split(filepath.ToSlash(filepath.Clean(socketDir)), "/") dir := "/" for _, dirElem := range dirChain { dir = filepath.Join(dir, dirElem) if err := waitForDir(dir); err != nil { return err } log.Debugf("dir ready: %v", dir) } // check if socket already exists if _, err := os.Stat(c.socketPath); err == nil { return nil // socket exists, we are ready } else if !errors.Is(err, fs.ErrNotExist) { log.Debugf("error is: %+v", err) return err // some other error occurred } log.Debugf("waiting for file: %v", c.socketPath) // socket does not exist, watch for it watcher, err := fsnotify.NewWatcher() if err != nil { return err } defer func() { if err := watcher.Close(); err != nil { log.Debugf("failed to close file watcher: %v", err) } }() // start directory watcher d := filepath.Dir(c.socketPath) if err := watcher.Add(d); err != nil { log.Debugf("watcher add(%v) error: %v", d, err) return err } timeout := time.NewTimer(MaxWaitReady) defer timeout.Stop() for { select { case <-timeout.C: log.Debugf("watcher timeout after: %v", MaxWaitReady) return fmt.Errorf("timeout waiting (%s) for socket file: %s", MaxWaitReady, c.socketPath) case e := <-watcher.Errors: log.Debugf("watcher error: %+v", e) return e case ev := <-watcher.Events: log.Debugf("watcher event: %+v", ev) if ev.Name == c.socketPath && (ev.Op&fsnotify.Create) == fsnotify.Create { // socket created, we are ready return nil } } } } func waitForDir(dir string) error { // check if dir already exists if _, err := os.Stat(dir); err == nil { return nil // dir exists, we are ready } else if !errors.Is(err, fs.ErrNotExist) { log.Debugf("error is: %+v", err) return err // some other error occurred } log.Debugf("waiting for dir: %v", dir) // dir does not exist, watch for it watcher, err := fsnotify.NewWatcher() if err != nil { return err } defer func() { if err := watcher.Close(); err != nil { log.Debugf("failed to close file watcher: %v", err) } }() // start watching directory d := filepath.Dir(dir) if err := watcher.Add(d); err != nil { log.Debugf("watcher add (%v) error: %v", d, err) return err } timeout := time.NewTimer(MaxWaitReady) defer timeout.Stop() for { select { case <-timeout.C: log.Debugf("watcher timeout after: %v", MaxWaitReady) return fmt.Errorf("timeout waiting (%s) for directory: %s", MaxWaitReady, dir) case e := <-watcher.Errors: log.Debugf("watcher error: %+v", e) return e case ev := <-watcher.Events: log.Debugf("watcher event: %+v", ev) if ev.Name == dir && (ev.Op&fsnotify.Create) == fsnotify.Create { // socket created, we are ready return nil } } } } func (c *Client) Connect() error { // check if socket exists if _, err := os.Stat(c.socketPath); os.IsNotExist(err) { return fmt.Errorf("VPP API socket file %s does not exist", c.socketPath) } else if err != nil { return fmt.Errorf("VPP API socket error: %v", err) } if err := c.connect(c.socketPath); err != nil { return err } if err := c.open(); err != nil { _ = c.disconnect() return err } c.quit = make(chan struct{}) c.wg.Add(1) go c.readerLoop() return nil } func (c *Client) Disconnect() error { if c.conn == nil { return nil } log.Debugf("Disconnecting..") close(c.quit) if err := c.conn.CloseRead(); err != nil { log.Debugf("closing readMsg failed: %v", err) } // wait for readerLoop to return c.wg.Wait() // Don't bother sending a vl_api_sockclnt_delete_t message, // just close the socket. if err := c.disconnect(); err != nil { return err } return nil } const defaultBufferSize = 4096 func (c *Client) connect(sockAddr string) error { addr := &net.UnixAddr{Name: sockAddr, Net: "unix"} log.Debugf("Connecting to: %v", c.socketPath) conn, err := net.DialUnix("unix", nil, addr) if err != nil { // we try different type of socket for backwards compatbility with VPP<=19.04 if strings.Contains(err.Error(), "wrong type for socket") { addr.Net = "unixpacket" log.Debugf("%s, retrying connect with type unixpacket", err) conn, err = net.DialUnix("unixpacket", nil, addr) } if err != nil { log.Debugf("Connecting to socket %s failed: %s", addr, err) return err } } c.conn = conn log.Debugf("Connected to socket (local addr: %v)", c.conn.LocalAddr().(*net.UnixAddr)) c.reader = bufio.NewReaderSize(c.conn, defaultBufferSize) c.writer = bufio.NewWriterSize(c.conn, defaultBufferSize) return nil } func (c *Client) disconnect() error { log.Debugf("Closing socket") // cleanup msg table c.setMsgTable(make(map[string]uint16), 0) if err := c.conn.Close(); err != nil { log.Debugln("Closing socket failed:", err) return err } return nil } const ( sockCreateMsgId = 15 // hard-coded sockclnt_create message ID createMsgContext = byte(123) deleteMsgContext = byte(124)
func (c *Client) open() error { var msgCodec = codec.DefaultCodec // Request socket client create req := &memclnt.SockclntCreate{ Name: c.clientName, } msg, err := msgCodec.EncodeMsg(req, sockCreateMsgId) if err != nil { log.Debugln("Encode error:", err) return err } // set non-0 context msg[5] = createMsgContext if err := c.writeMsg(msg); err != nil { log.Debugln("Write error: ", err) return err } msgReply, err := c.readMsgTimeout(nil, c.connectTimeout) if err != nil { log.Println("Read error:", err) return err } reply := new(memclnt.SockclntCreateReply) if err := msgCodec.DecodeMsg(msgReply, reply); err != nil { log.Println("Decoding sockclnt_create_reply failed:", err) return err } else if reply.Response != 0 { return fmt.Errorf("sockclnt_create_reply: response error (%d)", reply.Response) } log.Debugf("SockclntCreateReply: Response=%v Index=%v Count=%v", reply.Response, reply.Index, reply.Count) c.clientIndex = reply.Index msgTable := make(map[string]uint16, reply.Count) var sockDelMsgId uint16 for _, x := range reply.MessageTable { msgName := strings.Split(x.Name, "\x00")[0] name := strings.TrimSuffix(msgName, "\x13") msgTable[name] = x.Index if strings.HasPrefix(name, "sockclnt_delete_") { sockDelMsgId = x.Index } if debugMsgIds { log.Debugf(" - %4d: %q", x.Index, name) } } c.setMsgTable(msgTable, sockDelMsgId) return nil } func (c *Client) setMsgTable(msgTable map[string]uint16, sockDelMsgId uint16) { c.msgTableMu.Lock() defer c.msgTableMu
)
random_line_split
socketclient.go
(logger logrus.FieldLogger) { log = logger } func init() { logger := logrus.New() if debug { logger.Level = logrus.DebugLevel logger.Debug("govpp: debug level enabled for socketclient") } log = logger.WithField("logger", "govpp/socketclient") } type Client struct { socketPath string clientName string conn *net.UnixConn reader *bufio.Reader writer *bufio.Writer connectTimeout time.Duration disconnectTimeout time.Duration msgCallback adapter.MsgCallback clientIndex uint32 msgTable map[string]uint16 msgTableMu sync.RWMutex sockDelMsgId uint16 writeMu sync.Mutex headerPool *sync.Pool quit chan struct{} wg sync.WaitGroup } // NewVppClient returns a new Client using socket. // If socket is empty string DefaultSocketName is used. func NewVppClient(socket string) *Client { if socket == "" { socket = DefaultSocketName } return &Client{ socketPath: socket, clientName: DefaultClientName, connectTimeout: DefaultConnectTimeout, disconnectTimeout: DefaultDisconnectTimeout, headerPool: &sync.Pool{New: func() interface{} { x := make([]byte, 16) return &x }}, msgCallback: func(msgID uint16, data []byte) { log.Debugf("no callback set, dropping message: ID=%v len=%d", msgID, len(data)) }, } } // SetClientName sets a client name used for identification. func (c *Client) SetClientName(name string) { c.clientName = name } // SetConnectTimeout sets timeout used during connecting. func (c *Client) SetConnectTimeout(t time.Duration) { c.connectTimeout = t } // SetDisconnectTimeout sets timeout used during disconnecting. func (c *Client) SetDisconnectTimeout(t time.Duration) { c.disconnectTimeout = t } // SetMsgCallback sets the callback for incoming messages. func (c *Client) SetMsgCallback(cb adapter.MsgCallback) { log.Debug("SetMsgCallback") c.msgCallback = cb } // WaitReady checks if the socket file exists and if it does not exist waits for // it for the duration defined by MaxWaitReady. func (c *Client) WaitReady() error { socketDir, _ := filepath.Split(c.socketPath) dirChain := strings.Split(filepath.ToSlash(filepath.Clean(socketDir)), "/") dir := "/" for _, dirElem := range dirChain { dir = filepath.Join(dir, dirElem) if err := waitForDir(dir); err != nil { return err } log.Debugf("dir ready: %v", dir) } // check if socket already exists if _, err := os.Stat(c.socketPath); err == nil { return nil // socket exists, we are ready } else if !errors.Is(err, fs.ErrNotExist) { log.Debugf("error is: %+v", err) return err // some other error occurred } log.Debugf("waiting for file: %v", c.socketPath) // socket does not exist, watch for it watcher, err := fsnotify.NewWatcher() if err != nil { return err } defer func() { if err := watcher.Close(); err != nil { log.Debugf("failed to close file watcher: %v", err) } }() // start directory watcher d := filepath.Dir(c.socketPath) if err := watcher.Add(d); err != nil { log.Debugf("watcher add(%v) error: %v", d, err) return err } timeout := time.NewTimer(MaxWaitReady) defer timeout.Stop() for { select { case <-timeout.C: log.Debugf("watcher timeout after: %v", MaxWaitReady) return fmt.Errorf("timeout waiting (%s) for socket file: %s", MaxWaitReady, c.socketPath) case e := <-watcher.Errors: log.Debugf("watcher error: %+v", e) return e case ev := <-watcher.Events: log.Debugf("watcher event: %+v", ev) if ev.Name == c.socketPath && (ev.Op&fsnotify.Create) == fsnotify.Create { // socket created, we are ready return nil } } } } func waitForDir(dir string) error { // check if dir already exists if _, err := os.Stat(dir); err == nil { return nil // dir exists, we are ready } else if !errors.Is(err, fs.ErrNotExist) { log.Debugf("error is: %+v", err) return err // some other error occurred } log.Debugf("waiting for dir: %v", dir) // dir does not exist, watch for it watcher, err := fsnotify.NewWatcher() if err != nil { return err } defer func() { if err := watcher.Close(); err != nil { log.Debugf("failed to close file watcher: %v", err) } }() // start watching directory d := filepath.Dir(dir) if err := watcher.Add(d); err != nil { log.Debugf("watcher add (%v) error: %v", d, err) return err } timeout := time.NewTimer(MaxWaitReady) defer timeout.Stop() for { select { case <-timeout.C: log.Debugf("watcher timeout after: %v", MaxWaitReady) return fmt.Errorf("timeout waiting (%s) for directory: %s", MaxWaitReady, dir) case e := <-watcher.Errors: log.Debugf("watcher error: %+v", e) return e case ev := <-watcher.Events: log.Debugf("watcher event: %+v", ev) if ev.Name == dir && (ev.Op&fsnotify.Create) == fsnotify.Create { // socket created, we are ready return nil } } } } func (c *Client) Connect() error { // check if socket exists if _, err := os.Stat(c.socketPath); os.IsNotExist(err) { return fmt.Errorf("VPP API socket file %s does not exist", c.socketPath) } else if err != nil { return fmt.Errorf("VPP API socket error: %v", err) } if err := c.connect(c.socketPath); err != nil { return err } if err := c.open(); err != nil { _ = c.disconnect() return err } c.quit = make(chan struct{}) c.wg.Add(1) go c.readerLoop() return nil } func (c *Client) Disconnect() error { if c.conn == nil { return nil } log.Debugf("Disconnecting..") close(c.quit) if err := c.conn.CloseRead(); err != nil { log.Debugf("closing readMsg failed: %v", err) } // wait for readerLoop to return c.wg.Wait() // Don't bother sending a vl_api_sockclnt_delete_t message, // just close the socket. if err := c.disconnect(); err != nil { return err } return nil } const defaultBufferSize = 4096 func (c *Client) connect(sockAddr string) error { addr := &net.UnixAddr{Name: sockAddr, Net: "unix"} log.Debugf("Connecting to: %v", c.socketPath) conn, err := net.DialUnix("unix", nil, addr) if err != nil { // we try different type of socket for backwards compatbility with VPP<=19.04 if strings.Contains(err.Error(), "wrong type for socket") { addr.Net = "unixpacket" log.Debugf("%s, retrying connect with type unixpacket", err) conn, err = net.DialUnix("unixpacket", nil, addr) } if err != nil { log.Debugf("Connecting to socket %s failed: %s", addr, err) return err } } c.conn = conn log.Debugf("Connected to socket (local addr: %v)", c.conn.LocalAddr().(*net.UnixAddr)) c.reader = bufio.NewReaderSize(c.conn, defaultBufferSize) c.writer = bufio.NewWriterSize(c.conn, defaultBufferSize) return nil } func (c *Client) disconnect() error { log.Debugf("Closing socket") // cleanup msg table c.setMsgTable(make(map[string]uint16), 0) if err := c.conn.Close(); err != nil { log.Debugln("Closing socket failed:", err) return err } return nil } const ( sockCreateMsgId = 15 // hard-coded sockclnt_create message ID createMsgContext = byte(123) deleteMsgContext = byte(124) ) func (c *Client) open() error { var msgCodec = codec.DefaultCodec // Request socket client create req := &memclnt.SockclntCreate{ Name: c.clientName, } msg, err := msgCodec.EncodeMsg(req, sockCreateMsgId) if err != nil { log.Debugln("Encode error:", err) return err } // set non-0 context msg[5] = createMsgContext if err := c.writeMsg(msg); err != nil { log.Debugln("Write
SetLogger
identifier_name
socketclient.go
// SetClientName sets a client name used for identification. func (c *Client) SetClientName(name string) { c.clientName = name } // SetConnectTimeout sets timeout used during connecting. func (c *Client) SetConnectTimeout(t time.Duration) { c.connectTimeout = t } // SetDisconnectTimeout sets timeout used during disconnecting. func (c *Client) SetDisconnectTimeout(t time.Duration) { c.disconnectTimeout = t } // SetMsgCallback sets the callback for incoming messages. func (c *Client) SetMsgCallback(cb adapter.MsgCallback) { log.Debug("SetMsgCallback") c.msgCallback = cb } // WaitReady checks if the socket file exists and if it does not exist waits for // it for the duration defined by MaxWaitReady. func (c *Client) WaitReady() error { socketDir, _ := filepath.Split(c.socketPath) dirChain := strings.Split(filepath.ToSlash(filepath.Clean(socketDir)), "/") dir := "/" for _, dirElem := range dirChain { dir = filepath.Join(dir, dirElem) if err := waitForDir(dir); err != nil { return err } log.Debugf("dir ready: %v", dir) } // check if socket already exists if _, err := os.Stat(c.socketPath); err == nil { return nil // socket exists, we are ready } else if !errors.Is(err, fs.ErrNotExist) { log.Debugf("error is: %+v", err) return err // some other error occurred } log.Debugf("waiting for file: %v", c.socketPath) // socket does not exist, watch for it watcher, err := fsnotify.NewWatcher() if err != nil { return err } defer func() { if err := watcher.Close(); err != nil { log.Debugf("failed to close file watcher: %v", err) } }() // start directory watcher d := filepath.Dir(c.socketPath) if err := watcher.Add(d); err != nil { log.Debugf("watcher add(%v) error: %v", d, err) return err } timeout := time.NewTimer(MaxWaitReady) defer timeout.Stop() for { select { case <-timeout.C: log.Debugf("watcher timeout after: %v", MaxWaitReady) return fmt.Errorf("timeout waiting (%s) for socket file: %s", MaxWaitReady, c.socketPath) case e := <-watcher.Errors: log.Debugf("watcher error: %+v", e) return e case ev := <-watcher.Events: log.Debugf("watcher event: %+v", ev) if ev.Name == c.socketPath && (ev.Op&fsnotify.Create) == fsnotify.Create { // socket created, we are ready return nil } } } } func waitForDir(dir string) error { // check if dir already exists if _, err := os.Stat(dir); err == nil { return nil // dir exists, we are ready } else if !errors.Is(err, fs.ErrNotExist) { log.Debugf("error is: %+v", err) return err // some other error occurred } log.Debugf("waiting for dir: %v", dir) // dir does not exist, watch for it watcher, err := fsnotify.NewWatcher() if err != nil { return err } defer func() { if err := watcher.Close(); err != nil { log.Debugf("failed to close file watcher: %v", err) } }() // start watching directory d := filepath.Dir(dir) if err := watcher.Add(d); err != nil { log.Debugf("watcher add (%v) error: %v", d, err) return err } timeout := time.NewTimer(MaxWaitReady) defer timeout.Stop() for { select { case <-timeout.C: log.Debugf("watcher timeout after: %v", MaxWaitReady) return fmt.Errorf("timeout waiting (%s) for directory: %s", MaxWaitReady, dir) case e := <-watcher.Errors: log.Debugf("watcher error: %+v", e) return e case ev := <-watcher.Events: log.Debugf("watcher event: %+v", ev) if ev.Name == dir && (ev.Op&fsnotify.Create) == fsnotify.Create { // socket created, we are ready return nil } } } } func (c *Client) Connect() error { // check if socket exists if _, err := os.Stat(c.socketPath); os.IsNotExist(err) { return fmt.Errorf("VPP API socket file %s does not exist", c.socketPath) } else if err != nil { return fmt.Errorf("VPP API socket error: %v", err) } if err := c.connect(c.socketPath); err != nil { return err } if err := c.open(); err != nil { _ = c.disconnect() return err } c.quit = make(chan struct{}) c.wg.Add(1) go c.readerLoop() return nil } func (c *Client) Disconnect() error { if c.conn == nil { return nil } log.Debugf("Disconnecting..") close(c.quit) if err := c.conn.CloseRead(); err != nil { log.Debugf("closing readMsg failed: %v", err) } // wait for readerLoop to return c.wg.Wait() // Don't bother sending a vl_api_sockclnt_delete_t message, // just close the socket. if err := c.disconnect(); err != nil { return err } return nil } const defaultBufferSize = 4096 func (c *Client) connect(sockAddr string) error { addr := &net.UnixAddr{Name: sockAddr, Net: "unix"} log.Debugf("Connecting to: %v", c.socketPath) conn, err := net.DialUnix("unix", nil, addr) if err != nil { // we try different type of socket for backwards compatbility with VPP<=19.04 if strings.Contains(err.Error(), "wrong type for socket") { addr.Net = "unixpacket" log.Debugf("%s, retrying connect with type unixpacket", err) conn, err = net.DialUnix("unixpacket", nil, addr) } if err != nil { log.Debugf("Connecting to socket %s failed: %s", addr, err) return err } } c.conn = conn log.Debugf("Connected to socket (local addr: %v)", c.conn.LocalAddr().(*net.UnixAddr)) c.reader = bufio.NewReaderSize(c.conn, defaultBufferSize) c.writer = bufio.NewWriterSize(c.conn, defaultBufferSize) return nil } func (c *Client) disconnect() error { log.Debugf("Closing socket") // cleanup msg table c.setMsgTable(make(map[string]uint16), 0) if err := c.conn.Close(); err != nil { log.Debugln("Closing socket failed:", err) return err } return nil } const ( sockCreateMsgId = 15 // hard-coded sockclnt_create message ID createMsgContext = byte(123) deleteMsgContext = byte(124) ) func (c *Client) open() error { var msgCodec = codec.DefaultCodec // Request socket client create req := &memclnt.SockclntCreate{ Name: c.clientName, } msg, err := msgCodec.EncodeMsg(req, sockCreateMsgId) if err != nil { log.Debugln("Encode error:", err) return err } // set non-0 context msg[5] = createMsgContext if err := c.writeMsg(msg); err != nil { log.Debugln("Write error: ", err) return err } msgReply, err := c.readMsgTimeout(nil, c.connectTimeout) if err != nil { log.Println("Read error:", err) return err } reply := new(memclnt.SockclntCreateReply) if err := msgCodec.DecodeMsg(msgReply, reply); err != nil { log.Println("Decoding sockclnt_create_reply failed:", err) return err } else if reply.Response != 0
log.Debugf("SockclntCreateReply: Response=%v Index=%v Count=%v", reply.Response, reply.Index, reply.Count) c.clientIndex = reply.Index msgTable := make(map[string]uint16, reply.Count) var sockDelMsgId uint16 for _, x := range reply.MessageTable { msgName := strings.Split(x.Name, "\x00")[0] name := strings.TrimSuffix(msgName, "\x13") msgTable[name] = x.Index if strings.HasPrefix(name, "sockclnt_delete_") { sockDelMsgId = x.Index } if debugMsgIds { log.Debugf(" - %4d: %q", x.Index, name) } } c.setMsgTable(msgTable, sockDelMsgId) return nil } func (c *Client) setMsgTable(msgTable map[string]uint16, sockDelMsgId uint16) { c.msgTableMu.Lock() defer c.msgTable
{ return fmt.Errorf("sockclnt_create_reply: response error (%d)", reply.Response) }
conditional_block
tx.go
(now time.Time) { tx.now = now } // CreateMappers will create a set of mappers that need to be run to execute the map phase of a MapReduceJob. func (tx *tx) CreateMapReduceJobs(stmt *influxql.SelectStatement, tagKeys []string) ([]*influxql.MapReduceJob, error) { jobs := []*influxql.MapReduceJob{} for _, src := range stmt.Sources { mm, ok := src.(*influxql.Measurement) if !ok { return nil, fmt.Errorf("invalid source type: %#v", src) } // get the index and the retention policy rp, err := tx.meta.RetentionPolicy(mm.Database, mm.RetentionPolicy) if err != nil { return nil, err } m := tx.store.Measurement(mm.Database, mm.Name) if m == nil { return nil, ErrMeasurementNotFound(influxql.QuoteIdent([]string{mm.Database, "", mm.Name}...)) } tx.measurement = m // Validate the fields and tags asked for exist and keep track of which are in the select vs the where var selectFields []string var whereFields []string var selectTags []string for _, n := range stmt.NamesInSelect() { if m.HasField(n) { selectFields = append(selectFields, n) continue } if !m.HasTagKey(n) { return nil, fmt.Errorf("unknown field or tag name in select clause: %s", n) } selectTags = append(selectTags, n) tagKeys = append(tagKeys, n) } for _, n := range stmt.NamesInWhere() { if n == "time" { continue } if m.HasField(n) { whereFields = append(whereFields, n) continue } if !m.HasTagKey(n) { return nil, fmt.Errorf("unknown field or tag name in where clause: %s", n) } } if len(selectFields) == 0 && len(stmt.FunctionCalls()) == 0 { return nil, fmt.Errorf("select statement must include at least one field or function call") } // Validate that group by is not a field for _, d := range stmt.Dimensions { switch e := d.Expr.(type) { case *influxql.VarRef: if !m.HasTagKey(e.Val) { return nil, fmt.Errorf("can not use field in group by clause: %s", e.Val) } } } // Grab time range from statement. tmin, tmax := influxql.TimeRange(stmt.Condition) if tmax.IsZero() { tmax = tx.now } if tmin.IsZero() { tmin = time.Unix(0, 0) } // Find shard groups within time range. var shardGroups []*meta.ShardGroupInfo for _, group := range rp.ShardGroups { if group.Overlaps(tmin, tmax) { g := group shardGroups = append(shardGroups, &g) } } if len(shardGroups) == 0 { return nil, nil } // get the group by interval, if there is one var interval int64 if d, err := stmt.GroupByInterval(); err != nil { return nil, err } else { interval = d.Nanoseconds() } // get the sorted unique tag sets for this query. tagSets, err := m.TagSets(stmt, tagKeys) if err != nil { return nil, err } for _, t := range tagSets { // make a job for each tagset job := &influxql.MapReduceJob{ MeasurementName: m.Name, TagSet: t, TMin: tmin.UnixNano(), TMax: tmax.UnixNano(), } // make a mapper for each shard that must be hit. We may need to hit multiple shards within a shard group var mappers []influxql.Mapper // create mappers for each shard we need to hit for _, sg := range shardGroups { // TODO: implement distributed queries if len(sg.Shards) != 1 { return nil, fmt.Errorf("distributed queries aren't supported yet. You have a replication policy with RF < # of servers in cluster") } shard := tx.store.Shard(sg.Shards[0].ID) if shard == nil { // the store returned nil which means we haven't written any data into this shard yet, so ignore it continue } // get the codec for this measuremnt. If this is nil it just means this measurement was // never written into this shard, so we can skip it and continue. codec := shard.FieldCodec(m.Name) if codec == nil { continue } var mapper influxql.Mapper mapper = &LocalMapper{ seriesKeys: t.SeriesKeys, shard: shard, db: shard.DB(), job: job, decoder: codec, filters: t.Filters, whereFields: whereFields, selectFields: selectFields, selectTags: selectTags, tmin: tmin.UnixNano(), tmax: tmax.UnixNano(), interval: interval, // multiple mappers may need to be merged together to get the results // for a raw query. So each mapper will have to read at least the // limit plus the offset in data points to ensure we've hit our mark limit: uint64(stmt.Limit) + uint64(stmt.Offset), } mappers = append(mappers, mapper) } job.Mappers = mappers jobs = append(jobs, job) } } // always return them in sorted order so the results from running the jobs are returned in a deterministic order sort.Sort(influxql.MapReduceJobs(jobs)) return jobs, nil } // LocalMapper implements the influxql.Mapper interface for running map tasks over a shard that is local to this server type LocalMapper struct { cursorsEmpty bool // boolean that lets us know if the cursors are empty decoder *FieldCodec // decoder for the raw data bytes filters []influxql.Expr // filters for each series cursors []*shardCursor // bolt cursors for each series id seriesKeys []string // seriesKeys to be read from this shard shard *Shard // original shard db *bolt.DB // bolt store for the shard accessed by this mapper txn *bolt.Tx // read transactions by shard id job *influxql.MapReduceJob // the MRJob this mapper belongs to mapFunc influxql.MapFunc // the map func fieldID uint8 // the field ID associated with the mapFunc curently being run fieldName string // the field name associated with the mapFunc currently being run keyBuffer []int64 // the current timestamp key for each cursor valueBuffer [][]byte // the current value for each cursor tmin int64 // the min of the current group by interval being iterated over tmax int64 // the max of the current group by interval being iterated over additionalNames []string // additional field or tag names that might be requested from the map function whereFields []string // field names that occur in the where clause selectFields []string // field names that occur in the select clause selectTags []string // tag keys that occur in the select clause isRaw bool // if the query is a non-aggregate query interval int64 // the group by interval of the query, if any limit uint64 // used for raw queries for LIMIT perIntervalLimit int // used for raw queries to determine how far into a chunk we are chunkSize int // used for raw queries to determine how much data to read before flushing to client } // Open opens the LocalMapper. func (l *LocalMapper) Open() error { // Obtain shard lock to copy in-cache points. l.shard.mu.Lock() defer l.shard.mu.Unlock() // Open the data store txn, err := l.db.Begin(false) if err != nil { return err } l.txn = txn // create a bolt cursor for each unique series id l.cursors = make([]*shardCursor, len(l.seriesKeys)) for i, key := range l.seriesKeys { // Retrieve key bucket. b := l.txn.Bucket([]byte(key)) // Ignore if there is no bucket or points in the cache. partitionID := WALPartition([]byte(key)) if b == nil && len(l.shard.cache[partitionID][key]) == 0 { continue } // Retrieve a copy of the in-cache
SetNow
identifier_name
tx.go
MRJob this mapper belongs to mapFunc influxql.MapFunc // the map func fieldID uint8 // the field ID associated with the mapFunc curently being run fieldName string // the field name associated with the mapFunc currently being run keyBuffer []int64 // the current timestamp key for each cursor valueBuffer [][]byte // the current value for each cursor tmin int64 // the min of the current group by interval being iterated over tmax int64 // the max of the current group by interval being iterated over additionalNames []string // additional field or tag names that might be requested from the map function whereFields []string // field names that occur in the where clause selectFields []string // field names that occur in the select clause selectTags []string // tag keys that occur in the select clause isRaw bool // if the query is a non-aggregate query interval int64 // the group by interval of the query, if any limit uint64 // used for raw queries for LIMIT perIntervalLimit int // used for raw queries to determine how far into a chunk we are chunkSize int // used for raw queries to determine how much data to read before flushing to client } // Open opens the LocalMapper. func (l *LocalMapper) Open() error { // Obtain shard lock to copy in-cache points. l.shard.mu.Lock() defer l.shard.mu.Unlock() // Open the data store txn, err := l.db.Begin(false) if err != nil { return err } l.txn = txn // create a bolt cursor for each unique series id l.cursors = make([]*shardCursor, len(l.seriesKeys)) for i, key := range l.seriesKeys { // Retrieve key bucket. b := l.txn.Bucket([]byte(key)) // Ignore if there is no bucket or points in the cache. partitionID := WALPartition([]byte(key)) if b == nil && len(l.shard.cache[partitionID][key]) == 0 { continue } // Retrieve a copy of the in-cache points for the key. cache := make([][]byte, len(l.shard.cache[partitionID][key])) copy(cache, l.shard.cache[partitionID][key]) // Build a cursor that merges the bucket and cache together. cur := &shardCursor{cache: cache} if b != nil { cur.cursor = b.Cursor() } l.cursors[i] = cur } return nil } // Close closes the LocalMapper. func (l *LocalMapper) Close() { if l.txn != nil { _ = l.txn.Rollback() } } // Begin will set up the mapper to run the map function for a given aggregate call starting at the passed in time func (l *LocalMapper) Begin(c *influxql.Call, startingTime int64, chunkSize int) error { // set up the buffers. These ensure that we return data in time order mapFunc, err := influxql.InitializeMapFunc(c) if err != nil { return err } l.mapFunc = mapFunc l.keyBuffer = make([]int64, len(l.cursors)) l.valueBuffer = make([][]byte, len(l.cursors)) l.chunkSize = chunkSize l.tmin = startingTime var isCountDistinct bool // determine if this is a raw data query with a single field, multiple fields, or an aggregate var fieldName string if c == nil { // its a raw data query l.isRaw = true if len(l.selectFields) == 1 { fieldName = l.selectFields[0] } // if they haven't set a limit, just set it to the max int size if l.limit == 0 { l.limit = math.MaxUint64 } } else { // Check for calls like `derivative(mean(value), 1d)` var nested *influxql.Call = c if fn, ok := c.Args[0].(*influxql.Call); ok { nested = fn } switch lit := nested.Args[0].(type) { case *influxql.VarRef: fieldName = lit.Val case *influxql.Distinct: if c.Name != "count" { return fmt.Errorf("aggregate call didn't contain a field %s", c.String()) } isCountDistinct = true fieldName = lit.Val default: return fmt.Errorf("aggregate call didn't contain a field %s", c.String()) } isCountDistinct = isCountDistinct || (c.Name == "count" && nested.Name == "distinct") } // set up the field info if a specific field was set for this mapper if fieldName != "" { fid, err := l.decoder.FieldIDByName(fieldName) if err != nil { switch { case c != nil && c.Name == "distinct": return fmt.Errorf(`%s isn't a field on measurement %s; to query the unique values for a tag use SHOW TAG VALUES FROM %[2]s WITH KEY = "%[1]s`, fieldName, l.job.MeasurementName) case isCountDistinct: return fmt.Errorf("%s isn't a field on measurement %s; count(distinct) on tags isn't yet supported", fieldName, l.job.MeasurementName) } } l.fieldID = fid l.fieldName = fieldName } // seek the bolt cursors and fill the buffers for i, c := range l.cursors { // this series may have never been written in this shard group (time range) so the cursor would be nil if c == nil { l.keyBuffer[i] = 0 l.valueBuffer[i] = nil continue } k, v := c.Seek(u64tob(uint64(l.job.TMin))) if k == nil { l.keyBuffer[i] = 0 l.valueBuffer[i] = nil continue } l.cursorsEmpty = false t := int64(btou64(k)) l.keyBuffer[i] = t l.valueBuffer[i] = v } return nil } // NextInterval will get the time ordered next interval of the given interval size from the mapper. This is a // forward only operation from the start time passed into Begin. Will return nil when there is no more data to be read. // If this is a raw query, interval should be the max time to hit in the query func (l *LocalMapper) NextInterval() (interface{}, error) { if l.cursorsEmpty || l.tmin > l.job.TMax { return nil, nil } // after we call to the mapper, this will be the tmin for the next interval. nextMin := l.tmin + l.interval // Set the upper bound of the interval. if l.isRaw { l.perIntervalLimit = l.chunkSize } else if l.interval > 0 { // Set tmax to ensure that the interval lands on the boundary of the interval if l.tmin%l.interval != 0 { // the first interval in a query with a group by may be smaller than the others. This happens when they have a // where time > clause that is in the middle of the bucket that the group by time creates. That will be the // case on the first interval when the tmin % the interval isn't equal to zero nextMin = l.tmin/l.interval*l.interval + l.interval } l.tmax = nextMin - 1 } // Execute the map function. This local mapper acts as the iterator val := l.mapFunc(l) // see if all the cursors are empty l.cursorsEmpty = true for _, k := range l.keyBuffer { if k != 0 { l.cursorsEmpty = false break } } // Move the interval forward if it's not a raw query. For raw queries we use the limit to advance intervals. if !l.isRaw { l.tmin = nextMin } return val, nil } // Next returns the next matching timestamped value for the LocalMapper. func (l *LocalMapper) Next() (seriesKey string, timestamp int64, value interface{}) { for { // if it's a raw query and we've hit the limit of the number of points to read in // for either this chunk or for the absolute query, bail if l.isRaw && (l.limit == 0 || l.perIntervalLimit == 0) { return "", int64(0), nil } // find the minimum timestamp min := -1 minKey := int64(math.MaxInt64) for i, k := range l.keyBuffer { if k != 0 && k <= l.tmax && k < minKey && k >= l.tmin { min = i minKey = k
} } // return if there is no more data in this group by interval
random_line_split
tx.go
// SetNow sets the current time for the transaction. func (tx *tx) SetNow(now time.Time) { tx.now = now } // CreateMappers will create a set of mappers that need to be run to execute the map phase of a MapReduceJob. func (tx *tx) CreateMapReduceJobs(stmt *influxql.SelectStatement, tagKeys []string) ([]*influxql.MapReduceJob, error) { jobs := []*influxql.MapReduceJob{} for _, src := range stmt.Sources { mm, ok := src.(*influxql.Measurement) if !ok { return nil, fmt.Errorf("invalid source type: %#v", src) } // get the index and the retention policy rp, err := tx.meta.RetentionPolicy(mm.Database, mm.RetentionPolicy) if err != nil { return nil, err } m := tx.store.Measurement(mm.Database, mm.Name) if m == nil { return nil, ErrMeasurementNotFound(influxql.QuoteIdent([]string{mm.Database, "", mm.Name}...)) } tx.measurement = m // Validate the fields and tags asked for exist and keep track of which are in the select vs the where var selectFields []string var whereFields []string var selectTags []string for _, n := range stmt.NamesInSelect() { if m.HasField(n) { selectFields = append(selectFields, n) continue } if !m.HasTagKey(n) { return nil, fmt.Errorf("unknown field or tag name in select clause: %s", n) } selectTags = append(selectTags, n) tagKeys = append(tagKeys, n) } for _, n := range stmt.NamesInWhere() { if n == "time" { continue } if m.HasField(n) { whereFields = append(whereFields, n) continue } if !m.HasTagKey(n) { return nil, fmt.Errorf("unknown field or tag name in where clause: %s", n) } } if len(selectFields) == 0 && len(stmt.FunctionCalls()) == 0 { return nil, fmt.Errorf("select statement must include at least one field or function call") } // Validate that group by is not a field for _, d := range stmt.Dimensions { switch e := d.Expr.(type) { case *influxql.VarRef: if !m.HasTagKey(e.Val) { return nil, fmt.Errorf("can not use field in group by clause: %s", e.Val) } } } // Grab time range from statement. tmin, tmax := influxql.TimeRange(stmt.Condition) if tmax.IsZero() { tmax = tx.now } if tmin.IsZero() { tmin = time.Unix(0, 0) } // Find shard groups within time range. var shardGroups []*meta.ShardGroupInfo for _, group := range rp.ShardGroups { if group.Overlaps(tmin, tmax) { g := group shardGroups = append(shardGroups, &g) } } if len(shardGroups) == 0 { return nil, nil } // get the group by interval, if there is one var interval int64 if d, err := stmt.GroupByInterval(); err != nil { return nil, err } else { interval = d.Nanoseconds() } // get the sorted unique tag sets for this query. tagSets, err := m.TagSets(stmt, tagKeys) if err != nil { return nil, err } for _, t := range tagSets { // make a job for each tagset job := &influxql.MapReduceJob{ MeasurementName: m.Name, TagSet: t, TMin: tmin.UnixNano(), TMax: tmax.UnixNano(), } // make a mapper for each shard that must be hit. We may need to hit multiple shards within a shard group var mappers []influxql.Mapper // create mappers for each shard we need to hit for _, sg := range shardGroups { // TODO: implement distributed queries if len(sg.Shards) != 1 { return nil, fmt.Errorf("distributed queries aren't supported yet. You have a replication policy with RF < # of servers in cluster") } shard := tx.store.Shard(sg.Shards[0].ID) if shard == nil { // the store returned nil which means we haven't written any data into this shard yet, so ignore it continue } // get the codec for this measuremnt. If this is nil it just means this measurement was // never written into this shard, so we can skip it and continue. codec := shard.FieldCodec(m.Name) if codec == nil { continue } var mapper influxql.Mapper mapper = &LocalMapper{ seriesKeys: t.SeriesKeys, shard: shard, db: shard.DB(), job: job, decoder: codec, filters: t.Filters, whereFields: whereFields, selectFields: selectFields, selectTags: selectTags, tmin: tmin.UnixNano(), tmax: tmax.UnixNano(), interval: interval, // multiple mappers may need to be merged together to get the results // for a raw query. So each mapper will have to read at least the // limit plus the offset in data points to ensure we've hit our mark limit: uint64(stmt.Limit) + uint64(stmt.Offset), } mappers = append(mappers, mapper) } job.Mappers = mappers jobs = append(jobs, job) } } // always return them in sorted order so the results from running the jobs are returned in a deterministic order sort.Sort(influxql.MapReduceJobs(jobs)) return jobs, nil } // LocalMapper implements the influxql.Mapper interface for running map tasks over a shard that is local to this server type LocalMapper struct { cursorsEmpty bool // boolean that lets us know if the cursors are empty decoder *FieldCodec // decoder for the raw data bytes filters []influxql.Expr // filters for each series cursors []*shardCursor // bolt cursors for each series id seriesKeys []string // seriesKeys to be read from this shard shard *Shard // original shard db *bolt.DB // bolt store for the shard accessed by this mapper txn *bolt.Tx // read transactions by shard id job *influxql.MapReduceJob // the MRJob this mapper belongs to mapFunc influxql.MapFunc // the map func fieldID uint8 // the field ID associated with the mapFunc curently being run fieldName string // the field name associated with the mapFunc currently being run keyBuffer []int64 // the current timestamp key for each cursor valueBuffer [][]byte // the current value for each cursor tmin int64 // the min of the current group by interval being iterated over tmax int64 // the max of the current group by interval being iterated over additionalNames []string // additional field or tag names that might be requested from the map function whereFields []string // field names that occur in the where clause selectFields []string // field names that occur in the select clause selectTags []string // tag keys that occur in the select clause isRaw bool // if the query is a non-aggregate query interval int64 // the group by interval of the query, if any limit uint64 // used for raw queries for LIMIT perIntervalLimit int // used for raw queries to determine how far into a chunk we are chunkSize int // used for raw queries to determine how much data to read before flushing to client } // Open opens the LocalMapper. func (l *LocalMapper) Open() error { // Obtain shard lock to copy in-cache points. l.shard.mu.Lock() defer l.shard.mu.Unlock() // Open the data store txn, err := l.db.Begin(false) if err != nil { return err } l.txn = txn // create a bolt cursor for each unique series id l.cursors = make([]*shardCursor, len(l.seriesKeys)) for i, key := range l.seriesKeys { // Retrieve key bucket. b := l.txn.Bucket([]byte(key)) // Ignore if there is no bucket or points in the cache. partitionID
{ return &tx{ meta: meta, store: store, now: time.Now(), } }
identifier_body
tx.go
if !m.HasTagKey(n) { return nil, fmt.Errorf("unknown field or tag name in select clause: %s", n) } selectTags = append(selectTags, n) tagKeys = append(tagKeys, n) } for _, n := range stmt.NamesInWhere() { if n == "time" { continue } if m.HasField(n) { whereFields = append(whereFields, n) continue } if !m.HasTagKey(n) { return nil, fmt.Errorf("unknown field or tag name in where clause: %s", n) } } if len(selectFields) == 0 && len(stmt.FunctionCalls()) == 0 { return nil, fmt.Errorf("select statement must include at least one field or function call") } // Validate that group by is not a field for _, d := range stmt.Dimensions { switch e := d.Expr.(type) { case *influxql.VarRef: if !m.HasTagKey(e.Val) { return nil, fmt.Errorf("can not use field in group by clause: %s", e.Val) } } } // Grab time range from statement. tmin, tmax := influxql.TimeRange(stmt.Condition) if tmax.IsZero() { tmax = tx.now } if tmin.IsZero() { tmin = time.Unix(0, 0) } // Find shard groups within time range. var shardGroups []*meta.ShardGroupInfo for _, group := range rp.ShardGroups { if group.Overlaps(tmin, tmax) { g := group shardGroups = append(shardGroups, &g) } } if len(shardGroups) == 0 { return nil, nil } // get the group by interval, if there is one var interval int64 if d, err := stmt.GroupByInterval(); err != nil { return nil, err } else { interval = d.Nanoseconds() } // get the sorted unique tag sets for this query. tagSets, err := m.TagSets(stmt, tagKeys) if err != nil { return nil, err } for _, t := range tagSets { // make a job for each tagset job := &influxql.MapReduceJob{ MeasurementName: m.Name, TagSet: t, TMin: tmin.UnixNano(), TMax: tmax.UnixNano(), } // make a mapper for each shard that must be hit. We may need to hit multiple shards within a shard group var mappers []influxql.Mapper // create mappers for each shard we need to hit for _, sg := range shardGroups { // TODO: implement distributed queries if len(sg.Shards) != 1 { return nil, fmt.Errorf("distributed queries aren't supported yet. You have a replication policy with RF < # of servers in cluster") } shard := tx.store.Shard(sg.Shards[0].ID) if shard == nil { // the store returned nil which means we haven't written any data into this shard yet, so ignore it continue } // get the codec for this measuremnt. If this is nil it just means this measurement was // never written into this shard, so we can skip it and continue. codec := shard.FieldCodec(m.Name) if codec == nil { continue } var mapper influxql.Mapper mapper = &LocalMapper{ seriesKeys: t.SeriesKeys, shard: shard, db: shard.DB(), job: job, decoder: codec, filters: t.Filters, whereFields: whereFields, selectFields: selectFields, selectTags: selectTags, tmin: tmin.UnixNano(), tmax: tmax.UnixNano(), interval: interval, // multiple mappers may need to be merged together to get the results // for a raw query. So each mapper will have to read at least the // limit plus the offset in data points to ensure we've hit our mark limit: uint64(stmt.Limit) + uint64(stmt.Offset), } mappers = append(mappers, mapper) } job.Mappers = mappers jobs = append(jobs, job) } } // always return them in sorted order so the results from running the jobs are returned in a deterministic order sort.Sort(influxql.MapReduceJobs(jobs)) return jobs, nil } // LocalMapper implements the influxql.Mapper interface for running map tasks over a shard that is local to this server type LocalMapper struct { cursorsEmpty bool // boolean that lets us know if the cursors are empty decoder *FieldCodec // decoder for the raw data bytes filters []influxql.Expr // filters for each series cursors []*shardCursor // bolt cursors for each series id seriesKeys []string // seriesKeys to be read from this shard shard *Shard // original shard db *bolt.DB // bolt store for the shard accessed by this mapper txn *bolt.Tx // read transactions by shard id job *influxql.MapReduceJob // the MRJob this mapper belongs to mapFunc influxql.MapFunc // the map func fieldID uint8 // the field ID associated with the mapFunc curently being run fieldName string // the field name associated with the mapFunc currently being run keyBuffer []int64 // the current timestamp key for each cursor valueBuffer [][]byte // the current value for each cursor tmin int64 // the min of the current group by interval being iterated over tmax int64 // the max of the current group by interval being iterated over additionalNames []string // additional field or tag names that might be requested from the map function whereFields []string // field names that occur in the where clause selectFields []string // field names that occur in the select clause selectTags []string // tag keys that occur in the select clause isRaw bool // if the query is a non-aggregate query interval int64 // the group by interval of the query, if any limit uint64 // used for raw queries for LIMIT perIntervalLimit int // used for raw queries to determine how far into a chunk we are chunkSize int // used for raw queries to determine how much data to read before flushing to client } // Open opens the LocalMapper. func (l *LocalMapper) Open() error { // Obtain shard lock to copy in-cache points. l.shard.mu.Lock() defer l.shard.mu.Unlock() // Open the data store txn, err := l.db.Begin(false) if err != nil { return err } l.txn = txn // create a bolt cursor for each unique series id l.cursors = make([]*shardCursor, len(l.seriesKeys)) for i, key := range l.seriesKeys { // Retrieve key bucket. b := l.txn.Bucket([]byte(key)) // Ignore if there is no bucket or points in the cache. partitionID := WALPartition([]byte(key)) if b == nil && len(l.shard.cache[partitionID][key]) == 0 { continue } // Retrieve a copy of the in-cache points for the key. cache := make([][]byte, len(l.shard.cache[partitionID][key])) copy(cache, l.shard.cache[partitionID][key]) // Build a cursor that merges the bucket and cache together. cur := &shardCursor{cache: cache} if b != nil { cur.cursor = b.Cursor() } l.cursors[i] = cur } return nil } // Close closes the LocalMapper. func (l *LocalMapper) Close() { if l.txn != nil { _ = l.txn.Rollback() } } // Begin will set up the mapper to run the map function for a given aggregate call starting at the passed in time func (l *LocalMapper) Begin(c *influxql.Call, startingTime int64, chunkSize int) error { // set up the buffers. These ensure that we return data in time order mapFunc, err := influxql.InitializeMapFunc(c) if err != nil { return err } l.mapFunc = mapFunc l.keyBuffer = make([]int64, len(l.cursors)) l.valueBuffer = make([][]byte, len(l.cursors)) l.chunkSize = chunkSize l.tmin = startingTime var isCountDistinct bool // determine if this is a raw data query with a single
{ selectFields = append(selectFields, n) continue }
conditional_block
testing.py
bslash", "|": "pipe", "?": "qmark", "*": "asterisk", } def _response_to_json(resp: Response, ip_dict: Dict[str, int]) -> str: """Converts a Response object into json string readable by the responses mocking library Args: resp: Response from a Tamr API call ip_dict: Mapping of previously encountered IP addresses to their anonymization number Returns: The response represented as a json string """ if resp.encoding is None: resp.encoding = "utf-8" resp_log = { "method": resp.request.method, "url": _anonymize_url(resp.request.url, ip_dict), "status": resp.status_code, "content_type": resp.headers.get("Content-Type"), "body": resp.text, } return json.dumps(resp_log, ensure_ascii=False) def _anonymize_url(url: str, ip_dict: Dict[str, int]) -> str: """Returns a anonymized url. Updates the dictionary inplace if a new ip is encountered Args: url: A URL ip_dict: Previously encountered IP addresses and an assigned numeric value Returns: URL with the IP address anonymized """ regex_match = re.match(r"(?i)(^https?://)(.*?)([/:].*$)", url) ip = regex_match.group(2) try: num = ip_dict[ip] except KeyError: ip_dict[ip] = len(ip_dict.values()) + 1 num = ip_dict[ip] return f"{regex_match.group(1)}ip-{num:05d}{regex_match.group(3)}" def _collect_operation_calls( *, response: Response, poll_interval_seconds: int = 3 ) -> List[Response]: """If the provided response is an Operation, wait for the operation to complete and return responses related to that operation. Args: response: A previous Response generated from the same Tamr client poll_interval_seconds: Time interval (in seconds) between subsequent polls Returns: Responses related to polling the operation """ client = utils.client._from_response(response) op = Operation.from_response(client, response) LOGGER.info(f"Waiting for operation to complete: {op}") request_while_pending = client.get(endpoint=f"/api/versioned/v1/operations/{op.resource_id}") while op.state == "PENDING": op = op.poll() sleep(poll_interval_seconds) request_while_running = client.get(endpoint=f"/api/versioned/v1/operations/{op.resource_id}") op.wait() request_when_complete = client.get(endpoint=f"/api/versioned/v1/operations/{op.resource_id}") return [request_while_pending, request_while_running, request_when_complete] def _log_response(*, log_path: Path, ip_dict: Dict[str, int], response: Response) -> None: """Appends a response to a file. If the response returned is a Tamr Operation, poll the operation until complete and log those responses as well Args: log_path: File to write the response to ip_dict: Mapping of previously encountered IP addresses to their anonymization number response: The response to log """ LOGGER.info(f"logged request: {response.url}") with log_path.open(mode="a", encoding="utf-8") as f: all_responses = [response] # Poll and wait for operations, if applicable is_operation_request = bool( re.match(re.compile(".*/api/versioned/v1/operations/.*"), response.url) ) is_get_request = response.request.method == "GET" if is_get_request and is_operation_request: wait_resp = _collect_operation_calls(response=response) all_responses.extend(wait_resp) all_json = [_response_to_json(r, ip_dict) for r in all_responses] f.writelines([f"{j}\n" for j in all_json]) def _build_response_log_path( *, test_func: Callable, response_logs_dir: Optional[Union[str, Path]], **kwargs, ) -> Path: """Returns a file path for API response logs for a given test and test parameters Args: test_func: The test function **kwargs: Arguments to the test function Returns: File path for the API response logs """ # Convert test arguments and their values to a string, skipping ignored arguments test_params = "_".join([f"{k}={v}" for k, v in {**kwargs}.items()]) # Remove reserved characters from ndjson name for char in WINDOWS_RESERVED_CHARACTER_MAP: test_params = test_params.replace(char, WINDOWS_RESERVED_CHARACTER_MAP[char]) if len(test_params) > 0: test_params = "__" + test_params if response_logs_dir is None: # If no directory is provided, create a directory with the name of the test file # in a directory called "response_logs" located in the same directory as the test file dir_matcher = re.match(r"(.*)(?:\\|/)(.*).py", str(Path(getfile(test_func)))) response_logs_dir = f"{dir_matcher.group(1)}/response_logs/{dir_matcher.group(2)}" return Path(f"{response_logs_dir}/{test_func.__name__}{test_params}.ndjson") def mock_api( *, response_logs_dir: Optional[Union[str, Path]] = None, enforce_online_test=False ) -> Callable: """Decorator for `pytest` tests that mocks API requests by reading a file of pre-generated responses. Will generate responses file based on a real connection if pre-generated responses are not found. Args: response_logs_dir: Directory to read/write response logs enforce_online_test: Whether an online test should be run, even if a response log already exists Returns: Decorated function """ def wrap(test_function: Callable): @wraps(test_function) def wrapped(**kwargs): response_log_path = _build_response_log_path( test_func=test_function, response_logs_dir=response_logs_dir, **kwargs, ) if response_log_path.exists() and enforce_online_test: # Delete the file to enforce an online test response_log_path.unlink() if response_log_path.exists(): try: LOGGER.info(f"Running offline test based on file at {response_log_path}") _run_offline_test( response_log_path=response_log_path, test_function=test_function, **kwargs, ) except ConnectionError as e: msg = ( f"A required API call was missing from response logs file for this " f"offline test ({response_log_path}). The response log file must be " f"regenerated. Delete the existing file to automatically regenerate a " f"new one. Caused by: {e}" ) LOGGER.error(msg) raise ConnectionError(e) else: _run_online_test( response_log_path=response_log_path, test_function=test_function, **kwargs ) return wrapped return wrap # Handle ModuleNotFoundError to allow tamr_toolbox to be used when the optional dependency # `responses` is not installed try: import responses # Stores the original _real_send function of requests _BASE_FIND_MATCH = responses.RequestsMock._find_match # Stores the original _real_send function of responses _BASE_SEND_REAL = responses._real_send @responses.activate @mock.patch.object(Operation.wait, "__defaults__", (0, None)) # sets operation wait time to 0 def _run_offline_test(response_log_path: Path, test_function: Callable, **kwargs) -> None: """Runs a test function against saved API responses located in a file Args:
response_log_path: Location of saved API responses test_function: The function to test **kwargs: Keyword arguments for the test function """ with response_log_path.open(encoding="utf-8") as f: for line in f: response = json.loads(line) responses.add(**response) ip_lookup = {} def _find_anonymized_match(self, request): """Allows responses library to match requests for an ip address to match to an anonymized ip address """ request.url = _anonymize_url(request.url, ip_lookup) return _BASE_FIND_MATCH(self, request) with mock.patch("responses.RequestsMock._find_match", new=_find_anonymized_match): test_function(**kwargs) @responses.activate def _run_online_test(response_log_path: Path, test_function: Callable, **kwargs) -> None: """Runs a test function against a Tamr instance and saves the API responses to a file Args: response_log_path: Location to save API responses test_function: The function to test **kwargs: Keyword arguments for the test function """ LOGGER.info( f"Online test running against Tamr instance. " f"Creating new file at {response_log_path}. This may take a while ..." ) os.makedirs(response_log_path.parent, exist_ok=True) response_log_path.touch() # Each time an API call is made, allow it to pass through responses and make a real call # Each time a real call is made, log the response in the response file responses.add_passthru(re.compile(".*")) ip_lookup = {} def _send_real_with_log(*args, **kwargs
random_line_split
testing.py
lash", "|": "pipe", "?": "qmark", "*": "asterisk", } def _response_to_json(resp: Response, ip_dict: Dict[str, int]) -> str: """Converts a Response object into json string readable by the responses mocking library Args: resp: Response from a Tamr API call ip_dict: Mapping of previously encountered IP addresses to their anonymization number Returns: The response represented as a json string """ if resp.encoding is None: resp.encoding = "utf-8" resp_log = { "method": resp.request.method, "url": _anonymize_url(resp.request.url, ip_dict), "status": resp.status_code, "content_type": resp.headers.get("Content-Type"), "body": resp.text, } return json.dumps(resp_log, ensure_ascii=False) def _anonymize_url(url: str, ip_dict: Dict[str, int]) -> str: """Returns a anonymized url. Updates the dictionary inplace if a new ip is encountered Args: url: A URL ip_dict: Previously encountered IP addresses and an assigned numeric value Returns: URL with the IP address anonymized """ regex_match = re.match(r"(?i)(^https?://)(.*?)([/:].*$)", url) ip = regex_match.group(2) try: num = ip_dict[ip] except KeyError: ip_dict[ip] = len(ip_dict.values()) + 1 num = ip_dict[ip] return f"{regex_match.group(1)}ip-{num:05d}{regex_match.group(3)}" def _collect_operation_calls( *, response: Response, poll_interval_seconds: int = 3 ) -> List[Response]: """If the provided response is an Operation, wait for the operation to complete and return responses related to that operation. Args: response: A previous Response generated from the same Tamr client poll_interval_seconds: Time interval (in seconds) between subsequent polls Returns: Responses related to polling the operation """ client = utils.client._from_response(response) op = Operation.from_response(client, response) LOGGER.info(f"Waiting for operation to complete: {op}") request_while_pending = client.get(endpoint=f"/api/versioned/v1/operations/{op.resource_id}") while op.state == "PENDING": op = op.poll() sleep(poll_interval_seconds) request_while_running = client.get(endpoint=f"/api/versioned/v1/operations/{op.resource_id}") op.wait() request_when_complete = client.get(endpoint=f"/api/versioned/v1/operations/{op.resource_id}") return [request_while_pending, request_while_running, request_when_complete] def _log_response(*, log_path: Path, ip_dict: Dict[str, int], response: Response) -> None: """Appends a response to a file. If the response returned is a Tamr Operation, poll the operation until complete and log those responses as well Args: log_path: File to write the response to ip_dict: Mapping of previously encountered IP addresses to their anonymization number response: The response to log """ LOGGER.info(f"logged request: {response.url}") with log_path.open(mode="a", encoding="utf-8") as f: all_responses = [response] # Poll and wait for operations, if applicable is_operation_request = bool( re.match(re.compile(".*/api/versioned/v1/operations/.*"), response.url) ) is_get_request = response.request.method == "GET" if is_get_request and is_operation_request: wait_resp = _collect_operation_calls(response=response) all_responses.extend(wait_resp) all_json = [_response_to_json(r, ip_dict) for r in all_responses] f.writelines([f"{j}\n" for j in all_json]) def _build_response_log_path( *, test_func: Callable, response_logs_dir: Optional[Union[str, Path]], **kwargs, ) -> Path: """Returns a file path for API response logs for a given test and test parameters Args: test_func: The test function **kwargs: Arguments to the test function Returns: File path for the API response logs """ # Convert test arguments and their values to a string, skipping ignored arguments test_params = "_".join([f"{k}={v}" for k, v in {**kwargs}.items()]) # Remove reserved characters from ndjson name for char in WINDOWS_RESERVED_CHARACTER_MAP: test_params = test_params.replace(char, WINDOWS_RESERVED_CHARACTER_MAP[char]) if len(test_params) > 0: test_params = "__" + test_params if response_logs_dir is None: # If no directory is provided, create a directory with the name of the test file # in a directory called "response_logs" located in the same directory as the test file dir_matcher = re.match(r"(.*)(?:\\|/)(.*).py", str(Path(getfile(test_func)))) response_logs_dir = f"{dir_matcher.group(1)}/response_logs/{dir_matcher.group(2)}" return Path(f"{response_logs_dir}/{test_func.__name__}{test_params}.ndjson") def mock_api( *, response_logs_dir: Optional[Union[str, Path]] = None, enforce_online_test=False ) -> Callable:
) if response_log_path.exists() and enforce_online_test: # Delete the file to enforce an online test response_log_path.unlink() if response_log_path.exists(): try: LOGGER.info(f"Running offline test based on file at {response_log_path}") _run_offline_test( response_log_path=response_log_path, test_function=test_function, **kwargs, ) except ConnectionError as e: msg = ( f"A required API call was missing from response logs file for this " f"offline test ({response_log_path}). The response log file must be " f"regenerated. Delete the existing file to automatically regenerate a " f"new one. Caused by: {e}" ) LOGGER.error(msg) raise ConnectionError(e) else: _run_online_test( response_log_path=response_log_path, test_function=test_function, **kwargs ) return wrapped return wrap # Handle ModuleNotFoundError to allow tamr_toolbox to be used when the optional dependency # `responses` is not installed try: import responses # Stores the original _real_send function of requests _BASE_FIND_MATCH = responses.RequestsMock._find_match # Stores the original _real_send function of responses _BASE_SEND_REAL = responses._real_send @responses.activate @mock.patch.object(Operation.wait, "__defaults__", (0, None)) # sets operation wait time to 0 def _run_offline_test(response_log_path: Path, test_function: Callable, **kwargs) -> None: """Runs a test function against saved API responses located in a file Args: response_log_path: Location of saved API responses test_function: The function to test **kwargs: Keyword arguments for the test function """ with response_log_path.open(encoding="utf-8") as f: for line in f: response = json.loads(line) responses.add(**response) ip_lookup = {} def _find_anonymized_match(self, request): """Allows responses library to match requests for an ip address to match to an anonymized ip address """ request.url = _anonymize_url(request.url, ip_lookup) return _BASE_FIND_MATCH(self, request) with mock.patch("responses.RequestsMock._find_match", new=_find_anonymized_match): test_function(**kwargs) @responses.activate def _run_online_test(response_log_path: Path, test_function: Callable, **kwargs) -> None: """Runs a test function against a Tamr instance and saves the API responses to a file Args: response_log_path: Location to save API responses test_function: The function to test **kwargs: Keyword arguments for the test function """ LOGGER.info( f"Online test running against Tamr instance. " f"Creating new file at {response_log_path}. This may take a while ..." ) os.makedirs(response_log_path.parent, exist_ok=True) response_log_path.touch() # Each time an API call is made, allow it to pass through responses and make a real call # Each time a real call is made, log the response in the response file responses.add_passthru(re.compile(".*")) ip_lookup = {} def _send_real_with_log(*args, **kwargs)
"""Decorator for `pytest` tests that mocks API requests by reading a file of pre-generated responses. Will generate responses file based on a real connection if pre-generated responses are not found. Args: response_logs_dir: Directory to read/write response logs enforce_online_test: Whether an online test should be run, even if a response log already exists Returns: Decorated function """ def wrap(test_function: Callable): @wraps(test_function) def wrapped(**kwargs): response_log_path = _build_response_log_path( test_func=test_function, response_logs_dir=response_logs_dir, **kwargs,
identifier_body
testing.py
lash", "|": "pipe", "?": "qmark", "*": "asterisk", } def _response_to_json(resp: Response, ip_dict: Dict[str, int]) -> str: """Converts a Response object into json string readable by the responses mocking library Args: resp: Response from a Tamr API call ip_dict: Mapping of previously encountered IP addresses to their anonymization number Returns: The response represented as a json string """ if resp.encoding is None: resp.encoding = "utf-8" resp_log = { "method": resp.request.method, "url": _anonymize_url(resp.request.url, ip_dict), "status": resp.status_code, "content_type": resp.headers.get("Content-Type"), "body": resp.text, } return json.dumps(resp_log, ensure_ascii=False) def _anonymize_url(url: str, ip_dict: Dict[str, int]) -> str: """Returns a anonymized url. Updates the dictionary inplace if a new ip is encountered Args: url: A URL ip_dict: Previously encountered IP addresses and an assigned numeric value Returns: URL with the IP address anonymized """ regex_match = re.match(r"(?i)(^https?://)(.*?)([/:].*$)", url) ip = regex_match.group(2) try: num = ip_dict[ip] except KeyError: ip_dict[ip] = len(ip_dict.values()) + 1 num = ip_dict[ip] return f"{regex_match.group(1)}ip-{num:05d}{regex_match.group(3)}" def _collect_operation_calls( *, response: Response, poll_interval_seconds: int = 3 ) -> List[Response]: """If the provided response is an Operation, wait for the operation to complete and return responses related to that operation. Args: response: A previous Response generated from the same Tamr client poll_interval_seconds: Time interval (in seconds) between subsequent polls Returns: Responses related to polling the operation """ client = utils.client._from_response(response) op = Operation.from_response(client, response) LOGGER.info(f"Waiting for operation to complete: {op}") request_while_pending = client.get(endpoint=f"/api/versioned/v1/operations/{op.resource_id}") while op.state == "PENDING": op = op.poll() sleep(poll_interval_seconds) request_while_running = client.get(endpoint=f"/api/versioned/v1/operations/{op.resource_id}") op.wait() request_when_complete = client.get(endpoint=f"/api/versioned/v1/operations/{op.resource_id}") return [request_while_pending, request_while_running, request_when_complete] def _log_response(*, log_path: Path, ip_dict: Dict[str, int], response: Response) -> None: """Appends a response to a file. If the response returned is a Tamr Operation, poll the operation until complete and log those responses as well Args: log_path: File to write the response to ip_dict: Mapping of previously encountered IP addresses to their anonymization number response: The response to log """ LOGGER.info(f"logged request: {response.url}") with log_path.open(mode="a", encoding="utf-8") as f: all_responses = [response] # Poll and wait for operations, if applicable is_operation_request = bool( re.match(re.compile(".*/api/versioned/v1/operations/.*"), response.url) ) is_get_request = response.request.method == "GET" if is_get_request and is_operation_request: wait_resp = _collect_operation_calls(response=response) all_responses.extend(wait_resp) all_json = [_response_to_json(r, ip_dict) for r in all_responses] f.writelines([f"{j}\n" for j in all_json]) def _build_response_log_path( *, test_func: Callable, response_logs_dir: Optional[Union[str, Path]], **kwargs, ) -> Path: """Returns a file path for API response logs for a given test and test parameters Args: test_func: The test function **kwargs: Arguments to the test function Returns: File path for the API response logs """ # Convert test arguments and their values to a string, skipping ignored arguments test_params = "_".join([f"{k}={v}" for k, v in {**kwargs}.items()]) # Remove reserved characters from ndjson name for char in WINDOWS_RESERVED_CHARACTER_MAP: test_params = test_params.replace(char, WINDOWS_RESERVED_CHARACTER_MAP[char]) if len(test_params) > 0: test_params = "__" + test_params if response_logs_dir is None: # If no directory is provided, create a directory with the name of the test file # in a directory called "response_logs" located in the same directory as the test file dir_matcher = re.match(r"(.*)(?:\\|/)(.*).py", str(Path(getfile(test_func)))) response_logs_dir = f"{dir_matcher.group(1)}/response_logs/{dir_matcher.group(2)}" return Path(f"{response_logs_dir}/{test_func.__name__}{test_params}.ndjson") def mock_api( *, response_logs_dir: Optional[Union[str, Path]] = None, enforce_online_test=False ) -> Callable: """Decorator for `pytest` tests that mocks API requests by reading a file of pre-generated responses. Will generate responses file based on a real connection if pre-generated responses are not found. Args: response_logs_dir: Directory to read/write response logs enforce_online_test: Whether an online test should be run, even if a response log already exists Returns: Decorated function """ def wrap(test_function: Callable): @wraps(test_function) def
(**kwargs): response_log_path = _build_response_log_path( test_func=test_function, response_logs_dir=response_logs_dir, **kwargs, ) if response_log_path.exists() and enforce_online_test: # Delete the file to enforce an online test response_log_path.unlink() if response_log_path.exists(): try: LOGGER.info(f"Running offline test based on file at {response_log_path}") _run_offline_test( response_log_path=response_log_path, test_function=test_function, **kwargs, ) except ConnectionError as e: msg = ( f"A required API call was missing from response logs file for this " f"offline test ({response_log_path}). The response log file must be " f"regenerated. Delete the existing file to automatically regenerate a " f"new one. Caused by: {e}" ) LOGGER.error(msg) raise ConnectionError(e) else: _run_online_test( response_log_path=response_log_path, test_function=test_function, **kwargs ) return wrapped return wrap # Handle ModuleNotFoundError to allow tamr_toolbox to be used when the optional dependency # `responses` is not installed try: import responses # Stores the original _real_send function of requests _BASE_FIND_MATCH = responses.RequestsMock._find_match # Stores the original _real_send function of responses _BASE_SEND_REAL = responses._real_send @responses.activate @mock.patch.object(Operation.wait, "__defaults__", (0, None)) # sets operation wait time to 0 def _run_offline_test(response_log_path: Path, test_function: Callable, **kwargs) -> None: """Runs a test function against saved API responses located in a file Args: response_log_path: Location of saved API responses test_function: The function to test **kwargs: Keyword arguments for the test function """ with response_log_path.open(encoding="utf-8") as f: for line in f: response = json.loads(line) responses.add(**response) ip_lookup = {} def _find_anonymized_match(self, request): """Allows responses library to match requests for an ip address to match to an anonymized ip address """ request.url = _anonymize_url(request.url, ip_lookup) return _BASE_FIND_MATCH(self, request) with mock.patch("responses.RequestsMock._find_match", new=_find_anonymized_match): test_function(**kwargs) @responses.activate def _run_online_test(response_log_path: Path, test_function: Callable, **kwargs) -> None: """Runs a test function against a Tamr instance and saves the API responses to a file Args: response_log_path: Location to save API responses test_function: The function to test **kwargs: Keyword arguments for the test function """ LOGGER.info( f"Online test running against Tamr instance. " f"Creating new file at {response_log_path}. This may take a while ..." ) os.makedirs(response_log_path.parent, exist_ok=True) response_log_path.touch() # Each time an API call is made, allow it to pass through responses and make a real call # Each time a real call is made, log the response in the response file responses.add_passthru(re.compile(".*")) ip_lookup = {} def _send_real_with_log(*args, **kwargs
wrapped
identifier_name
testing.py
lash", "|": "pipe", "?": "qmark", "*": "asterisk", } def _response_to_json(resp: Response, ip_dict: Dict[str, int]) -> str: """Converts a Response object into json string readable by the responses mocking library Args: resp: Response from a Tamr API call ip_dict: Mapping of previously encountered IP addresses to their anonymization number Returns: The response represented as a json string """ if resp.encoding is None: resp.encoding = "utf-8" resp_log = { "method": resp.request.method, "url": _anonymize_url(resp.request.url, ip_dict), "status": resp.status_code, "content_type": resp.headers.get("Content-Type"), "body": resp.text, } return json.dumps(resp_log, ensure_ascii=False) def _anonymize_url(url: str, ip_dict: Dict[str, int]) -> str: """Returns a anonymized url. Updates the dictionary inplace if a new ip is encountered Args: url: A URL ip_dict: Previously encountered IP addresses and an assigned numeric value Returns: URL with the IP address anonymized """ regex_match = re.match(r"(?i)(^https?://)(.*?)([/:].*$)", url) ip = regex_match.group(2) try: num = ip_dict[ip] except KeyError: ip_dict[ip] = len(ip_dict.values()) + 1 num = ip_dict[ip] return f"{regex_match.group(1)}ip-{num:05d}{regex_match.group(3)}" def _collect_operation_calls( *, response: Response, poll_interval_seconds: int = 3 ) -> List[Response]: """If the provided response is an Operation, wait for the operation to complete and return responses related to that operation. Args: response: A previous Response generated from the same Tamr client poll_interval_seconds: Time interval (in seconds) between subsequent polls Returns: Responses related to polling the operation """ client = utils.client._from_response(response) op = Operation.from_response(client, response) LOGGER.info(f"Waiting for operation to complete: {op}") request_while_pending = client.get(endpoint=f"/api/versioned/v1/operations/{op.resource_id}") while op.state == "PENDING": op = op.poll() sleep(poll_interval_seconds) request_while_running = client.get(endpoint=f"/api/versioned/v1/operations/{op.resource_id}") op.wait() request_when_complete = client.get(endpoint=f"/api/versioned/v1/operations/{op.resource_id}") return [request_while_pending, request_while_running, request_when_complete] def _log_response(*, log_path: Path, ip_dict: Dict[str, int], response: Response) -> None: """Appends a response to a file. If the response returned is a Tamr Operation, poll the operation until complete and log those responses as well Args: log_path: File to write the response to ip_dict: Mapping of previously encountered IP addresses to their anonymization number response: The response to log """ LOGGER.info(f"logged request: {response.url}") with log_path.open(mode="a", encoding="utf-8") as f: all_responses = [response] # Poll and wait for operations, if applicable is_operation_request = bool( re.match(re.compile(".*/api/versioned/v1/operations/.*"), response.url) ) is_get_request = response.request.method == "GET" if is_get_request and is_operation_request: wait_resp = _collect_operation_calls(response=response) all_responses.extend(wait_resp) all_json = [_response_to_json(r, ip_dict) for r in all_responses] f.writelines([f"{j}\n" for j in all_json]) def _build_response_log_path( *, test_func: Callable, response_logs_dir: Optional[Union[str, Path]], **kwargs, ) -> Path: """Returns a file path for API response logs for a given test and test parameters Args: test_func: The test function **kwargs: Arguments to the test function Returns: File path for the API response logs """ # Convert test arguments and their values to a string, skipping ignored arguments test_params = "_".join([f"{k}={v}" for k, v in {**kwargs}.items()]) # Remove reserved characters from ndjson name for char in WINDOWS_RESERVED_CHARACTER_MAP: test_params = test_params.replace(char, WINDOWS_RESERVED_CHARACTER_MAP[char]) if len(test_params) > 0: test_params = "__" + test_params if response_logs_dir is None: # If no directory is provided, create a directory with the name of the test file # in a directory called "response_logs" located in the same directory as the test file dir_matcher = re.match(r"(.*)(?:\\|/)(.*).py", str(Path(getfile(test_func)))) response_logs_dir = f"{dir_matcher.group(1)}/response_logs/{dir_matcher.group(2)}" return Path(f"{response_logs_dir}/{test_func.__name__}{test_params}.ndjson") def mock_api( *, response_logs_dir: Optional[Union[str, Path]] = None, enforce_online_test=False ) -> Callable: """Decorator for `pytest` tests that mocks API requests by reading a file of pre-generated responses. Will generate responses file based on a real connection if pre-generated responses are not found. Args: response_logs_dir: Directory to read/write response logs enforce_online_test: Whether an online test should be run, even if a response log already exists Returns: Decorated function """ def wrap(test_function: Callable): @wraps(test_function) def wrapped(**kwargs): response_log_path = _build_response_log_path( test_func=test_function, response_logs_dir=response_logs_dir, **kwargs, ) if response_log_path.exists() and enforce_online_test: # Delete the file to enforce an online test response_log_path.unlink() if response_log_path.exists(): try: LOGGER.info(f"Running offline test based on file at {response_log_path}") _run_offline_test( response_log_path=response_log_path, test_function=test_function, **kwargs, ) except ConnectionError as e: msg = ( f"A required API call was missing from response logs file for this " f"offline test ({response_log_path}). The response log file must be " f"regenerated. Delete the existing file to automatically regenerate a " f"new one. Caused by: {e}" ) LOGGER.error(msg) raise ConnectionError(e) else: _run_online_test( response_log_path=response_log_path, test_function=test_function, **kwargs ) return wrapped return wrap # Handle ModuleNotFoundError to allow tamr_toolbox to be used when the optional dependency # `responses` is not installed try: import responses # Stores the original _real_send function of requests _BASE_FIND_MATCH = responses.RequestsMock._find_match # Stores the original _real_send function of responses _BASE_SEND_REAL = responses._real_send @responses.activate @mock.patch.object(Operation.wait, "__defaults__", (0, None)) # sets operation wait time to 0 def _run_offline_test(response_log_path: Path, test_function: Callable, **kwargs) -> None: """Runs a test function against saved API responses located in a file Args: response_log_path: Location of saved API responses test_function: The function to test **kwargs: Keyword arguments for the test function """ with response_log_path.open(encoding="utf-8") as f: for line in f:
ip_lookup = {} def _find_anonymized_match(self, request): """Allows responses library to match requests for an ip address to match to an anonymized ip address """ request.url = _anonymize_url(request.url, ip_lookup) return _BASE_FIND_MATCH(self, request) with mock.patch("responses.RequestsMock._find_match", new=_find_anonymized_match): test_function(**kwargs) @responses.activate def _run_online_test(response_log_path: Path, test_function: Callable, **kwargs) -> None: """Runs a test function against a Tamr instance and saves the API responses to a file Args: response_log_path: Location to save API responses test_function: The function to test **kwargs: Keyword arguments for the test function """ LOGGER.info( f"Online test running against Tamr instance. " f"Creating new file at {response_log_path}. This may take a while ..." ) os.makedirs(response_log_path.parent, exist_ok=True) response_log_path.touch() # Each time an API call is made, allow it to pass through responses and make a real call # Each time a real call is made, log the response in the response file responses.add_passthru(re.compile(".*")) ip_lookup = {} def _send_real_with_log(*args, **kwargs
response = json.loads(line) responses.add(**response)
conditional_block
current_models.py
6)) minf = alpha_m/(alpha_m + beta_m) mtau = 1/(alpha_m + beta_m) alpha_h = 0.011 + 1.39/(1 + np.exp((v+78.04)/11.32)) beta_h = 0.56 - 0.56/(1 + np.exp((v-21.82)/20.03)) hinf = alpha_h/(alpha_h + beta_h) htau = 1/(alpha_h + beta_h) dm = (minf-m)/mtau dh = (hinf-h)/htau return [dm, dh] def nav18tf(Y,t,voltage_clamp_func,voltage_clamp_params): " Rat? Nav 1.8 used in Tigerholm model " v = voltage_clamp_func(t,voltage_clamp_params) m = Y[0] h = Y[1] s = Y[2] u = Y[3] alpha_m = 2.85 - 2.839/(1 + np.exp((v-1.159)/13.95)) beta_m = 7.6205/(1 + np.exp((v+46.463)/8.8289)) minf = alpha_m/(alpha_m + beta_m) mtau = 1/(alpha_m + beta_m) hinf = 1/(1+np.exp((v+32.2)/4)) htau = 1.218 + 42.043*np.exp(-((v+38.1)**2)/(2*15.19**2)) alpha_s = 0.001 * 5.4203 / (1 + np.exp((v+79.816)/16.269)) beta_s = 0.001 * 5.0757 / (1 + np.exp(-(v+15.968)/11.542)) sinf = 1/(1+np.exp((v+45.0)/8)) stau = 1/(alpha_s + beta_s) alpha_u = 0.002 * 2.0434 / (1 + np.exp((v+67.499)/19.51)) beta_u = 0.002 * 1.9952 / (1 + np.exp(-(v+30.963)/14.792)) uinf = 1/(1+np.exp((v+51.0)/8)) utau = 1.0/(alpha_u + beta_u) dm = (minf-m)/mtau dh = (hinf-h)/htau ds = (sinf-s)/stau du = (uinf-u)/utau return [dm, dh, ds, du] def nav18cw(Y,t,voltage_clamp_func,voltage_clamp_params): " Nav 1.8 model used in Choi Waxman 2011 " v = voltage_clamp_func(t,voltage_clamp_params) m = Y[0] h = Y[1] alpha_m = 2.85 - 2.839/(1 + np.exp((v-1.159)/13.95)) beta_m = 7.6205/(1 + np.exp((v+46.463)/8.8289)) minf = alpha_m/(alpha_m + beta_m) mtau = 1/(alpha_m + beta_m) hinf = 1/(1+np.exp((v+32.2)/4)) htau = 1.218 + 42.043*np.exp(-((v+38.1)**2)/(2*15.19**2)) dm = (minf-m)/mtau dh = (hinf-h)/htau return [dm, dh] " -- Nav 1.9 models -- " def nav19hw(Y,t,voltage_clamp_func,voltage_clamp_params): " Nav 1.9 model from Huang Waxman 2014" m = Y[0] h = Y[1] s = Y[2] v = voltage_clamp_func(t,voltage_clamp_params) alpha_m = 0.751/(1 + np.exp(-(v+32.26)/13.71)) beta_m = 5.68/(1 + np.exp((v+123.71)/13.94)) minf = alpha_m/(alpha_m + beta_m) mtau = 1/(alpha_m + beta_m) alpha_h = 0.082/(1 + np.exp((v+113.69)/17.4)) beta_h = 0.24/(1 + np.exp(-(v-10.1)/17.2)) hinf = alpha_h/(alpha_h + beta_h) htau = 1/(alpha_h + beta_h) alpha_s = 0.019/(1 + np.exp((v+154.51)/11.46)) beta_s = 0.000376/(1 + np.exp(-(v+60.92)/15.79)) sinf = alpha_s/(alpha_s + beta_s) stau = 1/(alpha_s + beta_s) dm = (minf-m)/mtau dh = (hinf-h)/htau ds = (sinf-s)/stau return [dm, dh, ds] def nav19md(Y,t,voltage_clamp_func,voltage_clamp_params): " Nav 1.9 model from Maingret 2008" m = Y[0] h = Y[1] s = Y[2] v = voltage_clamp_func(t,voltage_clamp_params) return [dm, dh, ds] def nav16zm(Y,t,voltage_clamp_func,voltage_clamp_params): " Nav 1.6 model from Zach Mainen 1994 " m = Y[0]
h = Y[1] v = voltage_clamp_func(t,voltage_clamp_params) vhalf = -43.0 a_m = 0.182*(v-vhalf)/(1-np.exp((vhalf-v)/6.)) b_m = 0.124*(-v+vhalf)/(1-np.exp((-vhalf+v)/6.)) m_inf = a_m/(a_m + b_m) m_tau = 1./(a_m + b_m) vhalf_ha = -50.0 vhalf_hb = -75.0 q_h = 5.0 vhalf_inf = -72.0 qinf = 6.2 rate_ha = 0.0091 rate_hb = 0.024 a_h = rate_ha*(v-vhalf_ha)/(1-np.exp((vhalf_ha-v)/q_h)) b_h = rate_hb*(-v+vhalf_hb)/(1-np.exp((-vhalf_hb+v)/q_h)) h_inf = 1.0/(1.0 + np.exp((v-vhalf_inf)/qinf)) h_tau = 1./(a_h + b_h) dm = (m_inf-m)/m_tau dh = (h_inf-h)/h_tau return [dm, dh] " Kv models " def kdr_tf(Y,t,voltage_clamp_func,voltage_clamp_params): " Tigerholm version of the Sheets et al. IKdr model " " Model was developed from data recorded at 21 oC " v = voltage_clamp_func(t,voltage_clamp_params) n = Y[0] q10 = 1.0#3.3 # Preserved in case it is useful but disabled if v > -31.0: tau = 0.16+0.8*np.exp(-0.0267*(v+11)) else: tau = 1000*(0.000688 + 1/(np.exp((v+75.2)/6.5) + np.exp(-(v-131.5)/(34.8)))) ninf = 1/(1 + np.exp(-(v+45)/15.4)) ntau = tau/q10 dn = (ninf-n)/ntau return [dn] def km_tf(Y,t,voltage_clamp_func,voltage_clamp_params): """ Tigerholm version of the IM current. Current is from multiple sources: The voltage dependence of steady-state
random_line_split
current_models.py
376/(1 + np.exp(-(v+60.92)/15.79)) sinf = alpha_s/(alpha_s + beta_s) stau = 1/(alpha_s + beta_s) dm = (minf-m)/mtau dh = (hinf-h)/htau ds = (sinf-s)/stau return [dm, dh, ds] def nav19md(Y,t,voltage_clamp_func,voltage_clamp_params): " Nav 1.9 model from Maingret 2008" m = Y[0] h = Y[1] s = Y[2] v = voltage_clamp_func(t,voltage_clamp_params) return [dm, dh, ds] def nav16zm(Y,t,voltage_clamp_func,voltage_clamp_params): " Nav 1.6 model from Zach Mainen 1994 " m = Y[0] h = Y[1] v = voltage_clamp_func(t,voltage_clamp_params) vhalf = -43.0 a_m = 0.182*(v-vhalf)/(1-np.exp((vhalf-v)/6.)) b_m = 0.124*(-v+vhalf)/(1-np.exp((-vhalf+v)/6.)) m_inf = a_m/(a_m + b_m) m_tau = 1./(a_m + b_m) vhalf_ha = -50.0 vhalf_hb = -75.0 q_h = 5.0 vhalf_inf = -72.0 qinf = 6.2 rate_ha = 0.0091 rate_hb = 0.024 a_h = rate_ha*(v-vhalf_ha)/(1-np.exp((vhalf_ha-v)/q_h)) b_h = rate_hb*(-v+vhalf_hb)/(1-np.exp((-vhalf_hb+v)/q_h)) h_inf = 1.0/(1.0 + np.exp((v-vhalf_inf)/qinf)) h_tau = 1./(a_h + b_h) dm = (m_inf-m)/m_tau dh = (h_inf-h)/h_tau return [dm, dh] " Kv models " def kdr_tf(Y,t,voltage_clamp_func,voltage_clamp_params): " Tigerholm version of the Sheets et al. IKdr model " " Model was developed from data recorded at 21 oC " v = voltage_clamp_func(t,voltage_clamp_params) n = Y[0] q10 = 1.0#3.3 # Preserved in case it is useful but disabled if v > -31.0: tau = 0.16+0.8*np.exp(-0.0267*(v+11)) else: tau = 1000*(0.000688 + 1/(np.exp((v+75.2)/6.5) + np.exp(-(v-131.5)/(34.8)))) ninf = 1/(1 + np.exp(-(v+45)/15.4)) ntau = tau/q10 dn = (ninf-n)/ntau return [dn] def km_tf(Y,t,voltage_clamp_func,voltage_clamp_params): """ Tigerholm version of the IM current. Current is from multiple sources: The voltage dependence of steady-state activation forthe KM current is from Maingret et al. (2008), which was derived from Passmore 2003. The KM channel activation has a fast and a slow time constant as described by Passmore et al. (2003). To account for the two time constants, weimplemented one fast (nf) and one slow (ns) gate, combined as follows. """ # g = gbar * (0.25*ns + 0.75*nf) v = voltage_clamp_func(t,voltage_clamp_params) ns = Y[0] nf = Y[1] q10 = 1.0#3.3 # Preserved in case it is useful but disabled if v < -60.0: nstau = 219.0*q10 else: nstau = 13.0*v + 1000.0*q10 nftau_alpha = 0.00395*np.exp((v+30.0)/40.0) nftau_beta = 0.00395*np.exp(-(v+30.0)/20.0)*q10 nftau = 1.0/(nftau_alpha + nftau_beta) ninf = 1.0/(1.0 + np.exp(-(v+30.0)/6.0)) # Threshold is around -30 mV dns = (ninf-ns)/nstau dnf = (ninf-nf)/nftau return [dns,dnf] def ka_tf(Y,t,voltage_clamp_func,voltage_clamp_params): """ Tigerholm version of IA. """ # g = gbar * n * h v = voltage_clamp_func(t,voltage_clamp_params) n = Y[0] h = Y[1] q10 = 1.0#3.3 # Preserved in case it is useful but disabled ninf = (1.0/(1.0 + np.exp(-(v+5.4+15)/16.4)))**4 ntau = 0.25 + 10.04*np.exp((-(v+24.67)**2)/(2*34.8**2))*q10 hinf = 1.0/(1.0 + np.exp((v+49.9 + 15.0)/4.6)) htau = 20.0 + 50.0 * np.exp((-(v+40.0)**2)/(2.0*40.0**2))*q10 # Trap for htau following Sheets /ChoiWaxman/Tigerholm - set it to 5 ms if less than 5 ms if htau < 5.0: htau = 5.0 dn = (ninf-n)/ntau dh = (hinf-h)/htau return [dn,dh] """ Ca models Implemented: cal_ja - Jaffe et al. 1994 ICaL model. can_mi - Model of N-type Ca current from Migliore 95 To do: SK BK Ca diffusion """ def cal_ja(Y,t,voltage_clamp_func,voltage_clamp_params): """ Jaffe et al. 1994 ICaL model. """ v = voltage_clamp_func(t,voltage_clamp_params) m = Y[0] tfa = 1. ki = 0.001 # (mM) cao = 2.5 # Davidson (mM) " To do: make cai variable as an input like voltage " cai = 1.e-4 # (mM) Roughly values (100 nM) from Intracellular calcium regulation among subpopulations of rat dorsal root ganglion neurons by Lu, Zhang, Gold 2007 celsius = 37. def alpha(v): return 15.69*(81.5 - v)/(np.exp((-1.0*v+81.5)/10.0)-1.0) def beta(v): return 0.29*np.exp(-v/10.86) def KTF(celsius): return ((25./293.15)*(celsius + 273.15)) def efun(z): return np.array([1 - i/2 if i < 1e-4 else i/(np.exp(i)-1) for i in z]) def calc_ghk(v, cai, cao): f = KTF(celsius)/2 nu = v/f return -f*(1. - (cai/cao)*np.exp(nu))*efun(nu) a = alpha(v) b = beta(v) tau = 1./(tfa*(a + b)) minf = a/(a+b) dm = (minf - m)/tau """ Calculating the current # h gate h2 = ki/(ki+cai) gcalbar = 0.003 ghk = calc_ghk(v,cai,cao) ical = gcalbar*m*m*h2*ghk """ return [dm] def
can_mi
identifier_name
current_models.py
np.exp((v+46.463)/8.8289)) minf = alpha_m/(alpha_m + beta_m) mtau = 1/(alpha_m + beta_m) hinf = 1/(1+np.exp((v+32.2)/4)) htau = 1.218 + 42.043*np.exp(-((v+38.1)**2)/(2*15.19**2)) alpha_s = 0.001 * 5.4203 / (1 + np.exp((v+79.816)/16.269)) beta_s = 0.001 * 5.0757 / (1 + np.exp(-(v+15.968)/11.542)) sinf = 1/(1+np.exp((v+45.0)/8)) stau = 1/(alpha_s + beta_s) alpha_u = 0.002 * 2.0434 / (1 + np.exp((v+67.499)/19.51)) beta_u = 0.002 * 1.9952 / (1 + np.exp(-(v+30.963)/14.792)) uinf = 1/(1+np.exp((v+51.0)/8)) utau = 1.0/(alpha_u + beta_u) dm = (minf-m)/mtau dh = (hinf-h)/htau ds = (sinf-s)/stau du = (uinf-u)/utau return [dm, dh, ds, du] def nav18cw(Y,t,voltage_clamp_func,voltage_clamp_params): " Nav 1.8 model used in Choi Waxman 2011 " v = voltage_clamp_func(t,voltage_clamp_params) m = Y[0] h = Y[1] alpha_m = 2.85 - 2.839/(1 + np.exp((v-1.159)/13.95)) beta_m = 7.6205/(1 + np.exp((v+46.463)/8.8289)) minf = alpha_m/(alpha_m + beta_m) mtau = 1/(alpha_m + beta_m) hinf = 1/(1+np.exp((v+32.2)/4)) htau = 1.218 + 42.043*np.exp(-((v+38.1)**2)/(2*15.19**2)) dm = (minf-m)/mtau dh = (hinf-h)/htau return [dm, dh] " -- Nav 1.9 models -- " def nav19hw(Y,t,voltage_clamp_func,voltage_clamp_params): " Nav 1.9 model from Huang Waxman 2014" m = Y[0] h = Y[1] s = Y[2] v = voltage_clamp_func(t,voltage_clamp_params) alpha_m = 0.751/(1 + np.exp(-(v+32.26)/13.71)) beta_m = 5.68/(1 + np.exp((v+123.71)/13.94)) minf = alpha_m/(alpha_m + beta_m) mtau = 1/(alpha_m + beta_m) alpha_h = 0.082/(1 + np.exp((v+113.69)/17.4)) beta_h = 0.24/(1 + np.exp(-(v-10.1)/17.2)) hinf = alpha_h/(alpha_h + beta_h) htau = 1/(alpha_h + beta_h) alpha_s = 0.019/(1 + np.exp((v+154.51)/11.46)) beta_s = 0.000376/(1 + np.exp(-(v+60.92)/15.79)) sinf = alpha_s/(alpha_s + beta_s) stau = 1/(alpha_s + beta_s) dm = (minf-m)/mtau dh = (hinf-h)/htau ds = (sinf-s)/stau return [dm, dh, ds] def nav19md(Y,t,voltage_clamp_func,voltage_clamp_params): " Nav 1.9 model from Maingret 2008" m = Y[0] h = Y[1] s = Y[2] v = voltage_clamp_func(t,voltage_clamp_params) return [dm, dh, ds] def nav16zm(Y,t,voltage_clamp_func,voltage_clamp_params): " Nav 1.6 model from Zach Mainen 1994 " m = Y[0] h = Y[1] v = voltage_clamp_func(t,voltage_clamp_params) vhalf = -43.0 a_m = 0.182*(v-vhalf)/(1-np.exp((vhalf-v)/6.)) b_m = 0.124*(-v+vhalf)/(1-np.exp((-vhalf+v)/6.)) m_inf = a_m/(a_m + b_m) m_tau = 1./(a_m + b_m) vhalf_ha = -50.0 vhalf_hb = -75.0 q_h = 5.0 vhalf_inf = -72.0 qinf = 6.2 rate_ha = 0.0091 rate_hb = 0.024 a_h = rate_ha*(v-vhalf_ha)/(1-np.exp((vhalf_ha-v)/q_h)) b_h = rate_hb*(-v+vhalf_hb)/(1-np.exp((-vhalf_hb+v)/q_h)) h_inf = 1.0/(1.0 + np.exp((v-vhalf_inf)/qinf)) h_tau = 1./(a_h + b_h) dm = (m_inf-m)/m_tau dh = (h_inf-h)/h_tau return [dm, dh] " Kv models " def kdr_tf(Y,t,voltage_clamp_func,voltage_clamp_params): " Tigerholm version of the Sheets et al. IKdr model " " Model was developed from data recorded at 21 oC " v = voltage_clamp_func(t,voltage_clamp_params) n = Y[0] q10 = 1.0#3.3 # Preserved in case it is useful but disabled if v > -31.0: tau = 0.16+0.8*np.exp(-0.0267*(v+11)) else: tau = 1000*(0.000688 + 1/(np.exp((v+75.2)/6.5) + np.exp(-(v-131.5)/(34.8)))) ninf = 1/(1 + np.exp(-(v+45)/15.4)) ntau = tau/q10 dn = (ninf-n)/ntau return [dn] def km_tf(Y,t,voltage_clamp_func,voltage_clamp_params):
""" Tigerholm version of the IM current. Current is from multiple sources: The voltage dependence of steady-state activation forthe KM current is from Maingret et al. (2008), which was derived from Passmore 2003. The KM channel activation has a fast and a slow time constant as described by Passmore et al. (2003). To account for the two time constants, weimplemented one fast (nf) and one slow (ns) gate, combined as follows. """ # g = gbar * (0.25*ns + 0.75*nf) v = voltage_clamp_func(t,voltage_clamp_params) ns = Y[0] nf = Y[1] q10 = 1.0#3.3 # Preserved in case it is useful but disabled if v < -60.0: nstau = 219.0*q10 else: nstau = 13.0*v + 1000.0*q10 nftau_alpha = 0.00395*np.exp((v+30.0)/40.0) nftau_beta = 0.00395*np.exp(-(v+30.0)/20.0)*q10
identifier_body
current_models.py
_params) n = Y[0] q10 = 1.0#3.3 # Preserved in case it is useful but disabled if v > -31.0: tau = 0.16+0.8*np.exp(-0.0267*(v+11)) else: tau = 1000*(0.000688 + 1/(np.exp((v+75.2)/6.5) + np.exp(-(v-131.5)/(34.8)))) ninf = 1/(1 + np.exp(-(v+45)/15.4)) ntau = tau/q10 dn = (ninf-n)/ntau return [dn] def km_tf(Y,t,voltage_clamp_func,voltage_clamp_params): """ Tigerholm version of the IM current. Current is from multiple sources: The voltage dependence of steady-state activation forthe KM current is from Maingret et al. (2008), which was derived from Passmore 2003. The KM channel activation has a fast and a slow time constant as described by Passmore et al. (2003). To account for the two time constants, weimplemented one fast (nf) and one slow (ns) gate, combined as follows. """ # g = gbar * (0.25*ns + 0.75*nf) v = voltage_clamp_func(t,voltage_clamp_params) ns = Y[0] nf = Y[1] q10 = 1.0#3.3 # Preserved in case it is useful but disabled if v < -60.0: nstau = 219.0*q10 else: nstau = 13.0*v + 1000.0*q10 nftau_alpha = 0.00395*np.exp((v+30.0)/40.0) nftau_beta = 0.00395*np.exp(-(v+30.0)/20.0)*q10 nftau = 1.0/(nftau_alpha + nftau_beta) ninf = 1.0/(1.0 + np.exp(-(v+30.0)/6.0)) # Threshold is around -30 mV dns = (ninf-ns)/nstau dnf = (ninf-nf)/nftau return [dns,dnf] def ka_tf(Y,t,voltage_clamp_func,voltage_clamp_params): """ Tigerholm version of IA. """ # g = gbar * n * h v = voltage_clamp_func(t,voltage_clamp_params) n = Y[0] h = Y[1] q10 = 1.0#3.3 # Preserved in case it is useful but disabled ninf = (1.0/(1.0 + np.exp(-(v+5.4+15)/16.4)))**4 ntau = 0.25 + 10.04*np.exp((-(v+24.67)**2)/(2*34.8**2))*q10 hinf = 1.0/(1.0 + np.exp((v+49.9 + 15.0)/4.6)) htau = 20.0 + 50.0 * np.exp((-(v+40.0)**2)/(2.0*40.0**2))*q10 # Trap for htau following Sheets /ChoiWaxman/Tigerholm - set it to 5 ms if less than 5 ms if htau < 5.0: htau = 5.0 dn = (ninf-n)/ntau dh = (hinf-h)/htau return [dn,dh] """ Ca models Implemented: cal_ja - Jaffe et al. 1994 ICaL model. can_mi - Model of N-type Ca current from Migliore 95 To do: SK BK Ca diffusion """ def cal_ja(Y,t,voltage_clamp_func,voltage_clamp_params): """ Jaffe et al. 1994 ICaL model. """ v = voltage_clamp_func(t,voltage_clamp_params) m = Y[0] tfa = 1. ki = 0.001 # (mM) cao = 2.5 # Davidson (mM) " To do: make cai variable as an input like voltage " cai = 1.e-4 # (mM) Roughly values (100 nM) from Intracellular calcium regulation among subpopulations of rat dorsal root ganglion neurons by Lu, Zhang, Gold 2007 celsius = 37. def alpha(v): return 15.69*(81.5 - v)/(np.exp((-1.0*v+81.5)/10.0)-1.0) def beta(v): return 0.29*np.exp(-v/10.86) def KTF(celsius): return ((25./293.15)*(celsius + 273.15)) def efun(z): return np.array([1 - i/2 if i < 1e-4 else i/(np.exp(i)-1) for i in z]) def calc_ghk(v, cai, cao): f = KTF(celsius)/2 nu = v/f return -f*(1. - (cai/cao)*np.exp(nu))*efun(nu) a = alpha(v) b = beta(v) tau = 1./(tfa*(a + b)) minf = a/(a+b) dm = (minf - m)/tau """ Calculating the current # h gate h2 = ki/(ki+cai) gcalbar = 0.003 ghk = calc_ghk(v,cai,cao) ical = gcalbar*m*m*h2*ghk """ return [dm] def can_mi(): """ Model of N-type Ca current from Migliore 95 """ pass " HCN models " def hcn_kn(Y,t,voltage_clamp_func,voltage_clamp_params): """ Kouranova Ih model with non-specific current (reversal potential should be set at -30 mV """ v = voltage_clamp_func(t,voltage_clamp_params) n_s = Y[0] n_f = Y[1] ninf_s = 1/(1 + np.exp((v+87.2)/9.7)) ninf_f = ninf_s if v > -70.0: tau_ns = 300.0 + 542.0 * np.exp((v+25.0)/20.0) tau_nf = 140.0 + 50.0 * np.exp(-(v+25.0)/20.0) else: tau_ns = 2500.0 + 100.0 * np.exp((v+240.0)/50.0) tau_nf = 250.0 + 12.0 * np.exp((v+240.0)/50.0) dns = (ninf_s - n_s)/tau_ns dnf = (ninf_f - n_f)/tau_nf return [dns, dnf] def hcn_tf(Y,t,voltage_clamp_func,voltage_clamp_params): """ Tigerholm version of the Kouranova Ih model which is identical except that when you calculate the current you don't use a nonspecific reversal potential and instead split the current between Na+ and K+, 50/50. """ v = voltage_clamp_func(t,voltage_clamp_params) n_s = Y[0] n_f = Y[1] ninf_s = 1/(1 + np.exp((v+87.2)/9.7)) ninf_f = ninf_s if v > -70.0: tau_ns = 300.0 + 542.0 * np.exp((v+25.0)/20.0) tau_nf = 140.0 + 50.0 * np.exp(-(v+25.0)/20.0) else:
tau_ns = 2500.0 + 100.0 * np.exp((v+240.0)/50.0) tau_nf = 250.0 + 12.0 * np.exp((v+240.0)/50.0)
conditional_block
keyword_plan.pb.go
func (m *KeywordPlan) XXX_DiscardUnknown() { xxx_messageInfo_KeywordPlan.DiscardUnknown(m) } var xxx_messageInfo_KeywordPlan proto.InternalMessageInfo func (m *KeywordPlan) GetResourceName() string { if m != nil { return m.ResourceName } return "" } func (m *KeywordPlan) GetId() *wrappers.Int64Value { if m != nil { return m.Id } return nil } func (m *KeywordPlan) GetName() *wrappers.StringValue { if m != nil { return m.Name } return nil } func (m *KeywordPlan) GetForecastPeriod() *KeywordPlanForecastPeriod { if m != nil { return m.ForecastPeriod } return nil } // The forecasting period associated with the keyword plan. type KeywordPlanForecastPeriod struct { // Required. The date used for forecasting the Plan. // // Types that are valid to be assigned to Interval: // *KeywordPlanForecastPeriod_DateInterval // *KeywordPlanForecastPeriod_DateRange Interval isKeywordPlanForecastPeriod_Interval `protobuf_oneof:"interval"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` } func (m *KeywordPlanForecastPeriod) Reset() { *m = KeywordPlanForecastPeriod{} } func (m *KeywordPlanForecastPeriod) String() string { return proto.CompactTextString(m) } func (*KeywordPlanForecastPeriod) ProtoMessage() {} func (*KeywordPlanForecastPeriod) Descriptor() ([]byte, []int) { return fileDescriptor_db2ef87e79a4b462, []int{1} } func (m *KeywordPlanForecastPeriod) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_KeywordPlanForecastPeriod.Unmarshal(m, b) } func (m *KeywordPlanForecastPeriod) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { return xxx_messageInfo_KeywordPlanForecastPeriod.Marshal(b, m, deterministic) } func (m *KeywordPlanForecastPeriod) XXX_Merge(src proto.Message) { xxx_messageInfo_KeywordPlanForecastPeriod.Merge(m, src) } func (m *KeywordPlanForecastPeriod) XXX_Size() int { return xxx_messageInfo_KeywordPlanForecastPeriod.Size(m) } func (m *KeywordPlanForecastPeriod) XXX_DiscardUnknown() { xxx_messageInfo_KeywordPlanForecastPeriod.DiscardUnknown(m) } var xxx_messageInfo_KeywordPlanForecastPeriod proto.InternalMessageInfo type isKeywordPlanForecastPeriod_Interval interface { isKeywordPlanForecastPeriod_Interval() } type KeywordPlanForecastPeriod_DateInterval struct { DateInterval enums.KeywordPlanForecastIntervalEnum_KeywordPlanForecastInterval `protobuf:"varint,1,opt,name=date_interval,json=dateInterval,proto3,enum=google.ads.googleads.v2.enums.KeywordPlanForecastIntervalEnum_KeywordPlanForecastInterval,oneof"` } type KeywordPlanForecastPeriod_DateRange struct { DateRange *common.DateRange `protobuf:"bytes,2,opt,name=date_range,json=dateRange,proto3,oneof"` } func (*KeywordPlanForecastPeriod_DateInterval) isKeywordPlanForecastPeriod_Interval() {} func (*KeywordPlanForecastPeriod_DateRange) isKeywordPlanForecastPeriod_Interval() {} func (m *KeywordPlanForecastPeriod) GetInterval() isKeywordPlanForecastPeriod_Interval { if m != nil { return m.Interval } return nil } func (m *KeywordPlanForecastPeriod) GetDateInterval() enums.KeywordPlanForecastIntervalEnum_KeywordPlanForecastInterval { if x, ok := m.GetInterval().(*KeywordPlanForecastPeriod_DateInterval); ok { return x.DateInterval } return enums.KeywordPlanForecastIntervalEnum_UNSPECIFIED } func (m *KeywordPlanForecastPeriod) GetDateRange() *common.DateRange { if x, ok := m.GetInterval().(*KeywordPlanForecastPeriod_DateRange); ok { return x.DateRange } return nil } // XXX_OneofWrappers is for the internal use of the proto package. func (*KeywordPlanForecastPeriod) XXX_OneofWrappers() []interface{} { return []interface{}{ (*KeywordPlanForecastPeriod_DateInterval)(nil), (*KeywordPlanForecastPeriod_DateRange)(nil), } } func init() { proto.RegisterType((*KeywordPlan)(nil), "google.ads.googleads.v2.resources.KeywordPlan") proto.RegisterType((*KeywordPlanForecastPeriod)(nil), "google.ads.googleads.v2.resources.KeywordPlanForecastPeriod") } func init() { proto.RegisterFile("google/ads/googleads/v2/resources/keyword_plan.proto", fileDescriptor_db2ef87e79a4b462) } var fileDescriptor_db2ef87e79a4b462 = []byte{ // 485 bytes of a gzipped FileDescriptorProto 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0x84, 0x93, 0xdd, 0x6a, 0xd4, 0x40, 0x14, 0xc7, 0x9b, 0xb4, 0x88, 0x9d, 0x7e, 0x28, 0xb9, 0x5a, 0x6b, 0x91, 0xb6, 0x52, 0xa8, 0x0a, 0x13, 0x89, 0xc5, 0x8b, 0xe8, 0x4d, 0x16, 0xb5, 0x1f, 0x82, 0x2c, 0x11, 0xf6, 0xa2, 0x2c, 0x2c, 0xd3, 0x9d, 0xb3, 0x21, 0x98, 0xcc, 0x84, 0x99, 0xc9, 0x16, 0x2f, 0x7d, 0x15, 0x2f, 0x7d, 0x14, 0x1f, 0xc5, 0x17, 0xd0, 0x1b, 0x41, 0x32, 0x5f, 0xb4, 0xd8, 0x74, 0xef, 0xce, 0xd9, 0xf9, 0x9d, 0xff, 0xff, 0x7c, 0x64, 0xd1, 0x71, 0xc1, 0x79, 0x51, 0x41, 0x4c, 0xa8, 0x8c, 0x4d, 0xd8, 0x45, 0x8b, 0x24, 0x16, 0x20, 0x79, 0x2b, 0x66, 0x20, 0xe3, 0x2f, 0xf0, 0xf5, 0x8a, 0x0b, 0x3a, 0x6d, 0x2a, 0xc2, 0x70, 0x23, 0xb8, 0xe2, 0xd1, 0xbe, 0x41, 0x31, 0xa1, 0x12, 0xfb, 0x2a, 0xbc, 0x48, 0xb0, 0xaf, 0xda, 0x79, 0xde, 0x27, 0x3c, 0xe3, 0x75, 0xcd, 0x59, 0x4c, 0x89, 0x02, 0x69, 0xe4, 0x76, 0x86, 0x7d, 0x2c, 0xb0, 0xb6, 0xbe, 0xd9, 0xc0, 0x74, 0xce, 0x05, 0xcc, 0x88, 0x54, 0xd3, 0x92, 0x29, 0x10, 0x0b, 0x52, 0x59, 0x8d, 0x27, 0x56, 0x43, 0x67, 0x97, 0xed, 0x3c, 0xbe, 0
{ return xxx_messageInfo_KeywordPlan.Size(m) }
identifier_body