id
stringlengths
2
7
text
stringlengths
17
51.2k
title
stringclasses
1 value
c176700
ops := []bakery.Op{{ Entity: r.URL.Path, Action: r.Method, }} info, err := authChecker.Allow(ctx, ops...) if err != nil { // Bad macaroon return false, "", "", err } if info != nil && info.Identity != nil { // Valid identity macaroon found return true, info.Identity.Id(), "candid", nil } // Valid macaroon with no identity information return true, "", "candid", nil } // Validate normal TLS access for i := range r.TLS.PeerCertificates { trusted, username := util.CheckTrustState(*r.TLS.PeerCertificates[i], d.clientCerts) if trusted { return true, username, "tls", nil } } // Reject unauthorized return false, "", "", nil }
c176701
state.NewState(d.db, d.cluster, d.maas, d.os, d.endpoints) }
c176702
path } return filepath.Join(d.os.VarDir, "unix.socket") }
c176703
} logger.Infof("Closing the database") err := d.cluster.Close() // If we got io.EOF the network connection was interrupted and // it's likely that the other node shutdown. Let's just log the // event and return cleanly. if errors.Cause(err) == driver.ErrBadConn { logger.Debugf("Could not close remote database cleanly: %v", err) } else { trackError(err) } } if d.db != nil { trackError(d.db.Close()) } if d.gateway != nil { trackError(d.gateway.Shutdown()) } if d.endpoints != nil { trackError(d.endpoints.Down()) } if d.endpoints != nil { trackError(d.endpoints.Down()) } if shouldUnmount { logger.Infof("Unmounting temporary filesystems") syscall.Unmount(shared.VarPath("devlxd"), syscall.MNT_DETACH) syscall.Unmount(shared.VarPath("shmounts"), syscall.MNT_DETACH) logger.Infof("Done unmounting temporary filesystems") } else { logger.Debugf( "Not unmounting temporary filesystems (containers are still running)") } var err error if n := len(errs); n > 0 { format := "%v" if n > 1 { format += fmt.Sprintf(" (and %d more errors)", n) } err = fmt.Errorf(format, errs[0]) } if err != nil { logger.Errorf("Failed to cleanly shutdown daemon: %v", err) } return err }
c176704
PublicKey: bakery.PublicKey{Key: pkKey}, Version: 3, }) // Allow http URLs if we have a public key set if strings.HasPrefix(authEndpoint, "http://") { pkLocator.AllowInsecure() } } // Setup the bakery bakery := identchecker.NewBakery(identchecker.BakeryParams{ Key: key, Location: authEndpoint, Locator: pkLocator, Checker: httpbakery.NewChecker(), IdentityClient: idmClientWrapper, Authorizer: identchecker.ACLAuthorizer{ GetACL: func(ctx context.Context, op bakery.Op) ([]string, bool, error) { return []string{identchecker.Everyone}, false, nil }, }, }) // Store our settings d.externalAuth = &externalAuth{ endpoint: authEndpoint, expiry: expiry, bakery: bakery, } return nil }
c176705
} } for _, i := range legacyPatchesNeedingDB { legacy[i].NeedsDB = true } // Hook to run when the local database is created from scratch. It will // create the default profile and mark all patches as applied. freshHook := func(db *db.Node) error { for _, patchName := range patchesGetNames() { err := db.PatchesMarkApplied(patchName) if err != nil { return err } } return nil } var err error var dump *db.Dump d.db, dump, err = db.OpenNode(filepath.Join(d.os.VarDir, "database"), freshHook, legacy) if err != nil { return nil, fmt.Errorf("Error creating database: %s", err) } return dump, nil }
c176706
} err := json.NewEncoder(output).Encode(body) if captured != nil { shared.DebugJson(captured) } return err }
c176707
"", err } return fmt.Sprintf("%x", etag.Sum(nil)), nil }
c176708
err } if hash != match { return fmt.Errorf("ETag doesn't match: %s vs %s", hash, match) } return nil }
c176709
DisableKeepAlives: true, } myhttp := http.Client{ Transport: tr, } // Setup redirect policy myhttp.CheckRedirect = func(req *http.Request, via []*http.Request) error { // Replicate the headers req.Header = via[len(via)-1].Header return nil } return &myhttp, nil }
c176710
recursion, err := strconv.Atoi(recursionStr) if err != nil { return false } return recursion != 0 }
c176711
file := os.NewFile(uintptr(i), fmt.Sprintf("inherited-fd%d", i)) listener, err := net.FileListener(file) if err != nil { continue } listeners = append(listeners, listener) } return listeners }
c176712
schema = node.FreshSchema() } tx, err := db.Begin() if err != nil { return SmartError(errors.Wrap(err, "failed to start transaction")) } defer tx.Rollback() dump, err := query.Dump(tx, schema, schemaOnly == 1) if err != nil { return SmartError(errors.Wrapf(err, "failed dump database %s", database)) } return SyncResponse(true, internalSQLDump{Text: dump}) }
c176713
!= nil { return SmartError(err) } if strings.HasPrefix(strings.ToUpper(query), "SELECT") { err = internalSQLSelect(tx, query, &result) tx.Rollback() } else { err = internalSQLExec(tx, query, &result) if err != nil { tx.Rollback() } else { err = tx.Commit() } } if err != nil { return SmartError(err) } batch.Results = append(batch.Results, result) } return SyncResponse(true, batch) }
c176714
pem.EncodeToMemory(&pem.Block{Type: "CERTIFICATE", Bytes: data}) }
c176715
ok { data := x509.MarshalPKCS1PrivateKey(rsaKey) return pem.EncodeToMemory(&pem.Block{Type: "RSA PRIVATE KEY", Bytes: data}) } return nil }
c176716
fail, since we generated the cert ourselves, // but let's check the error for good measure. if err != nil { panic("invalid public key material") } return fingerprint }
c176717
certf, err) } certOut.Write(certBytes) certOut.Close() keyOut, err := os.OpenFile(keyf, os.O_WRONLY|os.O_CREATE|os.O_TRUNC, 0600) if err != nil { return fmt.Errorf("Failed to open %s for writing: %v", keyf, err) } keyOut.Write(keyBytes) keyOut.Close() return nil }
c176718
%s\n", env.KernelVersion) fmt.Printf(" Storage backend: %s\n", env.Storage) fmt.Printf(" Storage version: %s\n", env.StorageVersion) fmt.Printf(" Container backend: %s\n", env.Driver) fmt.Printf(" Container version: %s\n", env.DriverVersion) fmt.Printf("\n") return nil }
c176719
*sync.WaitGroup) { defer wg.Done() name := getContainerName(count, index) err := createContainer(c, fingerprint, name, privileged) if err != nil { logf("Failed to launch container '%s': %s", name, err) return } if start { err := startContainer(c, name) if err != nil { logf("Failed to start container '%s': %s", name, err) return } if freeze { err := freezeContainer(c, name) if err != nil { logf("Failed to freeze container '%s': %s", name, err) return } } } } duration = processBatch(count, batchSize, batchStart) return duration, nil }
c176720
} batchCreate := func(index int, wg *sync.WaitGroup) { defer wg.Done() name := getContainerName(count, index) err := createContainer(c, fingerprint, name, privileged) if err != nil { logf("Failed to launch container '%s': %s", name, err) return } } duration = processBatch(count, batchSize, batchCreate) return duration, nil }
c176721
range allContainers { if container.Config[userConfigKey] == "true" { containers = append(containers, container) } } return containers, nil }
c176722
count) batchStart := func(index int, wg *sync.WaitGroup) { defer wg.Done() container := containers[index] if !container.IsActive() { err := startContainer(c, container.Name) if err != nil { logf("Failed to start container '%s': %s", container.Name, err) return } } } duration = processBatch(count, batchSize, batchStart) return duration, nil }
c176723
return "", err } values := fields.Query() values.Set(param, url.QueryEscape(value)) fields.RawQuery = values.Encode() return fields.String(), nil }
c176724
nil, "", &images) if err != nil { return nil, err } return images, nil }
c176725
return r.GetPrivateImageFile(fingerprint, "", req) }
c176726
"", err } opAPI := op.Get() return opAPI.Metadata["secret"].(string), nil }
c176727
!= nil { return nil, "", err } } // Fetch the raw value etag, err := r.queryStruct("GET", path, nil, "", &image) if err != nil { return nil, "", err } return &image, etag, nil }
c176728
if err == nil { resp, err := lxdDownloadImage(fingerprint, unixURI, r.httpUserAgent, devlxdHTTP, req) if err == nil { return resp, nil } } } // Build the URL uri = fmt.Sprintf("%s%s", r.httpHost, uri) if secret != "" { uri, err = setQueryParam(uri, "secret", secret) if err != nil { return nil, err } } return lxdDownloadImage(fingerprint, uri, r.httpUserAgent, r.http, req) }
c176729
err := r.queryStruct("GET", "/images/aliases?recursion=1", nil, "", &aliases) if err != nil { return nil, err } return aliases, nil }
c176730
// Forward targetOp to remote op go func() { success := false errors := map[string]error{} for _, serverURL := range urls { req.Source.Server = serverURL op, err := r.CreateImage(req, nil) if err != nil { errors[serverURL] = err continue } rop.targetOp = op for _, handler := range rop.handlers { rop.targetOp.AddHandler(handler) } err = rop.targetOp.Wait() if err != nil { errors[serverURL] = err continue } success = true break } if !success { rop.err = remoteOperationError("Failed remote image download", errors) } close(rop.chDone) }() return &rop, nil }
c176731
if err != nil { return nil, err } req.Source.Secret = secret } // Process the arguments if args != nil { req.Aliases = args.Aliases req.AutoUpdate = args.AutoUpdate req.Public = args.Public if args.CopyAliases { req.Aliases = image.Aliases if args.Aliases != nil { req.Aliases = append(req.Aliases, args.Aliases...) } } } return r.tryCopyImage(req, info.Addresses) }
c176732
request _, _, err := r.query("PUT", fmt.Sprintf("/images/%s", url.QueryEscape(fingerprint)), image, ETag) if err != nil { return err } return nil }
c176733
:= r.queryOperation("DELETE", fmt.Sprintf("/images/%s", url.QueryEscape(fingerprint)), nil, "") if err != nil { return nil, err } return op, nil }
c176734
r.queryOperation("POST", fmt.Sprintf("/images/%s/refresh", url.QueryEscape(fingerprint)), nil, "") if err != nil { return nil, err } return op, nil }
c176735
the request _, _, err := r.query("POST", "/images/aliases", alias, "") if err != nil { return err } return nil }
c176736
url.QueryEscape(name)), alias, ETag) if err != nil { return err } return nil }
c176737
request _, _, err := r.query("POST", fmt.Sprintf("/images/aliases/%s", url.QueryEscape(name)), alias, "") if err != nil { return err } return nil }
c176738
:= r.query("DELETE", fmt.Sprintf("/images/aliases/%s", url.QueryEscape(name)), nil, "") if err != nil { return err } return nil }
c176739
return nil, fmt.Errorf("cannot open node database: %v", err) } return db, nil }
c176740
:= shared.FileCopy(path, path+".bak") if err != nil { return err } backupDone = true } if version == -1 { logger.Debugf("Running pre-update queries from file for local DB schema") } else { logger.Debugf("Updating DB schema from %d to %d", version, version+1) } // Run the given hook only against actual update versions, not // when a custom query file is passed (signaled by version == -1). if hook != nil && version != -1 { err := hook(version, tx) if err != nil { } } return nil }) return schema.Ensure(db) }
c176741
return "ext4", nil case FilesystemSuperMagicXfs: return "xfs", nil case FilesystemSuperMagicNfs: return "nfs", nil default: logger.Debugf("Unknown backing filesystem type: 0x%x", fs.Type) return string(fs.Type), nil } }
c176742
schema.Fresh(freshSchema) return schema }
c176743
err = tx.Exec(` INSERT INTO config (key, value) SELECT 'cluster.https_address', value FROM config WHERE key = 'core.https_address' `) if err != nil { return errors.Wrap(err, "Insert cluster.https_address config") } return nil }
c176744
ArchitectureDefault, err } return uname.Machine, nil }
c176745
nil, err } // Find the right machine machines, err := srv.Machines(gomaasapi.MachinesArgs{Hostnames: []string{machine}}) if err != nil { return nil, err } if len(machines) != 1 { return nil, fmt.Errorf("Couldn't find the specified machine: %s", machine) } // Setup the struct c := Controller{} c.srv = srv c.srvRaw = *srvRaw c.machine = machines[0] c.url = baseURL return &c, err }
c176746
{ return } c.DeleteContainer(name) }() // Create the rest of the interfaces for _, iface := range interfaces[1:] { _, err := device.CreateInterface(gomaasapi.CreateInterfaceArgs{ Name: iface.Name, MACAddress: iface.MACAddress, VLAN: subnets[iface.Subnets[0].Name].VLAN(), }) if err != nil { return err } } // Get a fresh copy of the device device, err = c.getDevice(name) if err != nil { return err } // Setup the interfaces for _, entry := range device.InterfaceSet() { // Get our record iface, ok := macInterfaces[entry.MACAddress()] if !ok { return fmt.Errorf("MAAS created an interface with a bad MAC: %s", entry.MACAddress()) } // Add the subnets for _, subnet := range iface.Subnets { err := entry.LinkSubnet(gomaasapi.LinkSubnetArgs{ Mode: gomaasapi.LinkModeStatic, Subnet: subnets[subnet.Name], IPAddress: subnet.Address, }) if err != nil { return err } } } success = true return nil }
c176747
return false, err } if len(devs) == 1 { return true, nil } return false, nil }
c176748
_, ok := existingSubnets[subnet.Name] if ok { continue } // Add the link err := entry.LinkSubnet(gomaasapi.LinkSubnetArgs{ Mode: gomaasapi.LinkModeStatic, Subnet: subnets[subnet.Name], IPAddress: subnet.Address, }) if err != nil { return err } } // Record the interface has being configured existingInterfaces[entry.MACAddress()] = entry } // Iterate over expected interfaces, add any missing one for _, iface := range macInterfaces { _, ok := existingInterfaces[iface.MACAddress] if ok { // We already have it so just move on continue } // Create the new interface entry, err := device.CreateInterface(gomaasapi.CreateInterfaceArgs{ Name: iface.Name, MACAddress: iface.MACAddress, VLAN: subnets[iface.Subnets[0].Name].VLAN(), }) if err != nil { return err } // Add the subnets for _, subnet := range iface.Subnets { err := entry.LinkSubnet(gomaasapi.LinkSubnetArgs{ Mode: gomaasapi.LinkModeStatic, Subnet: subnets[subnet.Name], IPAddress: subnet.Address, }) if err != nil { return err } } } return nil }
c176749
return err } values := url.Values{} values.Set("hostname", newName) _, err = c.srvRaw.Put(uri, values) if err != nil { return err } return nil }
c176750
if err != nil { return err } err = device.Delete() if err != nil { return err } return nil }
c176751
append(s.updates, update) }
c176752
:= createSchemaTable(tx) if err != nil { return fmt.Errorf("failed to create schema table: %v", err) } } return nil }
c176753
if count == 1 { // Insert the missing version. err := insertSchemaVersion(tx, 38) if err != nil { return -1, fmt.Errorf("Failed to insert missing schema version 38") } versions = append(versions, 38) } } current := 0 if len(versions) > 0 { err = checkSchemaVersionsHaveNoHoles(versions) if err != nil { return -1, err } current = versions[len(versions)-1] // Highest recorded version } return current, nil }
c176754
for _, update := range updates[current:] { if hook != nil { err := hook(current, tx) if err != nil { return fmt.Errorf( "failed to execute hook (version %d): %v", current, err) } } err := update(tx) if err != nil { return fmt.Errorf("failed to apply update %d: %v", current, err) } current++ err = insertSchemaVersion(tx, current) if err != nil { return fmt.Errorf("failed to insert version %d", current) } } return nil }
c176755
for i := range versions[:len(versions)-1] { if versions[i+1] != versions[i]+1 { return fmt.Errorf("Missing updates: %d to %d", versions[i], versions[i+1]) } } return nil }
c176756
return fmt.Errorf("expected schema table to contain at least one row") } err = checkSchemaVersionsHaveNoHoles(versions) if err != nil { return err } current := versions[len(versions)-1] if current != len(updates) { return fmt.Errorf("update level is %d, expected %d", current, len(updates)) } return nil }
c176757
lines[i] = strings.Replace(line, ", ", ",\n ", -1) } return strings.Join(lines, "\n") }
c176758
split = split[:len(split)-1] } xattrs = make(map[string]string, len(split)) for _, x := range split { xattr := string(x) // Call Getxattr() twice: First, to determine the size of the // buffer we need to allocate to store the extended attributes, // second, to actually store the extended attributes in the // buffer. Also, check if the size of the extended attribute // hasn't changed between the two calls. pre, err = syscall.Getxattr(path, xattr, nil) if err != nil || pre < 0 { return nil, err } dest = make([]byte, pre) post := 0 if pre > 0 { post, err = syscall.Getxattr(path, xattr, dest) if err != nil || post < 0 { return nil, err } } if post != pre { return nil, e1 } xattrs[xattr] = string(dest) } return xattrs, nil }
c176759
} tmpErrno, ok := err.(syscall.Errno) if ok { return tmpErrno, true } return nil, false }
c176760
Sysname: intArrayToString(uname.Sysname), Nodename: intArrayToString(uname.Nodename), Release: intArrayToString(uname.Release), Version: intArrayToString(uname.Version), Machine: intArrayToString(uname.Machine), Domainname: intArrayToString(uname.Domainname), }, nil }
c176761
stmts[code] = sql return code }
c176762
return nil, errors.Wrapf(err, "%q", sql) } index[code] = stmt } return index, nil }
c176763
ctx: ctx, cancel: cancel, upgradeCh: make(chan struct{}, 16), acceptCh: make(chan net.Conn), store: &dqliteServerStore{}, } err := gateway.init() if err != nil { return nil, err } return gateway, nil }
c176764
g.memoryDial(ctx, address) } return dqliteNetworkDial(ctx, address, g.cert) } }
c176765
g.Sync() g.server.Close() // Unset the memory dial, since Shutdown() is also called for // switching between in-memory and network mode. g.memoryDial = nil } return nil }
c176766
g.server.Dump("db.bin", dir) if err != nil { // Just log a warning, since this is not fatal. logger.Warnf("Failed to dump database to disk: %v", err) } }
c176767
error { return tx.RaftNodesReplace(nil) }) if err != nil { return err } g.cert = cert return g.init() }
c176768
node.Address) } return nil }) if err != nil { return "", errors.Wrap(err, "Failed to fetch raft nodes addresses") } if len(addresses) == 0 { // This should never happen because the raft_nodes table should // be never empty for a clustered node, but check it for good // measure. return "", fmt.Errorf("No raft node known") } for _, address := range addresses { url := fmt.Sprintf("https://%s%s", address, databaseEndpoint) request, err := http.NewRequest("GET", url, nil) if err != nil { return "", err } request = request.WithContext(ctx) client := &http.Client{Transport: &http.Transport{TLSClientConfig: config}} response, err := client.Do(request) if err != nil { logger.Debugf("Failed to fetch leader address from %s", address) continue } if response.StatusCode != http.StatusOK { logger.Debugf("Request for leader address from %s failed", address) continue } info := map[string]string{} err = shared.ReadToJSON(response.Body, &info) if err != nil { logger.Debugf("Failed to parse leader address from %s", address) continue } leader := info["leader"] if leader == "" { logger.Debugf("Raft node %s returned no leader address", address) continue } return leader, nil } return "", fmt.Errorf("RAFT cluster is unavailable") }
c176769
return fmt.Errorf("RAFT node did not self-elect within %s", time.Duration(n)*sleep) }
c176770
{ if err != db.ErrNoSuchObject { return nil, errors.Wrap(err, "Failed to fetch raft server address") } // Use the initial address as fallback. This is an edge // case that happens when a new leader is elected and // its raft_nodes table is not fully up-to-date yet. address = server.Address } id, err := strconv.Atoi(string(server.ID)) if err != nil { return nil, errors.Wrap(err, "Non-numeric server ID") } nodes[i].ID = int64(id) nodes[i].Address = string(address) } return nodes, nil }
c176771
if err != nil { return nil, errors.Wrap(err, "Failed to fetch raft nodes") } return addresses, nil }
c176772
error) { return net.Dial("unix", listener.Addr().String()) } }
c176773
logger.Debugf(format, a...) case dqlite.LogWarn: logger.Warnf(format, a...) case dqlite.LogError: logger.Errorf(format, a...) } }
c176774
r.MetadataAsStruct(&ret) if err != nil { return nil, err } return ret, nil }
c176775
Operation{} err := r.MetadataAsStruct(&op) if err != nil { return nil, err } return &op, nil }
c176776
[]string{} err := r.MetadataAsStruct(&sl) if err != nil { return nil, err } return sl, nil }
c176777
json.Unmarshal(r.Metadata, &target) }
c176778
} err = r.addRecord(record) if err != nil { return err } } logf("Loaded report file %s", r.Filename) return nil }
c176779
writer.WriteAll(r.records) if err != nil { return err } logf("Written report file %s", r.Filename) return nil }
c176780
fmt.Sprintf("%d", time.Now().UnixNano()/int64(time.Millisecond)), // timestamp fmt.Sprintf("%d", elapsed/time.Millisecond), label, "", // responseCode is not used "true", // success" } return r.addRecord(record) }
c176781
c.Remotes[k] = v } // NOTE: Remove this once we only see a small fraction of non-simplestreams users // Upgrade users to the "simplestreams" protocol images, ok := c.Remotes["images"] if ok && images.Protocol != ImagesRemote.Protocol && images.Addr == ImagesRemote.Addr { c.Remotes["images"] = ImagesRemote c.SaveConfig(path) } return c, nil }
c176782
%v", err) } defer f.Close() // Write the new config data, err := yaml.Marshal(conf) if err != nil { return fmt.Errorf("Unable to marshal the configuration: %v", err) } _, err = f.Write(data) if err != nil { return fmt.Errorf("Unable to write the configuration: %v", err) } return nil }
c176783
return nil, err } // Validate that we're under the expected prefix if !strings.HasPrefix(path, basePath) { return nil, fmt.Errorf("Attempting to access a file outside the container") } // Open and read the file buf, err := ioutil.ReadFile(path) if err != nil { return nil, err } return bytes.NewReader(buf), nil }
c176784
= append(path, paths...) return filepath.Join(path...) }
c176785
{ return c.ConfigPath("servercerts", fmt.Sprintf("%s.crt", remote)) }
c176786
DefaultRemotes config.DefaultRemote = "local" } return config }
c176787
of memory pages is below the defined threshold. tmp = s.container.ExpandedConfig()["migration.incremental.memory.iterations"] if tmp != "" { max_iterations, _ = strconv.Atoi(tmp) } else { // default to 10 max_iterations = 10 } if max_iterations > 999 { // the pre-dump directory is hardcoded to a string // with maximal 3 digits. 999 pre-dumps makes no // sense at all, but let's make sure the number // is not higher than this. max_iterations = 999 } logger.Debugf("Using maximal %d iterations for pre-dumping", max_iterations) return use_pre_dumps, max_iterations }
c176788
be pre-copied for the pre-copy migration to stop. var threshold int tmp := s.container.ExpandedConfig()["migration.incremental.memory.goal"] if tmp != "" { threshold, _ = strconv.Atoi(tmp) } else { // defaults to 70% threshold = 70 } if percentage_skipped > threshold { logger.Debugf("Memory pages skipped (%d%%) due to pre-copy is larger than threshold (%d%%)", percentage_skipped, threshold) logger.Debugf("This was the last pre-dump; next dump is the final dump") final = true } // If in pre-dump mode, the receiving side // expects a message to know if this was the // last pre-dump logger.Debugf("Sending another header") sync := migration.MigrationSync{ FinalPreDump: proto.Bool(final), } data, err := proto.Marshal(&sync) if err != nil { return final, err } err = s.criuConn.WriteMessage(websocket.BinaryMessage, data) if err != nil { s.sendControl(err) return final, err } logger.Debugf("Sending another header done") return final, nil }
c176789
entry point for all "go:generate" directives used in LXD's source code.`, RunE: func(cmd *cobra.Command, args []string) error { return fmt.Errorf("Not implemented") }, } cmd.AddCommand(newDb()) return cmd }
c176790
artificialBump := os.Getenv("LXD_ARTIFICIALLY_BUMP_API_EXTENSIONS") if artificialBump != "" { n, err := strconv.Atoi(artificialBump) if err == nil { count += n } } return count }
c176791
:= []string{} for rows.Next() { err := rows.Scan(dest...) if err != nil { return nil, errors.Wrapf(err, "Failed to scan URI params") } uri := f(params...) uris = append(uris, uri) } err = rows.Err() if err != nil { return nil, errors.Wrapf(err, "Failed to close URI result set") } return uris, nil }
c176792
err := rows.Scan(&value) if err != nil { return err } values = append(values, value) return nil } err := scanSingleColumn(tx, query, args, "TEXT", scan) if err != nil { return nil, err } return values, nil }
c176793
for rows.Next() { err := scan(rows) if err != nil { return err } } err = rows.Err() if err != nil { return err } return nil }
c176794
:= r.Ctx[i].(Lazy) if ok { v, err := evaluateLazy(lz) if err != nil { hadErr = true r.Ctx[i] = err } else { if cs, ok := v.(stack.Trace); ok { v = cs.TrimBelow(stack.Call(r.CallPC[0])). TrimRuntime() } r.Ctx[i] = v } } } if hadErr { r.Ctx = append(r.Ctx, errorKey, "bad lazy") } return h.Log(r) }) }
c176795
pc := range pcs[:n] { cs[i] = Call(pc) } putPoolBuf(pcs) return cs }
c176796
if fn == nil { return "???" } return fn.Name() }
c176797
for len(pcs) > 0 && pcs[0] != pc { pcs = pcs[1:] } return pcs }
c176798
pcs = pcs[:len(pcs)-1] } return pcs }
c176799
for len(pcs) > 0 && pcs[0].name() != name { pcs = pcs[1:] } return pcs }