id
stringlengths
2
7
text
stringlengths
17
51.2k
title
stringclasses
1 value
c174800
if ev.Kv.ModRevision >= w.minRev { // don't double notify wb.add(w, ev) } } } return wb }
c174801
{ iv.Val.(watcherSet).add(wa) return } // not registered, put in interval tree ws := make(watcherSet) ws.add(wa) wg.ranges.Insert(ivl, ws) }
c174802
|| wg.ranges.Intersects(adt.NewStringAffinePoint(key)) }
c174803
ws := iv.Val.(watcherSet) delete(ws, wa) if len(ws) == 0 { // remove interval missing watchers if ok := wg.ranges.Delete(ivl); !ok { panic("could not remove watcher from interval tree") } } return true }
c174804
if maxWatchers <= 0 { break } maxWatchers-- ret.add(w) } return &ret, ret.chooseAll(curRev, compactRev) }
c174805
reuse single-key set return wkeys case len(wranges) == 0 && len(wkeys) == 0: return nil case len(wranges) == 1 && len(wkeys) == 0: return wranges[0].Val.(watcherSet) } // copy case ret := make(watcherSet) ret.union(wg.keyWatchers[key]) for _, item := range wranges { ret.union(item.Val.(watcherSet)) } return ret }
c174806
iveCmpBegin := ivl.End.Compare(ivl2.Begin) // ivl is left of ivl2 if ivbCmpBegin < 0 && iveCmpBegin <= 0 { return -1 } // iv is right of iv2 if ivbCmpEnd >= 0 { return 1 } return 0 }
c174807
for y != nil && x == y.right { x = y y = y.parent } return y }
c174808
&& x.right.max.Compare(max) > 0 { max = x.right.max } if oldmax.Compare(max) == 0 { break } x.max = max x = x.parent } }
c174809
== 0 { if !x.left.visit(iv, nv) || !x.right.visit(iv, nv) { return false } } default: if !x.left.visit(iv, nv) || !nv(x) || !x.right.visit(iv, nv) { return false } } return true }
c174810
y.parent.left = x } else { y.parent.right = x } y.parent.updateMax() } if y != z { z.iv = y.iv z.updateMax() } if y.color() == black && x != nil { ivt.deleteFixup(x) } ivt.count-- return true }
c174811
ivt.root = z } else { if z.iv.Ivl.Begin.Compare(y.iv.Ivl.Begin) < 0 { y.left = z } else { y.right = z } y.updateMax() } z.c = red ivt.insertFixup(z) ivt.count++ }
c174812
if y.left != nil { y.left.parent = x } x.updateMax() ivt.replaceParent(x, y) y.left = x y.updateMax() }
c174813
{ x.parent.left = y } else { x.parent.right = y } x.parent.updateMax() } x.parent = y }
c174814
* math.Log2(float64(ivt.Len()+1))) + 0.5) }
c174815
*intervalNode) bool { return ivv(&n.iv) }) }
c174816
ret = n return false } ivt.root.visit(&ivl, f) return ret }
c174817
{ n := ivt.find(ivl) if n == nil { return nil } return &n.iv }
c174818
&& x.left.max.Compare(iv.Begin) > 0 { x = x.left } else { x = x.right } } return x != nil }
c174819
> 0 { maxEnd = n.Ivl.End } return true }) return isContiguous && minBegin != nil && maxEnd.Compare(ivl.End) >= 0 && minBegin.Compare(ivl.Begin) <= 0 }
c174820
{ return nil } f := func(n *IntervalValue) bool { ivs = append(ivs, n); return true } ivt.Visit(iv, f) return ivs }
c174821
f := func(n *IntervalValue) bool { ivt.Insert(n.Ivl, n.Val) return true } inIvt.Visit(ivl, f) }
c174822
&exactReadCloser{rc: rc, totalBytes: totalBytes} }
c174823
{ return &Election{session: s, keyPrefix: pfx + "/"} }
c174824
s, leaderKey: leaderKey, leaderRev: leaderRev, leaderSession: s, } }
c174825
e.leaderRev) txn := client.Txn(ctx).If(cmp) txn = txn.Then(v3.OpPut(e.leaderKey, val, v3.WithLease(e.leaderSession.Lease()))) tresp, terr := txn.Commit() if terr != nil { return terr } if !tresp.Succeeded { e.leaderKey = "" return ErrElectionNotLeader } e.hdr = tresp.Header return nil }
c174826
err := client.Txn(ctx).If(cmp).Then(v3.OpDelete(e.leaderKey)).Commit() if err == nil { e.hdr = resp.Header } e.leaderKey = "" e.leaderSession = nil return err }
c174827
v3.WithFirstCreate()...) if err != nil { return nil, err } else if len(resp.Kvs) == 0 { // no leader currently elected return nil, ErrElectionNoLeader } return resp, nil }
c174828
make(chan v3.GetResponse) go e.observe(ctx, retc) return retc }
c174829
pb.AlarmRequest_ACTIVATE, Alarm: pb.AlarmType_NOSPACE, } qa.a.Alarm(ctx, req) return rpctypes.ErrGRPCNoSpace }
c174830
"recursive, r", Usage: "watch all values for key and child keys"}, }, Action: func(c *cli.Context) error { execWatchCommandFunc(c, mustNewKeyAPI(c)) return nil }, } }
c174831
!= nil { handleError(c, ExitServerError, err) } if resp.Node.Dir { fmt.Fprintf(os.Stderr, "Ignored dir %s change\n", resp.Node.Key) continue } cmd := exec.Command(cmdArgs[0], cmdArgs[1:]...) cmd.Env = environResponse(resp, os.Environ()) cmd.Stdout = os.Stdout cmd.Stderr = os.Stderr go func() { err := cmd.Start() if err != nil { fmt.Fprintf(os.Stderr, err.Error()) os.Exit(1) } cmd.Wait() }() } }
c174832
u.Scheme, tlsinfo, ConnReadTimeout, ConnWriteTimeout) }
c174833
time to write out before reading out the response. return transport.NewTimeoutTransport(tlsInfo, dialTimeout, 0, 0) }
c174834
request error (%v)", err) } req.Header.Set("Content-Type", ct) req.Header.Set("X-Server-From", from.String()) req.Header.Set("X-Server-Version", version.Version) req.Header.Set("X-Min-Cluster-Version", version.MinClusterVersion) req.Header.Set("X-Etcd-Cluster-ID", cid.String()) setPeerURLsHeader(req, urls) return req }
c174835
plog.Errorf("request sent was ignored (cluster ID mismatch: remote[%s]=%s, local=%s)", to, resp.Header.Get("X-Etcd-Cluster-ID"), req.Header.Get("X-Etcd-Cluster-ID")) return errClusterIDMismatch default: return fmt.Errorf("unhandled error %q when precondition failed", string(body)) } case http.StatusForbidden: return errMemberRemoved case http.StatusNoContent: return nil default: return fmt.Errorf("unexpected http status %s while posting to %q", http.StatusText(resp.StatusCode), req.URL.String()) } }
c174836
// backward compatibility with etcd 2.0 if verStr == "" { verStr = "2.0.0" } return semver.Must(semver.NewVersion(verStr)) }
c174837
local=%s", name, server, localServer) } if compareMajorMinorVersion(minCluster, localServer) == 1 { return localServer, localMinCluster, fmt.Errorf("local version is too low: remote[%s]=%s, local=%s", name, server, localServer) } return localServer, localMinCluster, nil }
c174838
} peerURLs := make([]string, urls.Len()) for i := range urls { peerURLs[i] = urls[i].String() } req.Header.Set("X-PeerURLs", strings.Join(peerURLs, ",")) }
c174839
r.Header.Get("X-PeerURLs"); urls != "" { tr.AddRemote(from, strings.Split(urls, ",")) } } }
c174840
&httpKeysAPI{ client: c, prefix: p, } }
c174841
{ return time.Duration(n.TTL) * time.Second }
c174842
if serr := setFlagFromEnv(fs, prefix, f.Name, usedEnvKey, alreadySet, false); serr != nil { err = serr } }) verifyEnv(prefix, usedEnvKey, alreadySet) return err }
c174843
strings.ToUpper(strings.Replace(name, "-", "_", -1)) }
c174844
return fmt.Sprintf("%q...%q", str[:pre], str[len(str)-suf:]) }
c174845
== raftpb.EntryConfChange, "ConfigChange" }
c174846
fmt.Printf("%4d\t%10d\tnorm\t%s", entry.Term, entry.Index, rr.String()) } }
c174847
range ents { passed := false currtype := "" for _, filter := range entryFilters { passed, currtype = filter(e) if passed { cnt++ break } } if passed { printer := printerMap[currtype] printer(e) if streamdecoder == "" { fmt.Println() continue } // if decoder is set, pass the e.Data to stdin and read the stdout from decoder io.WriteString(stdin, hex.EncodeToString(e.Data)) io.WriteString(stdin, "\n") outputReader := bufio.NewReader(stdout) decoderoutput, currerr := outputReader.ReadString('\n') if currerr != nil { fmt.Println(currerr) return } decoder_status, decoded_data := parseDecoderOutput(decoderoutput) fmt.Printf("\t%s\t%s", decoder_status, decoded_data) } } stdin.Close() err = cmd.Wait() if streamdecoder != "" { if err != nil { log.Panic(err) } if stderr.String() != "" { os.Stderr.WriteString("decoder stderr: " + stderr.String()) } } fmt.Printf("\nEntry types (%s) count is : %d", entrytype, cnt) }
c174848
newLogWithSize(storage, logger, noLimit) }
c174849
= lastIndex + 1 log.unstable.logger = logger // Initialize our committed and applied pointers to the time of the last compaction. log.committed = firstIndex - 1 log.applied = firstIndex - 1 return log }
c174850
ne.Index, l.zeroTermOnErrCompacted(l.term(ne.Index)), ne.Term) } return ne.Index } } return 0 }
c174851
nil { l.logger.Panicf("unexpected error when getting unapplied entries (%v)", err) } return ents } return nil }
c174852
a racing compaction return l.allEntries() } // TODO (xiangli): handle error? panic(err) }
c174853
lo, min(hi, l.unstable.offset)) } else if err != nil { panic(err) // TODO(bdarnell) } // check if ents has reached the size limitation if uint64(len(storedEnts)) < min(hi, l.unstable.offset)-lo { return storedEnts, nil } ents = storedEnts } if hi > l.unstable.offset { unstable := l.unstable.slice(max(lo, l.unstable.offset), hi) if len(ents) > 0 { combined := make([]pb.Entry, len(ents)+len(unstable)) n := copy(combined, ents) copy(combined[n:], unstable) ents = combined } else { ents = unstable } } return limitSize(ents, maxSize), nil }
c174854
donec := make(chan struct{}) s := &Session{client: client, opts: ops, id: id, cancel: cancel, donec: donec} // keep the lease alive until client error or cancelled context go func() { defer close(donec) for range keepAlive { // eat messages until keep alive channel closes } }() return s, nil }
c174855
:= context.WithTimeout(s.opts.ctx, time.Duration(s.opts.ttl)*time.Second) _, err := s.client.Revoke(ctx, s.id) cancel() return err }
c174856
if ttl > 0 { so.ttl = ttl } } }
c174857
*sessionOptions) { so.leaseID = leaseID } }
c174858
= &readIndexStatus{index: index, req: m, acks: make(map[uint64]struct{})} ro.readIndexQueue = append(ro.readIndexQueue, ctx) }
c174859
if !ok { return 0 } rs.acks[m.From] = struct{}{} // add one to include an ack from local node return len(rs.acks) + 1 }
c174860
= true break } } if found { ro.readIndexQueue = ro.readIndexQueue[i:] for _, rs := range rss { delete(ro.pendingReadIndex, string(rs.req.Entries[0].Data)) } return rss } return nil }
c174861
return ro.readIndexQueue[len(ro.readIndexQueue)-1] }
c174862
}) s.goAttach(s.purgeFile) s.goAttach(func() { monitorFileDescriptor(s.getLogger(), s.stopping) }) s.goAttach(s.monitorVersions) s.goAttach(s.linearizableReadLoop) s.goAttach(s.monitorKVHash) }
c174863
zap.String("cluster-version", version.Cluster(s.ClusterVersion().String())), ) } else { plog.Infof("starting server... [version: %v, cluster version: %v]", version.Version, version.Cluster(s.ClusterVersion().String())) } membership.ClusterVersionMetrics.With(prometheus.Labels{"cluster_version": s.ClusterVersion().String()}).Set(1) } else { if lg != nil { lg.Info( "starting etcd server", zap.String("local-member-id", s.ID().String()), zap.String("local-server-version", version.Version), zap.String("cluster-version", "to_be_decided"), ) } else { plog.Infof("starting server... [version: %v, cluster version: to_be_decided]", version.Version) } } // TODO: if this is an empty log, writes all peer infos // into the first entry go s.run() }
c174864
zap.String("removed-member-id", types.ID(m.From).String()), ) } else { plog.Warningf("reject message from removed member %s", types.ID(m.From).String()) } return httptypes.NewHTTPError(http.StatusForbidden, "cannot process message from removed member") } if m.Type == raftpb.MsgApp { s.stats.RecvAppendReq(types.ID(m.From).String(), m.Size()) } return s.r.Step(ctx, m) }
c174865
s.r.ReportSnapshot(id, status) }
c174866
zap.String("local-member-id", s.ID().String()), zap.String("old-leader-member-id", types.ID(lead).String()), zap.String("new-leader-member-id", types.ID(transferee).String()), zap.Duration("took", time.Since(now)), ) } else { plog.Infof("%s finished leadership transfer from %s to %s (took %v)", s.ID(), types.ID(lead), types.ID(transferee), time.Since(now)) } return nil }
c174867
zap.String("current-leader-member-id", types.ID(s.Lead()).String()), ) } else { plog.Printf("skipped leadership transfer for single member cluster") } return nil } transferee, ok := longestConnected(s.r.transport, s.cluster.MemberIDs()) if !ok { return ErrUnhealthy } tm := s.Cfg.ReqTimeout() ctx, cancel := context.WithTimeout(s.ctx, tm) err := s.MoveLeader(ctx, s.Lead(), uint64(transferee)) cancel() return err }
c174868
else { plog.Panicf("configure trigger value should never be nil") } } resp := x.(*confChangeResponse) if lg := s.getLogger(); lg != nil { lg.Info( "applied a configuration change through raft", zap.String("local-member-id", s.ID().String()), zap.String("raft-conf-change", cc.Type.String()), zap.String("raft-conf-change-node-id", types.ID(cc.NodeID).String()), ) } return resp.membs, resp.err case <-ctx.Done(): s.w.Trigger(cc.ID, nil) // GC wait return nil, s.parseProposeCtxErr(ctx.Err(), start) case <-s.stopping: return nil, ErrStopped } }
c174869
it uses goroutine to propose. ctx, cancel := context.WithTimeout(s.ctx, timeout) s.goAttach(func() { s.r.Propose(ctx, data) cancel() }) }
c174870
"stopped publish because server is stopped", zap.String("local-member-id", s.ID().String()), zap.String("local-member-attributes", fmt.Sprintf("%+v", s.attributes)), zap.Duration("publish-timeout", timeout), zap.Error(err), ) } else { plog.Infof("aborting publish because server is stopped") } return default: if lg := s.getLogger(); lg != nil { lg.Warn( "failed to publish local member to cluster through raft", zap.String("local-member-id", s.ID().String()), zap.String("local-member-attributes", fmt.Sprintf("%+v", s.attributes)), zap.String("request-path", req.Path), zap.Duration("publish-timeout", timeout), zap.Error(err), ) } else { plog.Errorf("publish error: %v", err) } } } }
c174871
if needResult || !noSideEffect(&raftReq) { if !needResult && raftReq.Txn != nil { removeNeedlessRangeReqs(raftReq.Txn) } ar = s.applyV3.Apply(&raftReq) } if ar == nil { return } if ar.err != ErrNoSpace || len(s.alarmStore.Get(pb.AlarmType_NOSPACE)) > 0 { s.w.Trigger(id, ar) return } if lg := s.getLogger(); lg != nil { lg.Warn( "message exceeded backend quota; raising alarm", zap.Int64("quota-size-bytes", s.Cfg.QuotaBackendBytes), zap.String("quota-size", humanize.Bytes(uint64(s.Cfg.QuotaBackendBytes))), zap.Error(ar.err), ) } else { plog.Errorf("applying raft message exceeded backend quota") } s.goAttach(func() { a := &pb.AlarmRequest{ MemberID: uint64(s.ID()), Action: pb.AlarmRequest_ACTIVATE, Alarm: pb.AlarmType_NOSPACE, } s.raftRequest(s.ctx, pb.InternalRaftRequest{Alarm: a}) s.w.Trigger(id, ar) }) }
c174872
{ return true, nil } s.r.transport.RemovePeer(id) case raftpb.ConfChangeUpdateNode: m := new(membership.Member) if err := json.Unmarshal(cc.Context, m); err != nil { if lg != nil { lg.Panic("failed to unmarshal member", zap.Error(err)) } else { plog.Panicf("unmarshal member should never fail: %v", err) } } if cc.NodeID != uint64(m.ID) { if lg != nil { lg.Panic( "got different member ID", zap.String("member-id-from-config-change-entry", types.ID(cc.NodeID).String()), zap.String("member-id-from-message", m.ID.String()), ) } else { plog.Panicf("nodeID should always be equal to member ID") } } s.cluster.UpdateRaftAttributes(m.ID, m.RaftAttributes) if m.ID != s.id { s.r.transport.UpdatePeer(m.ID, m.PeerURLs) } } return false, nil }
c174873
verStr = v.String() } s.goAttach(func() { s.updateClusterVersion(verStr) }) continue } // update cluster version only if the decided version is greater than // the current cluster version if v != nil && s.cluster.Version().LessThan(*v) { s.goAttach(func() { s.updateClusterVersion(v.String()) }) } } }
c174874
plog.Warning("server has stopped (skipping goAttach)") } return default: } // now safe to add since waitgroup wait has not started yet s.wg.Add(1) go func() { defer s.wg.Done() f() }() }
c174875
scToAddr map[balancer.SubConn]resolver.Address, ) Picker { return &rrBalanced{ lg: lg, scs: scs, addrToSc: addrToSc, scToAddr: scToAddr, } }
c174876
:= []zapcore.Field{ zap.Error(info.Err), zap.String("address", picked), zap.Bool("success", info.Err == nil), zap.Bool("bytes-sent", info.BytesSent), zap.Bool("bytes-received", info.BytesReceived), } if info.Err == nil { rb.lg.Debug("balancer done", fss...) } else { rb.lg.Warn("balancer failed", fss...) } } return sc, doneFunc, nil }
c174877
{ check := func(context.Context, *tls.Conn) error { return nil } return newTLSListener(l, tlsinfo, check) }
c174878
herr := tlsConn.Handshake() pendingMu.Lock() delete(pending, conn) pendingMu.Unlock() if herr != nil { l.handshakeFailure(tlsConn, herr) return } if err := l.check(ctx, tlsConn); err != nil { l.handshakeFailure(tlsConn, err) return } select { case l.connc <- tlsConn: conn = nil case <-ctx.Done(): } }() } }
c174879
r := range e.resolvers { r.cc.NewAddress(addrs) } e.mu.Unlock() }
c174880
return Target(e.id, endpoint) }
c174881
{ return fmt.Sprintf("%s://%s/%s", scheme, id, endpoint) }
c174882
:= b.getResolverGroup(id) if err != nil { return nil, fmt.Errorf("failed to build resolver: %v", err) } r := &Resolver{ endpointID: id, cc: cc, } es.addResolver(r) return r, nil }
c174883
r) case "QGET": return v2api.QGet(ctx, r) case "GET": return v2api.Get(ctx, r) case "HEAD": return v2api.Head(ctx, r) } return Response{}, ErrUnknownMethod }
c174884
} cmd.Flags().IntVar(&totalClientConnections, "total-client-connections", 10, "total number of client connections") return cmd }
c174885
return nil, fmt.Errorf("unmarshal raftAttributes error: %v", err) } } else { return nil, fmt.Errorf("raftAttributes key doesn't exist") } if data := attrs[attrKey]; data != nil { if err := json.Unmarshal(data, &m.Attributes); err != nil { return m, fmt.Errorf("unmarshal attributes error: %v", err) } } return m, nil }
c174886
bcfg.Path, bcfg.BatchInterval, bcfg.BatchLimit = tmpPath, batchInterval, batchLimit return newBackend(bcfg), tmpPath }
c174887
c: c, } rc.ctx, rc.cancel = context.WithCancel(context.Background()) return rc }
c174888
zap.Int64("revision-compaction-retention", rc.retention), zap.Duration("took", time.Since(now)), ) } else { plog.Noticef("Finished auto-compaction at revision %d", rev) } } else { if rc.lg != nil { rc.lg.Warn( "failed auto revision compaction", zap.Int64("revision", rev), zap.Int64("revision-compaction-retention", rc.retention), zap.Duration("retry-interval", revInterval), zap.Error(err), ) } else { plog.Noticef("Failed auto-compaction at revision %d (%v)", rev, err) plog.Noticef("Retry after %v", revInterval) } } } }() }
c174889
= true rc.mu.Unlock() }
c174890
= false rc.mu.Unlock() }
c174891
pb.MsgPreVoteResp default: panic(fmt.Sprintf("not a vote message: %s", msgt)) } }
c174892
if i != 0 { buf.WriteString(", ") } buf.WriteString(DescribeEntry(e, f)) } fmt.Fprintf(&buf, "]") } if !IsEmptySnap(m.Snapshot) { fmt.Fprintf(&buf, " Snapshot:%v", m.Snapshot) } return buf.String() }
c174893
fmt.Sprintf("%q", e.Data) } return fmt.Sprintf("%d/%d %s %s", e.Term, e.Index, e.Type, formatted) }
c174894
_, _ = buf.WriteString(DescribeEntry(e, f) + "\n") } return buf.String() }
c174895
with locking grpclog.SetLoggerV2(lg) lgMu.Unlock() }
c174896
lgMu.RLock() l := lg lgMu.RUnlock() return l }
c174897
u.snapshot != nil { return u.snapshot.Metadata.Index + 1, true } return 0, false }
c174898
return u.offset + uint64(l) - 1, true } if u.snapshot != nil { return u.snapshot.Metadata.Index, true } return 0, false }
c174899
} last, ok := u.maybeLastIndex() if !ok { return 0, false } if i > last { return 0, false } return u.entries[i-u.offset].Term, true }