repo
stringlengths
5
54
path
stringlengths
4
155
func_name
stringlengths
1
118
original_string
stringlengths
52
85.5k
language
stringclasses
1 value
code
stringlengths
52
85.5k
code_tokens
list
docstring
stringlengths
6
2.61k
docstring_tokens
list
sha
stringlengths
40
40
url
stringlengths
85
252
partition
stringclasses
1 value
syndtr/goleveldb
leveldb/cache/cache.go
Delete
func (r *Cache) Delete(ns, key uint64, onDel func()) bool { r.mu.RLock() defer r.mu.RUnlock() if r.closed { return false } hash := murmur32(ns, key, 0xf00) for { h, b := r.getBucket(hash) done, _, n := b.get(r, h, hash, ns, key, true) if done { if n != nil { if onDel != nil { n.mu.Lock() n.onDel = append(n.onDel, onDel) n.mu.Unlock() } if r.cacher != nil { r.cacher.Ban(n) } n.unref() return true } break } } if onDel != nil { onDel() } return false }
go
func (r *Cache) Delete(ns, key uint64, onDel func()) bool { r.mu.RLock() defer r.mu.RUnlock() if r.closed { return false } hash := murmur32(ns, key, 0xf00) for { h, b := r.getBucket(hash) done, _, n := b.get(r, h, hash, ns, key, true) if done { if n != nil { if onDel != nil { n.mu.Lock() n.onDel = append(n.onDel, onDel) n.mu.Unlock() } if r.cacher != nil { r.cacher.Ban(n) } n.unref() return true } break } } if onDel != nil { onDel() } return false }
[ "func", "(", "r", "*", "Cache", ")", "Delete", "(", "ns", ",", "key", "uint64", ",", "onDel", "func", "(", ")", ")", "bool", "{", "r", ".", "mu", ".", "RLock", "(", ")", "\n", "defer", "r", ".", "mu", ".", "RUnlock", "(", ")", "\n", "if", "r", ".", "closed", "{", "return", "false", "\n", "}", "\n\n", "hash", ":=", "murmur32", "(", "ns", ",", "key", ",", "0xf00", ")", "\n", "for", "{", "h", ",", "b", ":=", "r", ".", "getBucket", "(", "hash", ")", "\n", "done", ",", "_", ",", "n", ":=", "b", ".", "get", "(", "r", ",", "h", ",", "hash", ",", "ns", ",", "key", ",", "true", ")", "\n", "if", "done", "{", "if", "n", "!=", "nil", "{", "if", "onDel", "!=", "nil", "{", "n", ".", "mu", ".", "Lock", "(", ")", "\n", "n", ".", "onDel", "=", "append", "(", "n", ".", "onDel", ",", "onDel", ")", "\n", "n", ".", "mu", ".", "Unlock", "(", ")", "\n", "}", "\n", "if", "r", ".", "cacher", "!=", "nil", "{", "r", ".", "cacher", ".", "Ban", "(", "n", ")", "\n", "}", "\n", "n", ".", "unref", "(", ")", "\n", "return", "true", "\n", "}", "\n\n", "break", "\n", "}", "\n", "}", "\n\n", "if", "onDel", "!=", "nil", "{", "onDel", "(", ")", "\n", "}", "\n\n", "return", "false", "\n", "}" ]
// Delete removes and ban 'cache node' with the given namespace and key. // A banned 'cache node' will never inserted into the 'cache tree'. Ban // only attributed to the particular 'cache node', so when a 'cache node' // is recreated it will not be banned. // // If onDel is not nil, then it will be executed if such 'cache node' // doesn't exist or once the 'cache node' is released. // // Delete return true is such 'cache node' exist.
[ "Delete", "removes", "and", "ban", "cache", "node", "with", "the", "given", "namespace", "and", "key", ".", "A", "banned", "cache", "node", "will", "never", "inserted", "into", "the", "cache", "tree", ".", "Ban", "only", "attributed", "to", "the", "particular", "cache", "node", "so", "when", "a", "cache", "node", "is", "recreated", "it", "will", "not", "be", "banned", ".", "If", "onDel", "is", "not", "nil", "then", "it", "will", "be", "executed", "if", "such", "cache", "node", "doesn", "t", "exist", "or", "once", "the", "cache", "node", "is", "released", ".", "Delete", "return", "true", "is", "such", "cache", "node", "exist", "." ]
c3a204f8e96543bb0cc090385c001078f184fc46
https://github.com/syndtr/goleveldb/blob/c3a204f8e96543bb0cc090385c001078f184fc46/leveldb/cache/cache.go#L419-L453
train
syndtr/goleveldb
leveldb/cache/cache.go
Evict
func (r *Cache) Evict(ns, key uint64) bool { r.mu.RLock() defer r.mu.RUnlock() if r.closed { return false } hash := murmur32(ns, key, 0xf00) for { h, b := r.getBucket(hash) done, _, n := b.get(r, h, hash, ns, key, true) if done { if n != nil { if r.cacher != nil { r.cacher.Evict(n) } n.unref() return true } break } } return false }
go
func (r *Cache) Evict(ns, key uint64) bool { r.mu.RLock() defer r.mu.RUnlock() if r.closed { return false } hash := murmur32(ns, key, 0xf00) for { h, b := r.getBucket(hash) done, _, n := b.get(r, h, hash, ns, key, true) if done { if n != nil { if r.cacher != nil { r.cacher.Evict(n) } n.unref() return true } break } } return false }
[ "func", "(", "r", "*", "Cache", ")", "Evict", "(", "ns", ",", "key", "uint64", ")", "bool", "{", "r", ".", "mu", ".", "RLock", "(", ")", "\n", "defer", "r", ".", "mu", ".", "RUnlock", "(", ")", "\n", "if", "r", ".", "closed", "{", "return", "false", "\n", "}", "\n\n", "hash", ":=", "murmur32", "(", "ns", ",", "key", ",", "0xf00", ")", "\n", "for", "{", "h", ",", "b", ":=", "r", ".", "getBucket", "(", "hash", ")", "\n", "done", ",", "_", ",", "n", ":=", "b", ".", "get", "(", "r", ",", "h", ",", "hash", ",", "ns", ",", "key", ",", "true", ")", "\n", "if", "done", "{", "if", "n", "!=", "nil", "{", "if", "r", ".", "cacher", "!=", "nil", "{", "r", ".", "cacher", ".", "Evict", "(", "n", ")", "\n", "}", "\n", "n", ".", "unref", "(", ")", "\n", "return", "true", "\n", "}", "\n\n", "break", "\n", "}", "\n", "}", "\n\n", "return", "false", "\n", "}" ]
// Evict evicts 'cache node' with the given namespace and key. This will // simply call Cacher.Evict. // // Evict return true is such 'cache node' exist.
[ "Evict", "evicts", "cache", "node", "with", "the", "given", "namespace", "and", "key", ".", "This", "will", "simply", "call", "Cacher", ".", "Evict", ".", "Evict", "return", "true", "is", "such", "cache", "node", "exist", "." ]
c3a204f8e96543bb0cc090385c001078f184fc46
https://github.com/syndtr/goleveldb/blob/c3a204f8e96543bb0cc090385c001078f184fc46/leveldb/cache/cache.go#L459-L484
train
syndtr/goleveldb
leveldb/cache/cache.go
EvictNS
func (r *Cache) EvictNS(ns uint64) { r.mu.RLock() defer r.mu.RUnlock() if r.closed { return } if r.cacher != nil { r.cacher.EvictNS(ns) } }
go
func (r *Cache) EvictNS(ns uint64) { r.mu.RLock() defer r.mu.RUnlock() if r.closed { return } if r.cacher != nil { r.cacher.EvictNS(ns) } }
[ "func", "(", "r", "*", "Cache", ")", "EvictNS", "(", "ns", "uint64", ")", "{", "r", ".", "mu", ".", "RLock", "(", ")", "\n", "defer", "r", ".", "mu", ".", "RUnlock", "(", ")", "\n", "if", "r", ".", "closed", "{", "return", "\n", "}", "\n\n", "if", "r", ".", "cacher", "!=", "nil", "{", "r", ".", "cacher", ".", "EvictNS", "(", "ns", ")", "\n", "}", "\n", "}" ]
// EvictNS evicts 'cache node' with the given namespace. This will // simply call Cacher.EvictNS.
[ "EvictNS", "evicts", "cache", "node", "with", "the", "given", "namespace", ".", "This", "will", "simply", "call", "Cacher", ".", "EvictNS", "." ]
c3a204f8e96543bb0cc090385c001078f184fc46
https://github.com/syndtr/goleveldb/blob/c3a204f8e96543bb0cc090385c001078f184fc46/leveldb/cache/cache.go#L488-L498
train
syndtr/goleveldb
leveldb/cache/cache.go
EvictAll
func (r *Cache) EvictAll() { r.mu.RLock() defer r.mu.RUnlock() if r.closed { return } if r.cacher != nil { r.cacher.EvictAll() } }
go
func (r *Cache) EvictAll() { r.mu.RLock() defer r.mu.RUnlock() if r.closed { return } if r.cacher != nil { r.cacher.EvictAll() } }
[ "func", "(", "r", "*", "Cache", ")", "EvictAll", "(", ")", "{", "r", ".", "mu", ".", "RLock", "(", ")", "\n", "defer", "r", ".", "mu", ".", "RUnlock", "(", ")", "\n", "if", "r", ".", "closed", "{", "return", "\n", "}", "\n\n", "if", "r", ".", "cacher", "!=", "nil", "{", "r", ".", "cacher", ".", "EvictAll", "(", ")", "\n", "}", "\n", "}" ]
// EvictAll evicts all 'cache node'. This will simply call Cacher.EvictAll.
[ "EvictAll", "evicts", "all", "cache", "node", ".", "This", "will", "simply", "call", "Cacher", ".", "EvictAll", "." ]
c3a204f8e96543bb0cc090385c001078f184fc46
https://github.com/syndtr/goleveldb/blob/c3a204f8e96543bb0cc090385c001078f184fc46/leveldb/cache/cache.go#L501-L511
train
syndtr/goleveldb
leveldb/cache/cache.go
Close
func (r *Cache) Close() error { r.mu.Lock() if !r.closed { r.closed = true h := (*mNode)(r.mHead) h.initBuckets() for i := range h.buckets { b := (*mBucket)(h.buckets[i]) for _, n := range b.node { // Call releaser. if n.value != nil { if r, ok := n.value.(util.Releaser); ok { r.Release() } n.value = nil } // Call OnDel. for _, f := range n.onDel { f() } n.onDel = nil } } } r.mu.Unlock() // Avoid deadlock. if r.cacher != nil { if err := r.cacher.Close(); err != nil { return err } } return nil }
go
func (r *Cache) Close() error { r.mu.Lock() if !r.closed { r.closed = true h := (*mNode)(r.mHead) h.initBuckets() for i := range h.buckets { b := (*mBucket)(h.buckets[i]) for _, n := range b.node { // Call releaser. if n.value != nil { if r, ok := n.value.(util.Releaser); ok { r.Release() } n.value = nil } // Call OnDel. for _, f := range n.onDel { f() } n.onDel = nil } } } r.mu.Unlock() // Avoid deadlock. if r.cacher != nil { if err := r.cacher.Close(); err != nil { return err } } return nil }
[ "func", "(", "r", "*", "Cache", ")", "Close", "(", ")", "error", "{", "r", ".", "mu", ".", "Lock", "(", ")", "\n", "if", "!", "r", ".", "closed", "{", "r", ".", "closed", "=", "true", "\n\n", "h", ":=", "(", "*", "mNode", ")", "(", "r", ".", "mHead", ")", "\n", "h", ".", "initBuckets", "(", ")", "\n\n", "for", "i", ":=", "range", "h", ".", "buckets", "{", "b", ":=", "(", "*", "mBucket", ")", "(", "h", ".", "buckets", "[", "i", "]", ")", "\n", "for", "_", ",", "n", ":=", "range", "b", ".", "node", "{", "// Call releaser.", "if", "n", ".", "value", "!=", "nil", "{", "if", "r", ",", "ok", ":=", "n", ".", "value", ".", "(", "util", ".", "Releaser", ")", ";", "ok", "{", "r", ".", "Release", "(", ")", "\n", "}", "\n", "n", ".", "value", "=", "nil", "\n", "}", "\n\n", "// Call OnDel.", "for", "_", ",", "f", ":=", "range", "n", ".", "onDel", "{", "f", "(", ")", "\n", "}", "\n", "n", ".", "onDel", "=", "nil", "\n", "}", "\n", "}", "\n", "}", "\n", "r", ".", "mu", ".", "Unlock", "(", ")", "\n\n", "// Avoid deadlock.", "if", "r", ".", "cacher", "!=", "nil", "{", "if", "err", ":=", "r", ".", "cacher", ".", "Close", "(", ")", ";", "err", "!=", "nil", "{", "return", "err", "\n", "}", "\n", "}", "\n", "return", "nil", "\n", "}" ]
// Close closes the 'cache map' and forcefully releases all 'cache node'.
[ "Close", "closes", "the", "cache", "map", "and", "forcefully", "releases", "all", "cache", "node", "." ]
c3a204f8e96543bb0cc090385c001078f184fc46
https://github.com/syndtr/goleveldb/blob/c3a204f8e96543bb0cc090385c001078f184fc46/leveldb/cache/cache.go#L514-L550
train
syndtr/goleveldb
leveldb/cache/cache.go
CloseWeak
func (r *Cache) CloseWeak() error { r.mu.Lock() if !r.closed { r.closed = true } r.mu.Unlock() // Avoid deadlock. if r.cacher != nil { r.cacher.EvictAll() if err := r.cacher.Close(); err != nil { return err } } return nil }
go
func (r *Cache) CloseWeak() error { r.mu.Lock() if !r.closed { r.closed = true } r.mu.Unlock() // Avoid deadlock. if r.cacher != nil { r.cacher.EvictAll() if err := r.cacher.Close(); err != nil { return err } } return nil }
[ "func", "(", "r", "*", "Cache", ")", "CloseWeak", "(", ")", "error", "{", "r", ".", "mu", ".", "Lock", "(", ")", "\n", "if", "!", "r", ".", "closed", "{", "r", ".", "closed", "=", "true", "\n", "}", "\n", "r", ".", "mu", ".", "Unlock", "(", ")", "\n\n", "// Avoid deadlock.", "if", "r", ".", "cacher", "!=", "nil", "{", "r", ".", "cacher", ".", "EvictAll", "(", ")", "\n", "if", "err", ":=", "r", ".", "cacher", ".", "Close", "(", ")", ";", "err", "!=", "nil", "{", "return", "err", "\n", "}", "\n", "}", "\n", "return", "nil", "\n", "}" ]
// CloseWeak closes the 'cache map' and evict all 'cache node' from cacher, but // unlike Close it doesn't forcefully releases 'cache node'.
[ "CloseWeak", "closes", "the", "cache", "map", "and", "evict", "all", "cache", "node", "from", "cacher", "but", "unlike", "Close", "it", "doesn", "t", "forcefully", "releases", "cache", "node", "." ]
c3a204f8e96543bb0cc090385c001078f184fc46
https://github.com/syndtr/goleveldb/blob/c3a204f8e96543bb0cc090385c001078f184fc46/leveldb/cache/cache.go#L554-L569
train
syndtr/goleveldb
leveldb/cache/cache.go
GetHandle
func (n *Node) GetHandle() *Handle { if atomic.AddInt32(&n.ref, 1) <= 1 { panic("BUG: Node.GetHandle on zero ref") } return &Handle{unsafe.Pointer(n)} }
go
func (n *Node) GetHandle() *Handle { if atomic.AddInt32(&n.ref, 1) <= 1 { panic("BUG: Node.GetHandle on zero ref") } return &Handle{unsafe.Pointer(n)} }
[ "func", "(", "n", "*", "Node", ")", "GetHandle", "(", ")", "*", "Handle", "{", "if", "atomic", ".", "AddInt32", "(", "&", "n", ".", "ref", ",", "1", ")", "<=", "1", "{", "panic", "(", "\"", "\"", ")", "\n", "}", "\n", "return", "&", "Handle", "{", "unsafe", ".", "Pointer", "(", "n", ")", "}", "\n", "}" ]
// GetHandle returns an handle for this 'cache node'.
[ "GetHandle", "returns", "an", "handle", "for", "this", "cache", "node", "." ]
c3a204f8e96543bb0cc090385c001078f184fc46
https://github.com/syndtr/goleveldb/blob/c3a204f8e96543bb0cc090385c001078f184fc46/leveldb/cache/cache.go#L614-L619
train
syndtr/goleveldb
leveldb/cache/cache.go
Value
func (h *Handle) Value() Value { n := (*Node)(atomic.LoadPointer(&h.n)) if n != nil { return n.value } return nil }
go
func (h *Handle) Value() Value { n := (*Node)(atomic.LoadPointer(&h.n)) if n != nil { return n.value } return nil }
[ "func", "(", "h", "*", "Handle", ")", "Value", "(", ")", "Value", "{", "n", ":=", "(", "*", "Node", ")", "(", "atomic", ".", "LoadPointer", "(", "&", "h", ".", "n", ")", ")", "\n", "if", "n", "!=", "nil", "{", "return", "n", ".", "value", "\n", "}", "\n", "return", "nil", "\n", "}" ]
// Value returns the value of the 'cache node'.
[ "Value", "returns", "the", "value", "of", "the", "cache", "node", "." ]
c3a204f8e96543bb0cc090385c001078f184fc46
https://github.com/syndtr/goleveldb/blob/c3a204f8e96543bb0cc090385c001078f184fc46/leveldb/cache/cache.go#L643-L649
train
syndtr/goleveldb
leveldb/cache/cache.go
Release
func (h *Handle) Release() { nPtr := atomic.LoadPointer(&h.n) if nPtr != nil && atomic.CompareAndSwapPointer(&h.n, nPtr, nil) { n := (*Node)(nPtr) n.unrefLocked() } }
go
func (h *Handle) Release() { nPtr := atomic.LoadPointer(&h.n) if nPtr != nil && atomic.CompareAndSwapPointer(&h.n, nPtr, nil) { n := (*Node)(nPtr) n.unrefLocked() } }
[ "func", "(", "h", "*", "Handle", ")", "Release", "(", ")", "{", "nPtr", ":=", "atomic", ".", "LoadPointer", "(", "&", "h", ".", "n", ")", "\n", "if", "nPtr", "!=", "nil", "&&", "atomic", ".", "CompareAndSwapPointer", "(", "&", "h", ".", "n", ",", "nPtr", ",", "nil", ")", "{", "n", ":=", "(", "*", "Node", ")", "(", "nPtr", ")", "\n", "n", ".", "unrefLocked", "(", ")", "\n", "}", "\n", "}" ]
// Release releases this 'cache handle'. // It is safe to call release multiple times.
[ "Release", "releases", "this", "cache", "handle", ".", "It", "is", "safe", "to", "call", "release", "multiple", "times", "." ]
c3a204f8e96543bb0cc090385c001078f184fc46
https://github.com/syndtr/goleveldb/blob/c3a204f8e96543bb0cc090385c001078f184fc46/leveldb/cache/cache.go#L653-L659
train
syndtr/goleveldb
leveldb/util/hash.go
Hash
func Hash(data []byte, seed uint32) uint32 { // Similar to murmur hash const ( m = uint32(0xc6a4a793) r = uint32(24) ) var ( h = seed ^ (uint32(len(data)) * m) i int ) for n := len(data) - len(data)%4; i < n; i += 4 { h += binary.LittleEndian.Uint32(data[i:]) h *= m h ^= (h >> 16) } switch len(data) - i { default: panic("not reached") case 3: h += uint32(data[i+2]) << 16 fallthrough case 2: h += uint32(data[i+1]) << 8 fallthrough case 1: h += uint32(data[i]) h *= m h ^= (h >> r) case 0: } return h }
go
func Hash(data []byte, seed uint32) uint32 { // Similar to murmur hash const ( m = uint32(0xc6a4a793) r = uint32(24) ) var ( h = seed ^ (uint32(len(data)) * m) i int ) for n := len(data) - len(data)%4; i < n; i += 4 { h += binary.LittleEndian.Uint32(data[i:]) h *= m h ^= (h >> 16) } switch len(data) - i { default: panic("not reached") case 3: h += uint32(data[i+2]) << 16 fallthrough case 2: h += uint32(data[i+1]) << 8 fallthrough case 1: h += uint32(data[i]) h *= m h ^= (h >> r) case 0: } return h }
[ "func", "Hash", "(", "data", "[", "]", "byte", ",", "seed", "uint32", ")", "uint32", "{", "// Similar to murmur hash", "const", "(", "m", "=", "uint32", "(", "0xc6a4a793", ")", "\n", "r", "=", "uint32", "(", "24", ")", "\n", ")", "\n", "var", "(", "h", "=", "seed", "^", "(", "uint32", "(", "len", "(", "data", ")", ")", "*", "m", ")", "\n", "i", "int", "\n", ")", "\n\n", "for", "n", ":=", "len", "(", "data", ")", "-", "len", "(", "data", ")", "%", "4", ";", "i", "<", "n", ";", "i", "+=", "4", "{", "h", "+=", "binary", ".", "LittleEndian", ".", "Uint32", "(", "data", "[", "i", ":", "]", ")", "\n", "h", "*=", "m", "\n", "h", "^=", "(", "h", ">>", "16", ")", "\n", "}", "\n\n", "switch", "len", "(", "data", ")", "-", "i", "{", "default", ":", "panic", "(", "\"", "\"", ")", "\n", "case", "3", ":", "h", "+=", "uint32", "(", "data", "[", "i", "+", "2", "]", ")", "<<", "16", "\n", "fallthrough", "\n", "case", "2", ":", "h", "+=", "uint32", "(", "data", "[", "i", "+", "1", "]", ")", "<<", "8", "\n", "fallthrough", "\n", "case", "1", ":", "h", "+=", "uint32", "(", "data", "[", "i", "]", ")", "\n", "h", "*=", "m", "\n", "h", "^=", "(", "h", ">>", "r", ")", "\n", "case", "0", ":", "}", "\n\n", "return", "h", "\n", "}" ]
// Hash return hash of the given data.
[ "Hash", "return", "hash", "of", "the", "given", "data", "." ]
c3a204f8e96543bb0cc090385c001078f184fc46
https://github.com/syndtr/goleveldb/blob/c3a204f8e96543bb0cc090385c001078f184fc46/leveldb/util/hash.go#L14-L48
train
syndtr/goleveldb
leveldb/db_write.go
writeLocked
func (db *DB) writeLocked(batch, ourBatch *Batch, merge, sync bool) error { // Try to flush memdb. This method would also trying to throttle writes // if it is too fast and compaction cannot catch-up. mdb, mdbFree, err := db.flush(batch.internalLen) if err != nil { db.unlockWrite(false, 0, err) return err } defer mdb.decref() var ( overflow bool merged int batches = []*Batch{batch} ) if merge { // Merge limit. var mergeLimit int if batch.internalLen > 128<<10 { mergeLimit = (1 << 20) - batch.internalLen } else { mergeLimit = 128 << 10 } mergeCap := mdbFree - batch.internalLen if mergeLimit > mergeCap { mergeLimit = mergeCap } merge: for mergeLimit > 0 { select { case incoming := <-db.writeMergeC: if incoming.batch != nil { // Merge batch. if incoming.batch.internalLen > mergeLimit { overflow = true break merge } batches = append(batches, incoming.batch) mergeLimit -= incoming.batch.internalLen } else { // Merge put. internalLen := len(incoming.key) + len(incoming.value) + 8 if internalLen > mergeLimit { overflow = true break merge } if ourBatch == nil { ourBatch = db.batchPool.Get().(*Batch) ourBatch.Reset() batches = append(batches, ourBatch) } // We can use same batch since concurrent write doesn't // guarantee write order. ourBatch.appendRec(incoming.keyType, incoming.key, incoming.value) mergeLimit -= internalLen } sync = sync || incoming.sync merged++ db.writeMergedC <- true default: break merge } } } // Release ourBatch if any. if ourBatch != nil { defer db.batchPool.Put(ourBatch) } // Seq number. seq := db.seq + 1 // Write journal. if err := db.writeJournal(batches, seq, sync); err != nil { db.unlockWrite(overflow, merged, err) return err } // Put batches. for _, batch := range batches { if err := batch.putMem(seq, mdb.DB); err != nil { panic(err) } seq += uint64(batch.Len()) } // Incr seq number. db.addSeq(uint64(batchesLen(batches))) // Rotate memdb if it's reach the threshold. if batch.internalLen >= mdbFree { db.rotateMem(0, false) } db.unlockWrite(overflow, merged, nil) return nil }
go
func (db *DB) writeLocked(batch, ourBatch *Batch, merge, sync bool) error { // Try to flush memdb. This method would also trying to throttle writes // if it is too fast and compaction cannot catch-up. mdb, mdbFree, err := db.flush(batch.internalLen) if err != nil { db.unlockWrite(false, 0, err) return err } defer mdb.decref() var ( overflow bool merged int batches = []*Batch{batch} ) if merge { // Merge limit. var mergeLimit int if batch.internalLen > 128<<10 { mergeLimit = (1 << 20) - batch.internalLen } else { mergeLimit = 128 << 10 } mergeCap := mdbFree - batch.internalLen if mergeLimit > mergeCap { mergeLimit = mergeCap } merge: for mergeLimit > 0 { select { case incoming := <-db.writeMergeC: if incoming.batch != nil { // Merge batch. if incoming.batch.internalLen > mergeLimit { overflow = true break merge } batches = append(batches, incoming.batch) mergeLimit -= incoming.batch.internalLen } else { // Merge put. internalLen := len(incoming.key) + len(incoming.value) + 8 if internalLen > mergeLimit { overflow = true break merge } if ourBatch == nil { ourBatch = db.batchPool.Get().(*Batch) ourBatch.Reset() batches = append(batches, ourBatch) } // We can use same batch since concurrent write doesn't // guarantee write order. ourBatch.appendRec(incoming.keyType, incoming.key, incoming.value) mergeLimit -= internalLen } sync = sync || incoming.sync merged++ db.writeMergedC <- true default: break merge } } } // Release ourBatch if any. if ourBatch != nil { defer db.batchPool.Put(ourBatch) } // Seq number. seq := db.seq + 1 // Write journal. if err := db.writeJournal(batches, seq, sync); err != nil { db.unlockWrite(overflow, merged, err) return err } // Put batches. for _, batch := range batches { if err := batch.putMem(seq, mdb.DB); err != nil { panic(err) } seq += uint64(batch.Len()) } // Incr seq number. db.addSeq(uint64(batchesLen(batches))) // Rotate memdb if it's reach the threshold. if batch.internalLen >= mdbFree { db.rotateMem(0, false) } db.unlockWrite(overflow, merged, nil) return nil }
[ "func", "(", "db", "*", "DB", ")", "writeLocked", "(", "batch", ",", "ourBatch", "*", "Batch", ",", "merge", ",", "sync", "bool", ")", "error", "{", "// Try to flush memdb. This method would also trying to throttle writes", "// if it is too fast and compaction cannot catch-up.", "mdb", ",", "mdbFree", ",", "err", ":=", "db", ".", "flush", "(", "batch", ".", "internalLen", ")", "\n", "if", "err", "!=", "nil", "{", "db", ".", "unlockWrite", "(", "false", ",", "0", ",", "err", ")", "\n", "return", "err", "\n", "}", "\n", "defer", "mdb", ".", "decref", "(", ")", "\n\n", "var", "(", "overflow", "bool", "\n", "merged", "int", "\n", "batches", "=", "[", "]", "*", "Batch", "{", "batch", "}", "\n", ")", "\n\n", "if", "merge", "{", "// Merge limit.", "var", "mergeLimit", "int", "\n", "if", "batch", ".", "internalLen", ">", "128", "<<", "10", "{", "mergeLimit", "=", "(", "1", "<<", "20", ")", "-", "batch", ".", "internalLen", "\n", "}", "else", "{", "mergeLimit", "=", "128", "<<", "10", "\n", "}", "\n", "mergeCap", ":=", "mdbFree", "-", "batch", ".", "internalLen", "\n", "if", "mergeLimit", ">", "mergeCap", "{", "mergeLimit", "=", "mergeCap", "\n", "}", "\n\n", "merge", ":", "for", "mergeLimit", ">", "0", "{", "select", "{", "case", "incoming", ":=", "<-", "db", ".", "writeMergeC", ":", "if", "incoming", ".", "batch", "!=", "nil", "{", "// Merge batch.", "if", "incoming", ".", "batch", ".", "internalLen", ">", "mergeLimit", "{", "overflow", "=", "true", "\n", "break", "merge", "\n", "}", "\n", "batches", "=", "append", "(", "batches", ",", "incoming", ".", "batch", ")", "\n", "mergeLimit", "-=", "incoming", ".", "batch", ".", "internalLen", "\n", "}", "else", "{", "// Merge put.", "internalLen", ":=", "len", "(", "incoming", ".", "key", ")", "+", "len", "(", "incoming", ".", "value", ")", "+", "8", "\n", "if", "internalLen", ">", "mergeLimit", "{", "overflow", "=", "true", "\n", "break", "merge", "\n", "}", "\n", "if", "ourBatch", "==", "nil", "{", "ourBatch", "=", "db", ".", "batchPool", ".", "Get", "(", ")", ".", "(", "*", "Batch", ")", "\n", "ourBatch", ".", "Reset", "(", ")", "\n", "batches", "=", "append", "(", "batches", ",", "ourBatch", ")", "\n", "}", "\n", "// We can use same batch since concurrent write doesn't", "// guarantee write order.", "ourBatch", ".", "appendRec", "(", "incoming", ".", "keyType", ",", "incoming", ".", "key", ",", "incoming", ".", "value", ")", "\n", "mergeLimit", "-=", "internalLen", "\n", "}", "\n", "sync", "=", "sync", "||", "incoming", ".", "sync", "\n", "merged", "++", "\n", "db", ".", "writeMergedC", "<-", "true", "\n\n", "default", ":", "break", "merge", "\n", "}", "\n", "}", "\n", "}", "\n\n", "// Release ourBatch if any.", "if", "ourBatch", "!=", "nil", "{", "defer", "db", ".", "batchPool", ".", "Put", "(", "ourBatch", ")", "\n", "}", "\n\n", "// Seq number.", "seq", ":=", "db", ".", "seq", "+", "1", "\n\n", "// Write journal.", "if", "err", ":=", "db", ".", "writeJournal", "(", "batches", ",", "seq", ",", "sync", ")", ";", "err", "!=", "nil", "{", "db", ".", "unlockWrite", "(", "overflow", ",", "merged", ",", "err", ")", "\n", "return", "err", "\n", "}", "\n\n", "// Put batches.", "for", "_", ",", "batch", ":=", "range", "batches", "{", "if", "err", ":=", "batch", ".", "putMem", "(", "seq", ",", "mdb", ".", "DB", ")", ";", "err", "!=", "nil", "{", "panic", "(", "err", ")", "\n", "}", "\n", "seq", "+=", "uint64", "(", "batch", ".", "Len", "(", ")", ")", "\n", "}", "\n\n", "// Incr seq number.", "db", ".", "addSeq", "(", "uint64", "(", "batchesLen", "(", "batches", ")", ")", ")", "\n\n", "// Rotate memdb if it's reach the threshold.", "if", "batch", ".", "internalLen", ">=", "mdbFree", "{", "db", ".", "rotateMem", "(", "0", ",", "false", ")", "\n", "}", "\n\n", "db", ".", "unlockWrite", "(", "overflow", ",", "merged", ",", "nil", ")", "\n", "return", "nil", "\n", "}" ]
// ourBatch is batch that we can modify.
[ "ourBatch", "is", "batch", "that", "we", "can", "modify", "." ]
c3a204f8e96543bb0cc090385c001078f184fc46
https://github.com/syndtr/goleveldb/blob/c3a204f8e96543bb0cc090385c001078f184fc46/leveldb/db_write.go#L154-L254
train
syndtr/goleveldb
leveldb/db_write.go
Write
func (db *DB) Write(batch *Batch, wo *opt.WriteOptions) error { if err := db.ok(); err != nil || batch == nil || batch.Len() == 0 { return err } // If the batch size is larger than write buffer, it may justified to write // using transaction instead. Using transaction the batch will be written // into tables directly, skipping the journaling. if batch.internalLen > db.s.o.GetWriteBuffer() && !db.s.o.GetDisableLargeBatchTransaction() { tr, err := db.OpenTransaction() if err != nil { return err } if err := tr.Write(batch, wo); err != nil { tr.Discard() return err } return tr.Commit() } merge := !wo.GetNoWriteMerge() && !db.s.o.GetNoWriteMerge() sync := wo.GetSync() && !db.s.o.GetNoSync() // Acquire write lock. if merge { select { case db.writeMergeC <- writeMerge{sync: sync, batch: batch}: if <-db.writeMergedC { // Write is merged. return <-db.writeAckC } // Write is not merged, the write lock is handed to us. Continue. case db.writeLockC <- struct{}{}: // Write lock acquired. case err := <-db.compPerErrC: // Compaction error. return err case <-db.closeC: // Closed return ErrClosed } } else { select { case db.writeLockC <- struct{}{}: // Write lock acquired. case err := <-db.compPerErrC: // Compaction error. return err case <-db.closeC: // Closed return ErrClosed } } return db.writeLocked(batch, nil, merge, sync) }
go
func (db *DB) Write(batch *Batch, wo *opt.WriteOptions) error { if err := db.ok(); err != nil || batch == nil || batch.Len() == 0 { return err } // If the batch size is larger than write buffer, it may justified to write // using transaction instead. Using transaction the batch will be written // into tables directly, skipping the journaling. if batch.internalLen > db.s.o.GetWriteBuffer() && !db.s.o.GetDisableLargeBatchTransaction() { tr, err := db.OpenTransaction() if err != nil { return err } if err := tr.Write(batch, wo); err != nil { tr.Discard() return err } return tr.Commit() } merge := !wo.GetNoWriteMerge() && !db.s.o.GetNoWriteMerge() sync := wo.GetSync() && !db.s.o.GetNoSync() // Acquire write lock. if merge { select { case db.writeMergeC <- writeMerge{sync: sync, batch: batch}: if <-db.writeMergedC { // Write is merged. return <-db.writeAckC } // Write is not merged, the write lock is handed to us. Continue. case db.writeLockC <- struct{}{}: // Write lock acquired. case err := <-db.compPerErrC: // Compaction error. return err case <-db.closeC: // Closed return ErrClosed } } else { select { case db.writeLockC <- struct{}{}: // Write lock acquired. case err := <-db.compPerErrC: // Compaction error. return err case <-db.closeC: // Closed return ErrClosed } } return db.writeLocked(batch, nil, merge, sync) }
[ "func", "(", "db", "*", "DB", ")", "Write", "(", "batch", "*", "Batch", ",", "wo", "*", "opt", ".", "WriteOptions", ")", "error", "{", "if", "err", ":=", "db", ".", "ok", "(", ")", ";", "err", "!=", "nil", "||", "batch", "==", "nil", "||", "batch", ".", "Len", "(", ")", "==", "0", "{", "return", "err", "\n", "}", "\n\n", "// If the batch size is larger than write buffer, it may justified to write", "// using transaction instead. Using transaction the batch will be written", "// into tables directly, skipping the journaling.", "if", "batch", ".", "internalLen", ">", "db", ".", "s", ".", "o", ".", "GetWriteBuffer", "(", ")", "&&", "!", "db", ".", "s", ".", "o", ".", "GetDisableLargeBatchTransaction", "(", ")", "{", "tr", ",", "err", ":=", "db", ".", "OpenTransaction", "(", ")", "\n", "if", "err", "!=", "nil", "{", "return", "err", "\n", "}", "\n", "if", "err", ":=", "tr", ".", "Write", "(", "batch", ",", "wo", ")", ";", "err", "!=", "nil", "{", "tr", ".", "Discard", "(", ")", "\n", "return", "err", "\n", "}", "\n", "return", "tr", ".", "Commit", "(", ")", "\n", "}", "\n\n", "merge", ":=", "!", "wo", ".", "GetNoWriteMerge", "(", ")", "&&", "!", "db", ".", "s", ".", "o", ".", "GetNoWriteMerge", "(", ")", "\n", "sync", ":=", "wo", ".", "GetSync", "(", ")", "&&", "!", "db", ".", "s", ".", "o", ".", "GetNoSync", "(", ")", "\n\n", "// Acquire write lock.", "if", "merge", "{", "select", "{", "case", "db", ".", "writeMergeC", "<-", "writeMerge", "{", "sync", ":", "sync", ",", "batch", ":", "batch", "}", ":", "if", "<-", "db", ".", "writeMergedC", "{", "// Write is merged.", "return", "<-", "db", ".", "writeAckC", "\n", "}", "\n", "// Write is not merged, the write lock is handed to us. Continue.", "case", "db", ".", "writeLockC", "<-", "struct", "{", "}", "{", "}", ":", "// Write lock acquired.", "case", "err", ":=", "<-", "db", ".", "compPerErrC", ":", "// Compaction error.", "return", "err", "\n", "case", "<-", "db", ".", "closeC", ":", "// Closed", "return", "ErrClosed", "\n", "}", "\n", "}", "else", "{", "select", "{", "case", "db", ".", "writeLockC", "<-", "struct", "{", "}", "{", "}", ":", "// Write lock acquired.", "case", "err", ":=", "<-", "db", ".", "compPerErrC", ":", "// Compaction error.", "return", "err", "\n", "case", "<-", "db", ".", "closeC", ":", "// Closed", "return", "ErrClosed", "\n", "}", "\n", "}", "\n\n", "return", "db", ".", "writeLocked", "(", "batch", ",", "nil", ",", "merge", ",", "sync", ")", "\n", "}" ]
// Write apply the given batch to the DB. The batch records will be applied // sequentially. Write might be used concurrently, when used concurrently and // batch is small enough, write will try to merge the batches. Set NoWriteMerge // option to true to disable write merge. // // It is safe to modify the contents of the arguments after Write returns but // not before. Write will not modify content of the batch.
[ "Write", "apply", "the", "given", "batch", "to", "the", "DB", ".", "The", "batch", "records", "will", "be", "applied", "sequentially", ".", "Write", "might", "be", "used", "concurrently", "when", "used", "concurrently", "and", "batch", "is", "small", "enough", "write", "will", "try", "to", "merge", "the", "batches", ".", "Set", "NoWriteMerge", "option", "to", "true", "to", "disable", "write", "merge", ".", "It", "is", "safe", "to", "modify", "the", "contents", "of", "the", "arguments", "after", "Write", "returns", "but", "not", "before", ".", "Write", "will", "not", "modify", "content", "of", "the", "batch", "." ]
c3a204f8e96543bb0cc090385c001078f184fc46
https://github.com/syndtr/goleveldb/blob/c3a204f8e96543bb0cc090385c001078f184fc46/leveldb/db_write.go#L263-L318
train
syndtr/goleveldb
leveldb/db_write.go
Put
func (db *DB) Put(key, value []byte, wo *opt.WriteOptions) error { return db.putRec(keyTypeVal, key, value, wo) }
go
func (db *DB) Put(key, value []byte, wo *opt.WriteOptions) error { return db.putRec(keyTypeVal, key, value, wo) }
[ "func", "(", "db", "*", "DB", ")", "Put", "(", "key", ",", "value", "[", "]", "byte", ",", "wo", "*", "opt", ".", "WriteOptions", ")", "error", "{", "return", "db", ".", "putRec", "(", "keyTypeVal", ",", "key", ",", "value", ",", "wo", ")", "\n", "}" ]
// Put sets the value for the given key. It overwrites any previous value // for that key; a DB is not a multi-map. Write merge also applies for Put, see // Write. // // It is safe to modify the contents of the arguments after Put returns but not // before.
[ "Put", "sets", "the", "value", "for", "the", "given", "key", ".", "It", "overwrites", "any", "previous", "value", "for", "that", "key", ";", "a", "DB", "is", "not", "a", "multi", "-", "map", ".", "Write", "merge", "also", "applies", "for", "Put", "see", "Write", ".", "It", "is", "safe", "to", "modify", "the", "contents", "of", "the", "arguments", "after", "Put", "returns", "but", "not", "before", "." ]
c3a204f8e96543bb0cc090385c001078f184fc46
https://github.com/syndtr/goleveldb/blob/c3a204f8e96543bb0cc090385c001078f184fc46/leveldb/db_write.go#L371-L373
train
syndtr/goleveldb
leveldb/db_write.go
Delete
func (db *DB) Delete(key []byte, wo *opt.WriteOptions) error { return db.putRec(keyTypeDel, key, nil, wo) }
go
func (db *DB) Delete(key []byte, wo *opt.WriteOptions) error { return db.putRec(keyTypeDel, key, nil, wo) }
[ "func", "(", "db", "*", "DB", ")", "Delete", "(", "key", "[", "]", "byte", ",", "wo", "*", "opt", ".", "WriteOptions", ")", "error", "{", "return", "db", ".", "putRec", "(", "keyTypeDel", ",", "key", ",", "nil", ",", "wo", ")", "\n", "}" ]
// Delete deletes the value for the given key. Delete will not returns error if // key doesn't exist. Write merge also applies for Delete, see Write. // // It is safe to modify the contents of the arguments after Delete returns but // not before.
[ "Delete", "deletes", "the", "value", "for", "the", "given", "key", ".", "Delete", "will", "not", "returns", "error", "if", "key", "doesn", "t", "exist", ".", "Write", "merge", "also", "applies", "for", "Delete", "see", "Write", ".", "It", "is", "safe", "to", "modify", "the", "contents", "of", "the", "arguments", "after", "Delete", "returns", "but", "not", "before", "." ]
c3a204f8e96543bb0cc090385c001078f184fc46
https://github.com/syndtr/goleveldb/blob/c3a204f8e96543bb0cc090385c001078f184fc46/leveldb/db_write.go#L380-L382
train
syndtr/goleveldb
leveldb/db_write.go
CompactRange
func (db *DB) CompactRange(r util.Range) error { if err := db.ok(); err != nil { return err } // Lock writer. select { case db.writeLockC <- struct{}{}: case err := <-db.compPerErrC: return err case <-db.closeC: return ErrClosed } // Check for overlaps in memdb. mdb := db.getEffectiveMem() if mdb == nil { return ErrClosed } defer mdb.decref() if isMemOverlaps(db.s.icmp, mdb.DB, r.Start, r.Limit) { // Memdb compaction. if _, err := db.rotateMem(0, false); err != nil { <-db.writeLockC return err } <-db.writeLockC if err := db.compTriggerWait(db.mcompCmdC); err != nil { return err } } else { <-db.writeLockC } // Table compaction. return db.compTriggerRange(db.tcompCmdC, -1, r.Start, r.Limit) }
go
func (db *DB) CompactRange(r util.Range) error { if err := db.ok(); err != nil { return err } // Lock writer. select { case db.writeLockC <- struct{}{}: case err := <-db.compPerErrC: return err case <-db.closeC: return ErrClosed } // Check for overlaps in memdb. mdb := db.getEffectiveMem() if mdb == nil { return ErrClosed } defer mdb.decref() if isMemOverlaps(db.s.icmp, mdb.DB, r.Start, r.Limit) { // Memdb compaction. if _, err := db.rotateMem(0, false); err != nil { <-db.writeLockC return err } <-db.writeLockC if err := db.compTriggerWait(db.mcompCmdC); err != nil { return err } } else { <-db.writeLockC } // Table compaction. return db.compTriggerRange(db.tcompCmdC, -1, r.Start, r.Limit) }
[ "func", "(", "db", "*", "DB", ")", "CompactRange", "(", "r", "util", ".", "Range", ")", "error", "{", "if", "err", ":=", "db", ".", "ok", "(", ")", ";", "err", "!=", "nil", "{", "return", "err", "\n", "}", "\n\n", "// Lock writer.", "select", "{", "case", "db", ".", "writeLockC", "<-", "struct", "{", "}", "{", "}", ":", "case", "err", ":=", "<-", "db", ".", "compPerErrC", ":", "return", "err", "\n", "case", "<-", "db", ".", "closeC", ":", "return", "ErrClosed", "\n", "}", "\n\n", "// Check for overlaps in memdb.", "mdb", ":=", "db", ".", "getEffectiveMem", "(", ")", "\n", "if", "mdb", "==", "nil", "{", "return", "ErrClosed", "\n", "}", "\n", "defer", "mdb", ".", "decref", "(", ")", "\n", "if", "isMemOverlaps", "(", "db", ".", "s", ".", "icmp", ",", "mdb", ".", "DB", ",", "r", ".", "Start", ",", "r", ".", "Limit", ")", "{", "// Memdb compaction.", "if", "_", ",", "err", ":=", "db", ".", "rotateMem", "(", "0", ",", "false", ")", ";", "err", "!=", "nil", "{", "<-", "db", ".", "writeLockC", "\n", "return", "err", "\n", "}", "\n", "<-", "db", ".", "writeLockC", "\n", "if", "err", ":=", "db", ".", "compTriggerWait", "(", "db", ".", "mcompCmdC", ")", ";", "err", "!=", "nil", "{", "return", "err", "\n", "}", "\n", "}", "else", "{", "<-", "db", ".", "writeLockC", "\n", "}", "\n\n", "// Table compaction.", "return", "db", ".", "compTriggerRange", "(", "db", ".", "tcompCmdC", ",", "-", "1", ",", "r", ".", "Start", ",", "r", ".", "Limit", ")", "\n", "}" ]
// CompactRange compacts the underlying DB for the given key range. // In particular, deleted and overwritten versions are discarded, // and the data is rearranged to reduce the cost of operations // needed to access the data. This operation should typically only // be invoked by users who understand the underlying implementation. // // A nil Range.Start is treated as a key before all keys in the DB. // And a nil Range.Limit is treated as a key after all keys in the DB. // Therefore if both is nil then it will compact entire DB.
[ "CompactRange", "compacts", "the", "underlying", "DB", "for", "the", "given", "key", "range", ".", "In", "particular", "deleted", "and", "overwritten", "versions", "are", "discarded", "and", "the", "data", "is", "rearranged", "to", "reduce", "the", "cost", "of", "operations", "needed", "to", "access", "the", "data", ".", "This", "operation", "should", "typically", "only", "be", "invoked", "by", "users", "who", "understand", "the", "underlying", "implementation", ".", "A", "nil", "Range", ".", "Start", "is", "treated", "as", "a", "key", "before", "all", "keys", "in", "the", "DB", ".", "And", "a", "nil", "Range", ".", "Limit", "is", "treated", "as", "a", "key", "after", "all", "keys", "in", "the", "DB", ".", "Therefore", "if", "both", "is", "nil", "then", "it", "will", "compact", "entire", "DB", "." ]
c3a204f8e96543bb0cc090385c001078f184fc46
https://github.com/syndtr/goleveldb/blob/c3a204f8e96543bb0cc090385c001078f184fc46/leveldb/db_write.go#L400-L436
train
syndtr/goleveldb
leveldb/db_write.go
SetReadOnly
func (db *DB) SetReadOnly() error { if err := db.ok(); err != nil { return err } // Lock writer. select { case db.writeLockC <- struct{}{}: db.compWriteLocking = true case err := <-db.compPerErrC: return err case <-db.closeC: return ErrClosed } // Set compaction read-only. select { case db.compErrSetC <- ErrReadOnly: case perr := <-db.compPerErrC: return perr case <-db.closeC: return ErrClosed } return nil }
go
func (db *DB) SetReadOnly() error { if err := db.ok(); err != nil { return err } // Lock writer. select { case db.writeLockC <- struct{}{}: db.compWriteLocking = true case err := <-db.compPerErrC: return err case <-db.closeC: return ErrClosed } // Set compaction read-only. select { case db.compErrSetC <- ErrReadOnly: case perr := <-db.compPerErrC: return perr case <-db.closeC: return ErrClosed } return nil }
[ "func", "(", "db", "*", "DB", ")", "SetReadOnly", "(", ")", "error", "{", "if", "err", ":=", "db", ".", "ok", "(", ")", ";", "err", "!=", "nil", "{", "return", "err", "\n", "}", "\n\n", "// Lock writer.", "select", "{", "case", "db", ".", "writeLockC", "<-", "struct", "{", "}", "{", "}", ":", "db", ".", "compWriteLocking", "=", "true", "\n", "case", "err", ":=", "<-", "db", ".", "compPerErrC", ":", "return", "err", "\n", "case", "<-", "db", ".", "closeC", ":", "return", "ErrClosed", "\n", "}", "\n\n", "// Set compaction read-only.", "select", "{", "case", "db", ".", "compErrSetC", "<-", "ErrReadOnly", ":", "case", "perr", ":=", "<-", "db", ".", "compPerErrC", ":", "return", "perr", "\n", "case", "<-", "db", ".", "closeC", ":", "return", "ErrClosed", "\n", "}", "\n\n", "return", "nil", "\n", "}" ]
// SetReadOnly makes DB read-only. It will stay read-only until reopened.
[ "SetReadOnly", "makes", "DB", "read", "-", "only", ".", "It", "will", "stay", "read", "-", "only", "until", "reopened", "." ]
c3a204f8e96543bb0cc090385c001078f184fc46
https://github.com/syndtr/goleveldb/blob/c3a204f8e96543bb0cc090385c001078f184fc46/leveldb/db_write.go#L439-L464
train
syndtr/goleveldb
leveldb/util/util.go
Release
func (r *BasicReleaser) Release() { if !r.released { if r.releaser != nil { r.releaser.Release() r.releaser = nil } r.released = true } }
go
func (r *BasicReleaser) Release() { if !r.released { if r.releaser != nil { r.releaser.Release() r.releaser = nil } r.released = true } }
[ "func", "(", "r", "*", "BasicReleaser", ")", "Release", "(", ")", "{", "if", "!", "r", ".", "released", "{", "if", "r", ".", "releaser", "!=", "nil", "{", "r", ".", "releaser", ".", "Release", "(", ")", "\n", "r", ".", "releaser", "=", "nil", "\n", "}", "\n", "r", ".", "released", "=", "true", "\n", "}", "\n", "}" ]
// Release implements Releaser.Release.
[ "Release", "implements", "Releaser", ".", "Release", "." ]
c3a204f8e96543bb0cc090385c001078f184fc46
https://github.com/syndtr/goleveldb/blob/c3a204f8e96543bb0cc090385c001078f184fc46/leveldb/util/util.go#L50-L58
train
syndtr/goleveldb
leveldb/util/util.go
SetReleaser
func (r *BasicReleaser) SetReleaser(releaser Releaser) { if r.released { panic(ErrReleased) } if r.releaser != nil && releaser != nil { panic(ErrHasReleaser) } r.releaser = releaser }
go
func (r *BasicReleaser) SetReleaser(releaser Releaser) { if r.released { panic(ErrReleased) } if r.releaser != nil && releaser != nil { panic(ErrHasReleaser) } r.releaser = releaser }
[ "func", "(", "r", "*", "BasicReleaser", ")", "SetReleaser", "(", "releaser", "Releaser", ")", "{", "if", "r", ".", "released", "{", "panic", "(", "ErrReleased", ")", "\n", "}", "\n", "if", "r", ".", "releaser", "!=", "nil", "&&", "releaser", "!=", "nil", "{", "panic", "(", "ErrHasReleaser", ")", "\n", "}", "\n", "r", ".", "releaser", "=", "releaser", "\n", "}" ]
// SetReleaser implements ReleaseSetter.SetReleaser.
[ "SetReleaser", "implements", "ReleaseSetter", ".", "SetReleaser", "." ]
c3a204f8e96543bb0cc090385c001078f184fc46
https://github.com/syndtr/goleveldb/blob/c3a204f8e96543bb0cc090385c001078f184fc46/leveldb/util/util.go#L61-L69
train
syndtr/goleveldb
leveldb/errors/errors.go
NewErrCorrupted
func NewErrCorrupted(fd storage.FileDesc, err error) error { return &ErrCorrupted{fd, err} }
go
func NewErrCorrupted(fd storage.FileDesc, err error) error { return &ErrCorrupted{fd, err} }
[ "func", "NewErrCorrupted", "(", "fd", "storage", ".", "FileDesc", ",", "err", "error", ")", "error", "{", "return", "&", "ErrCorrupted", "{", "fd", ",", "err", "}", "\n", "}" ]
// NewErrCorrupted creates new ErrCorrupted error.
[ "NewErrCorrupted", "creates", "new", "ErrCorrupted", "error", "." ]
c3a204f8e96543bb0cc090385c001078f184fc46
https://github.com/syndtr/goleveldb/blob/c3a204f8e96543bb0cc090385c001078f184fc46/leveldb/errors/errors.go#L45-L47
train
syndtr/goleveldb
leveldb/errors/errors.go
IsCorrupted
func IsCorrupted(err error) bool { switch err.(type) { case *ErrCorrupted: return true case *storage.ErrCorrupted: return true } return false }
go
func IsCorrupted(err error) bool { switch err.(type) { case *ErrCorrupted: return true case *storage.ErrCorrupted: return true } return false }
[ "func", "IsCorrupted", "(", "err", "error", ")", "bool", "{", "switch", "err", ".", "(", "type", ")", "{", "case", "*", "ErrCorrupted", ":", "return", "true", "\n", "case", "*", "storage", ".", "ErrCorrupted", ":", "return", "true", "\n", "}", "\n", "return", "false", "\n", "}" ]
// IsCorrupted returns a boolean indicating whether the error is indicating // a corruption.
[ "IsCorrupted", "returns", "a", "boolean", "indicating", "whether", "the", "error", "is", "indicating", "a", "corruption", "." ]
c3a204f8e96543bb0cc090385c001078f184fc46
https://github.com/syndtr/goleveldb/blob/c3a204f8e96543bb0cc090385c001078f184fc46/leveldb/errors/errors.go#L51-L59
train
syndtr/goleveldb
leveldb/errors/errors.go
SetFd
func SetFd(err error, fd storage.FileDesc) error { switch x := err.(type) { case *ErrCorrupted: x.Fd = fd return x } return err }
go
func SetFd(err error, fd storage.FileDesc) error { switch x := err.(type) { case *ErrCorrupted: x.Fd = fd return x } return err }
[ "func", "SetFd", "(", "err", "error", ",", "fd", "storage", ".", "FileDesc", ")", "error", "{", "switch", "x", ":=", "err", ".", "(", "type", ")", "{", "case", "*", "ErrCorrupted", ":", "x", ".", "Fd", "=", "fd", "\n", "return", "x", "\n", "}", "\n", "return", "err", "\n", "}" ]
// SetFd sets 'file info' of the given error with the given file. // Currently only ErrCorrupted is supported, otherwise will do nothing.
[ "SetFd", "sets", "file", "info", "of", "the", "given", "error", "with", "the", "given", "file", ".", "Currently", "only", "ErrCorrupted", "is", "supported", "otherwise", "will", "do", "nothing", "." ]
c3a204f8e96543bb0cc090385c001078f184fc46
https://github.com/syndtr/goleveldb/blob/c3a204f8e96543bb0cc090385c001078f184fc46/leveldb/errors/errors.go#L71-L78
train
syndtr/goleveldb
leveldb/db_util.go
Sum
func (sizes Sizes) Sum() int64 { var sum int64 for _, size := range sizes { sum += size } return sum }
go
func (sizes Sizes) Sum() int64 { var sum int64 for _, size := range sizes { sum += size } return sum }
[ "func", "(", "sizes", "Sizes", ")", "Sum", "(", ")", "int64", "{", "var", "sum", "int64", "\n", "for", "_", ",", "size", ":=", "range", "sizes", "{", "sum", "+=", "size", "\n", "}", "\n", "return", "sum", "\n", "}" ]
// Sum returns sum of the sizes.
[ "Sum", "returns", "sum", "of", "the", "sizes", "." ]
c3a204f8e96543bb0cc090385c001078f184fc46
https://github.com/syndtr/goleveldb/blob/c3a204f8e96543bb0cc090385c001078f184fc46/leveldb/db_util.go#L28-L34
train
syndtr/goleveldb
leveldb/db_util.go
checkAndCleanFiles
func (db *DB) checkAndCleanFiles() error { v := db.s.version() defer v.release() tmap := make(map[int64]bool) for _, tables := range v.levels { for _, t := range tables { tmap[t.fd.Num] = false } } fds, err := db.s.stor.List(storage.TypeAll) if err != nil { return err } var nt int var rem []storage.FileDesc for _, fd := range fds { keep := true switch fd.Type { case storage.TypeManifest: keep = fd.Num >= db.s.manifestFd.Num case storage.TypeJournal: if !db.frozenJournalFd.Zero() { keep = fd.Num >= db.frozenJournalFd.Num } else { keep = fd.Num >= db.journalFd.Num } case storage.TypeTable: _, keep = tmap[fd.Num] if keep { tmap[fd.Num] = true nt++ } } if !keep { rem = append(rem, fd) } } if nt != len(tmap) { var mfds []storage.FileDesc for num, present := range tmap { if !present { mfds = append(mfds, storage.FileDesc{Type: storage.TypeTable, Num: num}) db.logf("db@janitor table missing @%d", num) } } return errors.NewErrCorrupted(storage.FileDesc{}, &errors.ErrMissingFiles{Fds: mfds}) } db.logf("db@janitor F·%d G·%d", len(fds), len(rem)) for _, fd := range rem { db.logf("db@janitor removing %s-%d", fd.Type, fd.Num) if err := db.s.stor.Remove(fd); err != nil { return err } } return nil }
go
func (db *DB) checkAndCleanFiles() error { v := db.s.version() defer v.release() tmap := make(map[int64]bool) for _, tables := range v.levels { for _, t := range tables { tmap[t.fd.Num] = false } } fds, err := db.s.stor.List(storage.TypeAll) if err != nil { return err } var nt int var rem []storage.FileDesc for _, fd := range fds { keep := true switch fd.Type { case storage.TypeManifest: keep = fd.Num >= db.s.manifestFd.Num case storage.TypeJournal: if !db.frozenJournalFd.Zero() { keep = fd.Num >= db.frozenJournalFd.Num } else { keep = fd.Num >= db.journalFd.Num } case storage.TypeTable: _, keep = tmap[fd.Num] if keep { tmap[fd.Num] = true nt++ } } if !keep { rem = append(rem, fd) } } if nt != len(tmap) { var mfds []storage.FileDesc for num, present := range tmap { if !present { mfds = append(mfds, storage.FileDesc{Type: storage.TypeTable, Num: num}) db.logf("db@janitor table missing @%d", num) } } return errors.NewErrCorrupted(storage.FileDesc{}, &errors.ErrMissingFiles{Fds: mfds}) } db.logf("db@janitor F·%d G·%d", len(fds), len(rem)) for _, fd := range rem { db.logf("db@janitor removing %s-%d", fd.Type, fd.Num) if err := db.s.stor.Remove(fd); err != nil { return err } } return nil }
[ "func", "(", "db", "*", "DB", ")", "checkAndCleanFiles", "(", ")", "error", "{", "v", ":=", "db", ".", "s", ".", "version", "(", ")", "\n", "defer", "v", ".", "release", "(", ")", "\n\n", "tmap", ":=", "make", "(", "map", "[", "int64", "]", "bool", ")", "\n", "for", "_", ",", "tables", ":=", "range", "v", ".", "levels", "{", "for", "_", ",", "t", ":=", "range", "tables", "{", "tmap", "[", "t", ".", "fd", ".", "Num", "]", "=", "false", "\n", "}", "\n", "}", "\n\n", "fds", ",", "err", ":=", "db", ".", "s", ".", "stor", ".", "List", "(", "storage", ".", "TypeAll", ")", "\n", "if", "err", "!=", "nil", "{", "return", "err", "\n", "}", "\n\n", "var", "nt", "int", "\n", "var", "rem", "[", "]", "storage", ".", "FileDesc", "\n", "for", "_", ",", "fd", ":=", "range", "fds", "{", "keep", ":=", "true", "\n", "switch", "fd", ".", "Type", "{", "case", "storage", ".", "TypeManifest", ":", "keep", "=", "fd", ".", "Num", ">=", "db", ".", "s", ".", "manifestFd", ".", "Num", "\n", "case", "storage", ".", "TypeJournal", ":", "if", "!", "db", ".", "frozenJournalFd", ".", "Zero", "(", ")", "{", "keep", "=", "fd", ".", "Num", ">=", "db", ".", "frozenJournalFd", ".", "Num", "\n", "}", "else", "{", "keep", "=", "fd", ".", "Num", ">=", "db", ".", "journalFd", ".", "Num", "\n", "}", "\n", "case", "storage", ".", "TypeTable", ":", "_", ",", "keep", "=", "tmap", "[", "fd", ".", "Num", "]", "\n", "if", "keep", "{", "tmap", "[", "fd", ".", "Num", "]", "=", "true", "\n", "nt", "++", "\n", "}", "\n", "}", "\n\n", "if", "!", "keep", "{", "rem", "=", "append", "(", "rem", ",", "fd", ")", "\n", "}", "\n", "}", "\n\n", "if", "nt", "!=", "len", "(", "tmap", ")", "{", "var", "mfds", "[", "]", "storage", ".", "FileDesc", "\n", "for", "num", ",", "present", ":=", "range", "tmap", "{", "if", "!", "present", "{", "mfds", "=", "append", "(", "mfds", ",", "storage", ".", "FileDesc", "{", "Type", ":", "storage", ".", "TypeTable", ",", "Num", ":", "num", "}", ")", "\n", "db", ".", "logf", "(", "\"", "\"", ",", "num", ")", "\n", "}", "\n", "}", "\n", "return", "errors", ".", "NewErrCorrupted", "(", "storage", ".", "FileDesc", "{", "}", ",", "&", "errors", ".", "ErrMissingFiles", "{", "Fds", ":", "mfds", "}", ")", "\n", "}", "\n\n", "db", ".", "logf", "(", "\"", " ", "l", "n(f", "d", "s),", " ", "l", "n(r", "e", "m))", "", "", "\n", "for", "_", ",", "fd", ":=", "range", "rem", "{", "db", ".", "logf", "(", "\"", "\"", ",", "fd", ".", "Type", ",", "fd", ".", "Num", ")", "\n", "if", "err", ":=", "db", ".", "s", ".", "stor", ".", "Remove", "(", "fd", ")", ";", "err", "!=", "nil", "{", "return", "err", "\n", "}", "\n", "}", "\n", "return", "nil", "\n", "}" ]
// Check and clean files.
[ "Check", "and", "clean", "files", "." ]
c3a204f8e96543bb0cc090385c001078f184fc46
https://github.com/syndtr/goleveldb/blob/c3a204f8e96543bb0cc090385c001078f184fc46/leveldb/db_util.go#L41-L102
train
syndtr/goleveldb
leveldb/batch.go
Delete
func (b *Batch) Delete(key []byte) { b.appendRec(keyTypeDel, key, nil) }
go
func (b *Batch) Delete(key []byte) { b.appendRec(keyTypeDel, key, nil) }
[ "func", "(", "b", "*", "Batch", ")", "Delete", "(", "key", "[", "]", "byte", ")", "{", "b", ".", "appendRec", "(", "keyTypeDel", ",", "key", ",", "nil", ")", "\n", "}" ]
// Delete appends 'delete operation' of the given key to the batch. // It is safe to modify the contents of the argument after Delete returns but // not before.
[ "Delete", "appends", "delete", "operation", "of", "the", "given", "key", "to", "the", "batch", ".", "It", "is", "safe", "to", "modify", "the", "contents", "of", "the", "argument", "after", "Delete", "returns", "but", "not", "before", "." ]
c3a204f8e96543bb0cc090385c001078f184fc46
https://github.com/syndtr/goleveldb/blob/c3a204f8e96543bb0cc090385c001078f184fc46/leveldb/batch.go#L124-L126
train
syndtr/goleveldb
leveldb/batch.go
Replay
func (b *Batch) Replay(r BatchReplay) error { for _, index := range b.index { switch index.keyType { case keyTypeVal: r.Put(index.k(b.data), index.v(b.data)) case keyTypeDel: r.Delete(index.k(b.data)) } } return nil }
go
func (b *Batch) Replay(r BatchReplay) error { for _, index := range b.index { switch index.keyType { case keyTypeVal: r.Put(index.k(b.data), index.v(b.data)) case keyTypeDel: r.Delete(index.k(b.data)) } } return nil }
[ "func", "(", "b", "*", "Batch", ")", "Replay", "(", "r", "BatchReplay", ")", "error", "{", "for", "_", ",", "index", ":=", "range", "b", ".", "index", "{", "switch", "index", ".", "keyType", "{", "case", "keyTypeVal", ":", "r", ".", "Put", "(", "index", ".", "k", "(", "b", ".", "data", ")", ",", "index", ".", "v", "(", "b", ".", "data", ")", ")", "\n", "case", "keyTypeDel", ":", "r", ".", "Delete", "(", "index", ".", "k", "(", "b", ".", "data", ")", ")", "\n", "}", "\n", "}", "\n", "return", "nil", "\n", "}" ]
// Replay replays batch contents.
[ "Replay", "replays", "batch", "contents", "." ]
c3a204f8e96543bb0cc090385c001078f184fc46
https://github.com/syndtr/goleveldb/blob/c3a204f8e96543bb0cc090385c001078f184fc46/leveldb/batch.go#L145-L155
train
syndtr/goleveldb
leveldb/batch.go
Reset
func (b *Batch) Reset() { b.data = b.data[:0] b.index = b.index[:0] b.internalLen = 0 }
go
func (b *Batch) Reset() { b.data = b.data[:0] b.index = b.index[:0] b.internalLen = 0 }
[ "func", "(", "b", "*", "Batch", ")", "Reset", "(", ")", "{", "b", ".", "data", "=", "b", ".", "data", "[", ":", "0", "]", "\n", "b", ".", "index", "=", "b", ".", "index", "[", ":", "0", "]", "\n", "b", ".", "internalLen", "=", "0", "\n", "}" ]
// Reset resets the batch.
[ "Reset", "resets", "the", "batch", "." ]
c3a204f8e96543bb0cc090385c001078f184fc46
https://github.com/syndtr/goleveldb/blob/c3a204f8e96543bb0cc090385c001078f184fc46/leveldb/batch.go#L163-L167
train
syndtr/goleveldb
leveldb/db.go
Open
func Open(stor storage.Storage, o *opt.Options) (db *DB, err error) { s, err := newSession(stor, o) if err != nil { return } defer func() { if err != nil { s.close() s.release() } }() err = s.recover() if err != nil { if !os.IsNotExist(err) || s.o.GetErrorIfMissing() || s.o.GetReadOnly() { return } err = s.create() if err != nil { return } } else if s.o.GetErrorIfExist() { err = os.ErrExist return } return openDB(s) }
go
func Open(stor storage.Storage, o *opt.Options) (db *DB, err error) { s, err := newSession(stor, o) if err != nil { return } defer func() { if err != nil { s.close() s.release() } }() err = s.recover() if err != nil { if !os.IsNotExist(err) || s.o.GetErrorIfMissing() || s.o.GetReadOnly() { return } err = s.create() if err != nil { return } } else if s.o.GetErrorIfExist() { err = os.ErrExist return } return openDB(s) }
[ "func", "Open", "(", "stor", "storage", ".", "Storage", ",", "o", "*", "opt", ".", "Options", ")", "(", "db", "*", "DB", ",", "err", "error", ")", "{", "s", ",", "err", ":=", "newSession", "(", "stor", ",", "o", ")", "\n", "if", "err", "!=", "nil", "{", "return", "\n", "}", "\n", "defer", "func", "(", ")", "{", "if", "err", "!=", "nil", "{", "s", ".", "close", "(", ")", "\n", "s", ".", "release", "(", ")", "\n", "}", "\n", "}", "(", ")", "\n\n", "err", "=", "s", ".", "recover", "(", ")", "\n", "if", "err", "!=", "nil", "{", "if", "!", "os", ".", "IsNotExist", "(", "err", ")", "||", "s", ".", "o", ".", "GetErrorIfMissing", "(", ")", "||", "s", ".", "o", ".", "GetReadOnly", "(", ")", "{", "return", "\n", "}", "\n", "err", "=", "s", ".", "create", "(", ")", "\n", "if", "err", "!=", "nil", "{", "return", "\n", "}", "\n", "}", "else", "if", "s", ".", "o", ".", "GetErrorIfExist", "(", ")", "{", "err", "=", "os", ".", "ErrExist", "\n", "return", "\n", "}", "\n\n", "return", "openDB", "(", "s", ")", "\n", "}" ]
// Open opens or creates a DB for the given storage. // The DB will be created if not exist, unless ErrorIfMissing is true. // Also, if ErrorIfExist is true and the DB exist Open will returns // os.ErrExist error. // // Open will return an error with type of ErrCorrupted if corruption // detected in the DB. Use errors.IsCorrupted to test whether an error is // due to corruption. Corrupted DB can be recovered with Recover function. // // The returned DB instance is safe for concurrent use. // The DB must be closed after use, by calling Close method.
[ "Open", "opens", "or", "creates", "a", "DB", "for", "the", "given", "storage", ".", "The", "DB", "will", "be", "created", "if", "not", "exist", "unless", "ErrorIfMissing", "is", "true", ".", "Also", "if", "ErrorIfExist", "is", "true", "and", "the", "DB", "exist", "Open", "will", "returns", "os", ".", "ErrExist", "error", ".", "Open", "will", "return", "an", "error", "with", "type", "of", "ErrCorrupted", "if", "corruption", "detected", "in", "the", "DB", ".", "Use", "errors", ".", "IsCorrupted", "to", "test", "whether", "an", "error", "is", "due", "to", "corruption", ".", "Corrupted", "DB", "can", "be", "recovered", "with", "Recover", "function", ".", "The", "returned", "DB", "instance", "is", "safe", "for", "concurrent", "use", ".", "The", "DB", "must", "be", "closed", "after", "use", "by", "calling", "Close", "method", "." ]
c3a204f8e96543bb0cc090385c001078f184fc46
https://github.com/syndtr/goleveldb/blob/c3a204f8e96543bb0cc090385c001078f184fc46/leveldb/db.go#L171-L198
train
syndtr/goleveldb
leveldb/db.go
Recover
func Recover(stor storage.Storage, o *opt.Options) (db *DB, err error) { s, err := newSession(stor, o) if err != nil { return } defer func() { if err != nil { s.close() s.release() } }() err = recoverTable(s, o) if err != nil { return } return openDB(s) }
go
func Recover(stor storage.Storage, o *opt.Options) (db *DB, err error) { s, err := newSession(stor, o) if err != nil { return } defer func() { if err != nil { s.close() s.release() } }() err = recoverTable(s, o) if err != nil { return } return openDB(s) }
[ "func", "Recover", "(", "stor", "storage", ".", "Storage", ",", "o", "*", "opt", ".", "Options", ")", "(", "db", "*", "DB", ",", "err", "error", ")", "{", "s", ",", "err", ":=", "newSession", "(", "stor", ",", "o", ")", "\n", "if", "err", "!=", "nil", "{", "return", "\n", "}", "\n", "defer", "func", "(", ")", "{", "if", "err", "!=", "nil", "{", "s", ".", "close", "(", ")", "\n", "s", ".", "release", "(", ")", "\n", "}", "\n", "}", "(", ")", "\n\n", "err", "=", "recoverTable", "(", "s", ",", "o", ")", "\n", "if", "err", "!=", "nil", "{", "return", "\n", "}", "\n", "return", "openDB", "(", "s", ")", "\n", "}" ]
// Recover recovers and opens a DB with missing or corrupted manifest files // for the given storage. It will ignore any manifest files, valid or not. // The DB must already exist or it will returns an error. // Also, Recover will ignore ErrorIfMissing and ErrorIfExist options. // // The returned DB instance is safe for concurrent use. // The DB must be closed after use, by calling Close method.
[ "Recover", "recovers", "and", "opens", "a", "DB", "with", "missing", "or", "corrupted", "manifest", "files", "for", "the", "given", "storage", ".", "It", "will", "ignore", "any", "manifest", "files", "valid", "or", "not", ".", "The", "DB", "must", "already", "exist", "or", "it", "will", "returns", "an", "error", ".", "Also", "Recover", "will", "ignore", "ErrorIfMissing", "and", "ErrorIfExist", "options", ".", "The", "returned", "DB", "instance", "is", "safe", "for", "concurrent", "use", ".", "The", "DB", "must", "be", "closed", "after", "use", "by", "calling", "Close", "method", "." ]
c3a204f8e96543bb0cc090385c001078f184fc46
https://github.com/syndtr/goleveldb/blob/c3a204f8e96543bb0cc090385c001078f184fc46/leveldb/db.go#L235-L252
train
syndtr/goleveldb
leveldb/db.go
GetSnapshot
func (db *DB) GetSnapshot() (*Snapshot, error) { if err := db.ok(); err != nil { return nil, err } return db.newSnapshot(), nil }
go
func (db *DB) GetSnapshot() (*Snapshot, error) { if err := db.ok(); err != nil { return nil, err } return db.newSnapshot(), nil }
[ "func", "(", "db", "*", "DB", ")", "GetSnapshot", "(", ")", "(", "*", "Snapshot", ",", "error", ")", "{", "if", "err", ":=", "db", ".", "ok", "(", ")", ";", "err", "!=", "nil", "{", "return", "nil", ",", "err", "\n", "}", "\n\n", "return", "db", ".", "newSnapshot", "(", ")", ",", "nil", "\n", "}" ]
// GetSnapshot returns a latest snapshot of the underlying DB. A snapshot // is a frozen snapshot of a DB state at a particular point in time. The // content of snapshot are guaranteed to be consistent. // // The snapshot must be released after use, by calling Release method.
[ "GetSnapshot", "returns", "a", "latest", "snapshot", "of", "the", "underlying", "DB", ".", "A", "snapshot", "is", "a", "frozen", "snapshot", "of", "a", "DB", "state", "at", "a", "particular", "point", "in", "time", ".", "The", "content", "of", "snapshot", "are", "guaranteed", "to", "be", "consistent", ".", "The", "snapshot", "must", "be", "released", "after", "use", "by", "calling", "Release", "method", "." ]
c3a204f8e96543bb0cc090385c001078f184fc46
https://github.com/syndtr/goleveldb/blob/c3a204f8e96543bb0cc090385c001078f184fc46/leveldb/db.go#L899-L905
train
syndtr/goleveldb
leveldb/db.go
Stats
func (db *DB) Stats(s *DBStats) error { err := db.ok() if err != nil { return err } s.IORead = db.s.stor.reads() s.IOWrite = db.s.stor.writes() s.WriteDelayCount = atomic.LoadInt32(&db.cWriteDelayN) s.WriteDelayDuration = time.Duration(atomic.LoadInt64(&db.cWriteDelay)) s.WritePaused = atomic.LoadInt32(&db.inWritePaused) == 1 s.OpenedTablesCount = db.s.tops.cache.Size() if db.s.tops.bcache != nil { s.BlockCacheSize = db.s.tops.bcache.Size() } else { s.BlockCacheSize = 0 } s.AliveIterators = atomic.LoadInt32(&db.aliveIters) s.AliveSnapshots = atomic.LoadInt32(&db.aliveSnaps) s.LevelDurations = s.LevelDurations[:0] s.LevelRead = s.LevelRead[:0] s.LevelWrite = s.LevelWrite[:0] s.LevelSizes = s.LevelSizes[:0] s.LevelTablesCounts = s.LevelTablesCounts[:0] v := db.s.version() defer v.release() for level, tables := range v.levels { duration, read, write := db.compStats.getStat(level) if len(tables) == 0 && duration == 0 { continue } s.LevelDurations = append(s.LevelDurations, duration) s.LevelRead = append(s.LevelRead, read) s.LevelWrite = append(s.LevelWrite, write) s.LevelSizes = append(s.LevelSizes, tables.size()) s.LevelTablesCounts = append(s.LevelTablesCounts, len(tables)) } return nil }
go
func (db *DB) Stats(s *DBStats) error { err := db.ok() if err != nil { return err } s.IORead = db.s.stor.reads() s.IOWrite = db.s.stor.writes() s.WriteDelayCount = atomic.LoadInt32(&db.cWriteDelayN) s.WriteDelayDuration = time.Duration(atomic.LoadInt64(&db.cWriteDelay)) s.WritePaused = atomic.LoadInt32(&db.inWritePaused) == 1 s.OpenedTablesCount = db.s.tops.cache.Size() if db.s.tops.bcache != nil { s.BlockCacheSize = db.s.tops.bcache.Size() } else { s.BlockCacheSize = 0 } s.AliveIterators = atomic.LoadInt32(&db.aliveIters) s.AliveSnapshots = atomic.LoadInt32(&db.aliveSnaps) s.LevelDurations = s.LevelDurations[:0] s.LevelRead = s.LevelRead[:0] s.LevelWrite = s.LevelWrite[:0] s.LevelSizes = s.LevelSizes[:0] s.LevelTablesCounts = s.LevelTablesCounts[:0] v := db.s.version() defer v.release() for level, tables := range v.levels { duration, read, write := db.compStats.getStat(level) if len(tables) == 0 && duration == 0 { continue } s.LevelDurations = append(s.LevelDurations, duration) s.LevelRead = append(s.LevelRead, read) s.LevelWrite = append(s.LevelWrite, write) s.LevelSizes = append(s.LevelSizes, tables.size()) s.LevelTablesCounts = append(s.LevelTablesCounts, len(tables)) } return nil }
[ "func", "(", "db", "*", "DB", ")", "Stats", "(", "s", "*", "DBStats", ")", "error", "{", "err", ":=", "db", ".", "ok", "(", ")", "\n", "if", "err", "!=", "nil", "{", "return", "err", "\n", "}", "\n\n", "s", ".", "IORead", "=", "db", ".", "s", ".", "stor", ".", "reads", "(", ")", "\n", "s", ".", "IOWrite", "=", "db", ".", "s", ".", "stor", ".", "writes", "(", ")", "\n", "s", ".", "WriteDelayCount", "=", "atomic", ".", "LoadInt32", "(", "&", "db", ".", "cWriteDelayN", ")", "\n", "s", ".", "WriteDelayDuration", "=", "time", ".", "Duration", "(", "atomic", ".", "LoadInt64", "(", "&", "db", ".", "cWriteDelay", ")", ")", "\n", "s", ".", "WritePaused", "=", "atomic", ".", "LoadInt32", "(", "&", "db", ".", "inWritePaused", ")", "==", "1", "\n\n", "s", ".", "OpenedTablesCount", "=", "db", ".", "s", ".", "tops", ".", "cache", ".", "Size", "(", ")", "\n", "if", "db", ".", "s", ".", "tops", ".", "bcache", "!=", "nil", "{", "s", ".", "BlockCacheSize", "=", "db", ".", "s", ".", "tops", ".", "bcache", ".", "Size", "(", ")", "\n", "}", "else", "{", "s", ".", "BlockCacheSize", "=", "0", "\n", "}", "\n\n", "s", ".", "AliveIterators", "=", "atomic", ".", "LoadInt32", "(", "&", "db", ".", "aliveIters", ")", "\n", "s", ".", "AliveSnapshots", "=", "atomic", ".", "LoadInt32", "(", "&", "db", ".", "aliveSnaps", ")", "\n\n", "s", ".", "LevelDurations", "=", "s", ".", "LevelDurations", "[", ":", "0", "]", "\n", "s", ".", "LevelRead", "=", "s", ".", "LevelRead", "[", ":", "0", "]", "\n", "s", ".", "LevelWrite", "=", "s", ".", "LevelWrite", "[", ":", "0", "]", "\n", "s", ".", "LevelSizes", "=", "s", ".", "LevelSizes", "[", ":", "0", "]", "\n", "s", ".", "LevelTablesCounts", "=", "s", ".", "LevelTablesCounts", "[", ":", "0", "]", "\n\n", "v", ":=", "db", ".", "s", ".", "version", "(", ")", "\n", "defer", "v", ".", "release", "(", ")", "\n\n", "for", "level", ",", "tables", ":=", "range", "v", ".", "levels", "{", "duration", ",", "read", ",", "write", ":=", "db", ".", "compStats", ".", "getStat", "(", "level", ")", "\n", "if", "len", "(", "tables", ")", "==", "0", "&&", "duration", "==", "0", "{", "continue", "\n", "}", "\n", "s", ".", "LevelDurations", "=", "append", "(", "s", ".", "LevelDurations", ",", "duration", ")", "\n", "s", ".", "LevelRead", "=", "append", "(", "s", ".", "LevelRead", ",", "read", ")", "\n", "s", ".", "LevelWrite", "=", "append", "(", "s", ".", "LevelWrite", ",", "write", ")", "\n", "s", ".", "LevelSizes", "=", "append", "(", "s", ".", "LevelSizes", ",", "tables", ".", "size", "(", ")", ")", "\n", "s", ".", "LevelTablesCounts", "=", "append", "(", "s", ".", "LevelTablesCounts", ",", "len", "(", "tables", ")", ")", "\n", "}", "\n\n", "return", "nil", "\n", "}" ]
// Stats populates s with database statistics.
[ "Stats", "populates", "s", "with", "database", "statistics", "." ]
c3a204f8e96543bb0cc090385c001078f184fc46
https://github.com/syndtr/goleveldb/blob/c3a204f8e96543bb0cc090385c001078f184fc46/leveldb/db.go#L1040-L1084
train
syndtr/goleveldb
leveldb/db.go
SizeOf
func (db *DB) SizeOf(ranges []util.Range) (Sizes, error) { if err := db.ok(); err != nil { return nil, err } v := db.s.version() defer v.release() sizes := make(Sizes, 0, len(ranges)) for _, r := range ranges { imin := makeInternalKey(nil, r.Start, keyMaxSeq, keyTypeSeek) imax := makeInternalKey(nil, r.Limit, keyMaxSeq, keyTypeSeek) start, err := v.offsetOf(imin) if err != nil { return nil, err } limit, err := v.offsetOf(imax) if err != nil { return nil, err } var size int64 if limit >= start { size = limit - start } sizes = append(sizes, size) } return sizes, nil }
go
func (db *DB) SizeOf(ranges []util.Range) (Sizes, error) { if err := db.ok(); err != nil { return nil, err } v := db.s.version() defer v.release() sizes := make(Sizes, 0, len(ranges)) for _, r := range ranges { imin := makeInternalKey(nil, r.Start, keyMaxSeq, keyTypeSeek) imax := makeInternalKey(nil, r.Limit, keyMaxSeq, keyTypeSeek) start, err := v.offsetOf(imin) if err != nil { return nil, err } limit, err := v.offsetOf(imax) if err != nil { return nil, err } var size int64 if limit >= start { size = limit - start } sizes = append(sizes, size) } return sizes, nil }
[ "func", "(", "db", "*", "DB", ")", "SizeOf", "(", "ranges", "[", "]", "util", ".", "Range", ")", "(", "Sizes", ",", "error", ")", "{", "if", "err", ":=", "db", ".", "ok", "(", ")", ";", "err", "!=", "nil", "{", "return", "nil", ",", "err", "\n", "}", "\n\n", "v", ":=", "db", ".", "s", ".", "version", "(", ")", "\n", "defer", "v", ".", "release", "(", ")", "\n\n", "sizes", ":=", "make", "(", "Sizes", ",", "0", ",", "len", "(", "ranges", ")", ")", "\n", "for", "_", ",", "r", ":=", "range", "ranges", "{", "imin", ":=", "makeInternalKey", "(", "nil", ",", "r", ".", "Start", ",", "keyMaxSeq", ",", "keyTypeSeek", ")", "\n", "imax", ":=", "makeInternalKey", "(", "nil", ",", "r", ".", "Limit", ",", "keyMaxSeq", ",", "keyTypeSeek", ")", "\n", "start", ",", "err", ":=", "v", ".", "offsetOf", "(", "imin", ")", "\n", "if", "err", "!=", "nil", "{", "return", "nil", ",", "err", "\n", "}", "\n", "limit", ",", "err", ":=", "v", ".", "offsetOf", "(", "imax", ")", "\n", "if", "err", "!=", "nil", "{", "return", "nil", ",", "err", "\n", "}", "\n", "var", "size", "int64", "\n", "if", "limit", ">=", "start", "{", "size", "=", "limit", "-", "start", "\n", "}", "\n", "sizes", "=", "append", "(", "sizes", ",", "size", ")", "\n", "}", "\n\n", "return", "sizes", ",", "nil", "\n", "}" ]
// SizeOf calculates approximate sizes of the given key ranges. // The length of the returned sizes are equal with the length of the given // ranges. The returned sizes measure storage space usage, so if the user // data compresses by a factor of ten, the returned sizes will be one-tenth // the size of the corresponding user data size. // The results may not include the sizes of recently written data.
[ "SizeOf", "calculates", "approximate", "sizes", "of", "the", "given", "key", "ranges", ".", "The", "length", "of", "the", "returned", "sizes", "are", "equal", "with", "the", "length", "of", "the", "given", "ranges", ".", "The", "returned", "sizes", "measure", "storage", "space", "usage", "so", "if", "the", "user", "data", "compresses", "by", "a", "factor", "of", "ten", "the", "returned", "sizes", "will", "be", "one", "-", "tenth", "the", "size", "of", "the", "corresponding", "user", "data", "size", ".", "The", "results", "may", "not", "include", "the", "sizes", "of", "recently", "written", "data", "." ]
c3a204f8e96543bb0cc090385c001078f184fc46
https://github.com/syndtr/goleveldb/blob/c3a204f8e96543bb0cc090385c001078f184fc46/leveldb/db.go#L1092-L1120
train
syndtr/goleveldb
leveldb/db.go
Close
func (db *DB) Close() error { if !db.setClosed() { return ErrClosed } start := time.Now() db.log("db@close closing") // Clear the finalizer. runtime.SetFinalizer(db, nil) // Get compaction error. var err error select { case err = <-db.compErrC: if err == ErrReadOnly { err = nil } default: } // Signal all goroutines. close(db.closeC) // Discard open transaction. if db.tr != nil { db.tr.Discard() } // Acquire writer lock. db.writeLockC <- struct{}{} // Wait for all gorotines to exit. db.closeW.Wait() // Closes journal. if db.journal != nil { db.journal.Close() db.journalWriter.Close() db.journal = nil db.journalWriter = nil } if db.writeDelayN > 0 { db.logf("db@write was delayed N·%d T·%v", db.writeDelayN, db.writeDelay) } // Close session. db.s.close() db.logf("db@close done T·%v", time.Since(start)) db.s.release() if db.closer != nil { if err1 := db.closer.Close(); err == nil { err = err1 } db.closer = nil } // Clear memdbs. db.clearMems() return err }
go
func (db *DB) Close() error { if !db.setClosed() { return ErrClosed } start := time.Now() db.log("db@close closing") // Clear the finalizer. runtime.SetFinalizer(db, nil) // Get compaction error. var err error select { case err = <-db.compErrC: if err == ErrReadOnly { err = nil } default: } // Signal all goroutines. close(db.closeC) // Discard open transaction. if db.tr != nil { db.tr.Discard() } // Acquire writer lock. db.writeLockC <- struct{}{} // Wait for all gorotines to exit. db.closeW.Wait() // Closes journal. if db.journal != nil { db.journal.Close() db.journalWriter.Close() db.journal = nil db.journalWriter = nil } if db.writeDelayN > 0 { db.logf("db@write was delayed N·%d T·%v", db.writeDelayN, db.writeDelay) } // Close session. db.s.close() db.logf("db@close done T·%v", time.Since(start)) db.s.release() if db.closer != nil { if err1 := db.closer.Close(); err == nil { err = err1 } db.closer = nil } // Clear memdbs. db.clearMems() return err }
[ "func", "(", "db", "*", "DB", ")", "Close", "(", ")", "error", "{", "if", "!", "db", ".", "setClosed", "(", ")", "{", "return", "ErrClosed", "\n", "}", "\n\n", "start", ":=", "time", ".", "Now", "(", ")", "\n", "db", ".", "log", "(", "\"", "\"", ")", "\n\n", "// Clear the finalizer.", "runtime", ".", "SetFinalizer", "(", "db", ",", "nil", ")", "\n\n", "// Get compaction error.", "var", "err", "error", "\n", "select", "{", "case", "err", "=", "<-", "db", ".", "compErrC", ":", "if", "err", "==", "ErrReadOnly", "{", "err", "=", "nil", "\n", "}", "\n", "default", ":", "}", "\n\n", "// Signal all goroutines.", "close", "(", "db", ".", "closeC", ")", "\n\n", "// Discard open transaction.", "if", "db", ".", "tr", "!=", "nil", "{", "db", ".", "tr", ".", "Discard", "(", ")", "\n", "}", "\n\n", "// Acquire writer lock.", "db", ".", "writeLockC", "<-", "struct", "{", "}", "{", "}", "\n\n", "// Wait for all gorotines to exit.", "db", ".", "closeW", ".", "Wait", "(", ")", "\n\n", "// Closes journal.", "if", "db", ".", "journal", "!=", "nil", "{", "db", ".", "journal", ".", "Close", "(", ")", "\n", "db", ".", "journalWriter", ".", "Close", "(", ")", "\n", "db", ".", "journal", "=", "nil", "\n", "db", ".", "journalWriter", "=", "nil", "\n", "}", "\n\n", "if", "db", ".", "writeDelayN", ">", "0", "{", "db", ".", "logf", "(", "\"", " ", "d", ".w", "r", "iteDelayN, ", "d", ".w", "r", "iteDelay)", "", "\n", "}", "\n\n", "// Close session.", "db", ".", "s", ".", "close", "(", ")", "\n", "db", ".", "logf", "(", "\"", ",", " ", "ime.", "S", "ince(", "s", "tart)", ")", "", "\n", "db", ".", "s", ".", "release", "(", ")", "\n\n", "if", "db", ".", "closer", "!=", "nil", "{", "if", "err1", ":=", "db", ".", "closer", ".", "Close", "(", ")", ";", "err", "==", "nil", "{", "err", "=", "err1", "\n", "}", "\n", "db", ".", "closer", "=", "nil", "\n", "}", "\n\n", "// Clear memdbs.", "db", ".", "clearMems", "(", ")", "\n\n", "return", "err", "\n", "}" ]
// Close closes the DB. This will also releases any outstanding snapshot, // abort any in-flight compaction and discard open transaction. // // It is not safe to close a DB until all outstanding iterators are released. // It is valid to call Close multiple times. Other methods should not be // called after the DB has been closed.
[ "Close", "closes", "the", "DB", ".", "This", "will", "also", "releases", "any", "outstanding", "snapshot", "abort", "any", "in", "-", "flight", "compaction", "and", "discard", "open", "transaction", ".", "It", "is", "not", "safe", "to", "close", "a", "DB", "until", "all", "outstanding", "iterators", "are", "released", ".", "It", "is", "valid", "to", "call", "Close", "multiple", "times", ".", "Other", "methods", "should", "not", "be", "called", "after", "the", "DB", "has", "been", "closed", "." ]
c3a204f8e96543bb0cc090385c001078f184fc46
https://github.com/syndtr/goleveldb/blob/c3a204f8e96543bb0cc090385c001078f184fc46/leveldb/db.go#L1128-L1191
train
olivere/elastic
search_queries_match_all.go
Source
func (q MatchAllQuery) Source() (interface{}, error) { // { // "match_all" : { ... } // } source := make(map[string]interface{}) params := make(map[string]interface{}) source["match_all"] = params if q.boost != nil { params["boost"] = *q.boost } if q.queryName != "" { params["_name"] = q.queryName } return source, nil }
go
func (q MatchAllQuery) Source() (interface{}, error) { // { // "match_all" : { ... } // } source := make(map[string]interface{}) params := make(map[string]interface{}) source["match_all"] = params if q.boost != nil { params["boost"] = *q.boost } if q.queryName != "" { params["_name"] = q.queryName } return source, nil }
[ "func", "(", "q", "MatchAllQuery", ")", "Source", "(", ")", "(", "interface", "{", "}", ",", "error", ")", "{", "// {", "// \"match_all\" : { ... }", "// }", "source", ":=", "make", "(", "map", "[", "string", "]", "interface", "{", "}", ")", "\n", "params", ":=", "make", "(", "map", "[", "string", "]", "interface", "{", "}", ")", "\n", "source", "[", "\"", "\"", "]", "=", "params", "\n", "if", "q", ".", "boost", "!=", "nil", "{", "params", "[", "\"", "\"", "]", "=", "*", "q", ".", "boost", "\n", "}", "\n", "if", "q", ".", "queryName", "!=", "\"", "\"", "{", "params", "[", "\"", "\"", "]", "=", "q", ".", "queryName", "\n", "}", "\n", "return", "source", ",", "nil", "\n", "}" ]
// Source returns JSON for the match all query.
[ "Source", "returns", "JSON", "for", "the", "match", "all", "query", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/search_queries_match_all.go#L37-L51
train
olivere/elastic
search_aggs.go
UnmarshalJSON
func (a *AggregationValueMetric) UnmarshalJSON(data []byte) error { var aggs map[string]json.RawMessage if err := json.Unmarshal(data, &aggs); err != nil { return err } if v, ok := aggs["value"]; ok && v != nil { json.Unmarshal(v, &a.Value) } if v, ok := aggs["meta"]; ok && v != nil { json.Unmarshal(v, &a.Meta) } a.Aggregations = aggs return nil }
go
func (a *AggregationValueMetric) UnmarshalJSON(data []byte) error { var aggs map[string]json.RawMessage if err := json.Unmarshal(data, &aggs); err != nil { return err } if v, ok := aggs["value"]; ok && v != nil { json.Unmarshal(v, &a.Value) } if v, ok := aggs["meta"]; ok && v != nil { json.Unmarshal(v, &a.Meta) } a.Aggregations = aggs return nil }
[ "func", "(", "a", "*", "AggregationValueMetric", ")", "UnmarshalJSON", "(", "data", "[", "]", "byte", ")", "error", "{", "var", "aggs", "map", "[", "string", "]", "json", ".", "RawMessage", "\n", "if", "err", ":=", "json", ".", "Unmarshal", "(", "data", ",", "&", "aggs", ")", ";", "err", "!=", "nil", "{", "return", "err", "\n", "}", "\n", "if", "v", ",", "ok", ":=", "aggs", "[", "\"", "\"", "]", ";", "ok", "&&", "v", "!=", "nil", "{", "json", ".", "Unmarshal", "(", "v", ",", "&", "a", ".", "Value", ")", "\n", "}", "\n", "if", "v", ",", "ok", ":=", "aggs", "[", "\"", "\"", "]", ";", "ok", "&&", "v", "!=", "nil", "{", "json", ".", "Unmarshal", "(", "v", ",", "&", "a", ".", "Meta", ")", "\n", "}", "\n", "a", ".", "Aggregations", "=", "aggs", "\n", "return", "nil", "\n", "}" ]
// UnmarshalJSON decodes JSON data and initializes an AggregationValueMetric structure.
[ "UnmarshalJSON", "decodes", "JSON", "data", "and", "initializes", "an", "AggregationValueMetric", "structure", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/search_aggs.go#L778-L791
train
olivere/elastic
search_aggs.go
UnmarshalJSON
func (a *AggregationStatsMetric) UnmarshalJSON(data []byte) error { var aggs map[string]json.RawMessage if err := json.Unmarshal(data, &aggs); err != nil { return err } if v, ok := aggs["count"]; ok && v != nil { json.Unmarshal(v, &a.Count) } if v, ok := aggs["min"]; ok && v != nil { json.Unmarshal(v, &a.Min) } if v, ok := aggs["max"]; ok && v != nil { json.Unmarshal(v, &a.Max) } if v, ok := aggs["avg"]; ok && v != nil { json.Unmarshal(v, &a.Avg) } if v, ok := aggs["sum"]; ok && v != nil { json.Unmarshal(v, &a.Sum) } if v, ok := aggs["meta"]; ok && v != nil { json.Unmarshal(v, &a.Meta) } a.Aggregations = aggs return nil }
go
func (a *AggregationStatsMetric) UnmarshalJSON(data []byte) error { var aggs map[string]json.RawMessage if err := json.Unmarshal(data, &aggs); err != nil { return err } if v, ok := aggs["count"]; ok && v != nil { json.Unmarshal(v, &a.Count) } if v, ok := aggs["min"]; ok && v != nil { json.Unmarshal(v, &a.Min) } if v, ok := aggs["max"]; ok && v != nil { json.Unmarshal(v, &a.Max) } if v, ok := aggs["avg"]; ok && v != nil { json.Unmarshal(v, &a.Avg) } if v, ok := aggs["sum"]; ok && v != nil { json.Unmarshal(v, &a.Sum) } if v, ok := aggs["meta"]; ok && v != nil { json.Unmarshal(v, &a.Meta) } a.Aggregations = aggs return nil }
[ "func", "(", "a", "*", "AggregationStatsMetric", ")", "UnmarshalJSON", "(", "data", "[", "]", "byte", ")", "error", "{", "var", "aggs", "map", "[", "string", "]", "json", ".", "RawMessage", "\n", "if", "err", ":=", "json", ".", "Unmarshal", "(", "data", ",", "&", "aggs", ")", ";", "err", "!=", "nil", "{", "return", "err", "\n", "}", "\n", "if", "v", ",", "ok", ":=", "aggs", "[", "\"", "\"", "]", ";", "ok", "&&", "v", "!=", "nil", "{", "json", ".", "Unmarshal", "(", "v", ",", "&", "a", ".", "Count", ")", "\n", "}", "\n", "if", "v", ",", "ok", ":=", "aggs", "[", "\"", "\"", "]", ";", "ok", "&&", "v", "!=", "nil", "{", "json", ".", "Unmarshal", "(", "v", ",", "&", "a", ".", "Min", ")", "\n", "}", "\n", "if", "v", ",", "ok", ":=", "aggs", "[", "\"", "\"", "]", ";", "ok", "&&", "v", "!=", "nil", "{", "json", ".", "Unmarshal", "(", "v", ",", "&", "a", ".", "Max", ")", "\n", "}", "\n", "if", "v", ",", "ok", ":=", "aggs", "[", "\"", "\"", "]", ";", "ok", "&&", "v", "!=", "nil", "{", "json", ".", "Unmarshal", "(", "v", ",", "&", "a", ".", "Avg", ")", "\n", "}", "\n", "if", "v", ",", "ok", ":=", "aggs", "[", "\"", "\"", "]", ";", "ok", "&&", "v", "!=", "nil", "{", "json", ".", "Unmarshal", "(", "v", ",", "&", "a", ".", "Sum", ")", "\n", "}", "\n", "if", "v", ",", "ok", ":=", "aggs", "[", "\"", "\"", "]", ";", "ok", "&&", "v", "!=", "nil", "{", "json", ".", "Unmarshal", "(", "v", ",", "&", "a", ".", "Meta", ")", "\n", "}", "\n", "a", ".", "Aggregations", "=", "aggs", "\n", "return", "nil", "\n", "}" ]
// UnmarshalJSON decodes JSON data and initializes an AggregationStatsMetric structure.
[ "UnmarshalJSON", "decodes", "JSON", "data", "and", "initializes", "an", "AggregationStatsMetric", "structure", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/search_aggs.go#L808-L833
train
olivere/elastic
search_aggs.go
UnmarshalJSON
func (a *AggregationGeoCentroidMetric) UnmarshalJSON(data []byte) error { var aggs map[string]json.RawMessage if err := json.Unmarshal(data, &aggs); err != nil { return err } if v, ok := aggs["location"]; ok && v != nil { json.Unmarshal(v, &a.Location) } if v, ok := aggs["meta"]; ok && v != nil { json.Unmarshal(v, &a.Meta) } if v, ok := aggs["count"]; ok && v != nil { json.Unmarshal(v, &a.Count) } a.Aggregations = aggs return nil }
go
func (a *AggregationGeoCentroidMetric) UnmarshalJSON(data []byte) error { var aggs map[string]json.RawMessage if err := json.Unmarshal(data, &aggs); err != nil { return err } if v, ok := aggs["location"]; ok && v != nil { json.Unmarshal(v, &a.Location) } if v, ok := aggs["meta"]; ok && v != nil { json.Unmarshal(v, &a.Meta) } if v, ok := aggs["count"]; ok && v != nil { json.Unmarshal(v, &a.Count) } a.Aggregations = aggs return nil }
[ "func", "(", "a", "*", "AggregationGeoCentroidMetric", ")", "UnmarshalJSON", "(", "data", "[", "]", "byte", ")", "error", "{", "var", "aggs", "map", "[", "string", "]", "json", ".", "RawMessage", "\n", "if", "err", ":=", "json", ".", "Unmarshal", "(", "data", ",", "&", "aggs", ")", ";", "err", "!=", "nil", "{", "return", "err", "\n", "}", "\n", "if", "v", ",", "ok", ":=", "aggs", "[", "\"", "\"", "]", ";", "ok", "&&", "v", "!=", "nil", "{", "json", ".", "Unmarshal", "(", "v", ",", "&", "a", ".", "Location", ")", "\n", "}", "\n", "if", "v", ",", "ok", ":=", "aggs", "[", "\"", "\"", "]", ";", "ok", "&&", "v", "!=", "nil", "{", "json", ".", "Unmarshal", "(", "v", ",", "&", "a", ".", "Meta", ")", "\n", "}", "\n", "if", "v", ",", "ok", ":=", "aggs", "[", "\"", "\"", "]", ";", "ok", "&&", "v", "!=", "nil", "{", "json", ".", "Unmarshal", "(", "v", ",", "&", "a", ".", "Count", ")", "\n", "}", "\n", "a", ".", "Aggregations", "=", "aggs", "\n", "return", "nil", "\n", "}" ]
// UnmarshalJSON decodes JSON data and initializes an AggregationGeoCentroidMetric structure.
[ "UnmarshalJSON", "decodes", "JSON", "data", "and", "initializes", "an", "AggregationGeoCentroidMetric", "structure", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/search_aggs.go#L1033-L1049
train
olivere/elastic
search_aggs.go
UnmarshalJSON
func (a *AggregationBucketRangeItem) UnmarshalJSON(data []byte) error { var aggs map[string]json.RawMessage if err := json.Unmarshal(data, &aggs); err != nil { return err } if v, ok := aggs["key"]; ok && v != nil { json.Unmarshal(v, &a.Key) } if v, ok := aggs["doc_count"]; ok && v != nil { json.Unmarshal(v, &a.DocCount) } if v, ok := aggs["from"]; ok && v != nil { json.Unmarshal(v, &a.From) } if v, ok := aggs["from_as_string"]; ok && v != nil { json.Unmarshal(v, &a.FromAsString) } if v, ok := aggs["to"]; ok && v != nil { json.Unmarshal(v, &a.To) } if v, ok := aggs["to_as_string"]; ok && v != nil { json.Unmarshal(v, &a.ToAsString) } a.Aggregations = aggs return nil }
go
func (a *AggregationBucketRangeItem) UnmarshalJSON(data []byte) error { var aggs map[string]json.RawMessage if err := json.Unmarshal(data, &aggs); err != nil { return err } if v, ok := aggs["key"]; ok && v != nil { json.Unmarshal(v, &a.Key) } if v, ok := aggs["doc_count"]; ok && v != nil { json.Unmarshal(v, &a.DocCount) } if v, ok := aggs["from"]; ok && v != nil { json.Unmarshal(v, &a.From) } if v, ok := aggs["from_as_string"]; ok && v != nil { json.Unmarshal(v, &a.FromAsString) } if v, ok := aggs["to"]; ok && v != nil { json.Unmarshal(v, &a.To) } if v, ok := aggs["to_as_string"]; ok && v != nil { json.Unmarshal(v, &a.ToAsString) } a.Aggregations = aggs return nil }
[ "func", "(", "a", "*", "AggregationBucketRangeItem", ")", "UnmarshalJSON", "(", "data", "[", "]", "byte", ")", "error", "{", "var", "aggs", "map", "[", "string", "]", "json", ".", "RawMessage", "\n", "if", "err", ":=", "json", ".", "Unmarshal", "(", "data", ",", "&", "aggs", ")", ";", "err", "!=", "nil", "{", "return", "err", "\n", "}", "\n", "if", "v", ",", "ok", ":=", "aggs", "[", "\"", "\"", "]", ";", "ok", "&&", "v", "!=", "nil", "{", "json", ".", "Unmarshal", "(", "v", ",", "&", "a", ".", "Key", ")", "\n", "}", "\n", "if", "v", ",", "ok", ":=", "aggs", "[", "\"", "\"", "]", ";", "ok", "&&", "v", "!=", "nil", "{", "json", ".", "Unmarshal", "(", "v", ",", "&", "a", ".", "DocCount", ")", "\n", "}", "\n", "if", "v", ",", "ok", ":=", "aggs", "[", "\"", "\"", "]", ";", "ok", "&&", "v", "!=", "nil", "{", "json", ".", "Unmarshal", "(", "v", ",", "&", "a", ".", "From", ")", "\n", "}", "\n", "if", "v", ",", "ok", ":=", "aggs", "[", "\"", "\"", "]", ";", "ok", "&&", "v", "!=", "nil", "{", "json", ".", "Unmarshal", "(", "v", ",", "&", "a", ".", "FromAsString", ")", "\n", "}", "\n", "if", "v", ",", "ok", ":=", "aggs", "[", "\"", "\"", "]", ";", "ok", "&&", "v", "!=", "nil", "{", "json", ".", "Unmarshal", "(", "v", ",", "&", "a", ".", "To", ")", "\n", "}", "\n", "if", "v", ",", "ok", ":=", "aggs", "[", "\"", "\"", "]", ";", "ok", "&&", "v", "!=", "nil", "{", "json", ".", "Unmarshal", "(", "v", ",", "&", "a", ".", "ToAsString", ")", "\n", "}", "\n", "a", ".", "Aggregations", "=", "aggs", "\n", "return", "nil", "\n", "}" ]
// UnmarshalJSON decodes JSON data and initializes an AggregationBucketRangeItem structure.
[ "UnmarshalJSON", "decodes", "JSON", "data", "and", "initializes", "an", "AggregationBucketRangeItem", "structure", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/search_aggs.go#L1158-L1183
train
olivere/elastic
search_aggs.go
UnmarshalJSON
func (a *AggregationBucketHistogramItem) UnmarshalJSON(data []byte) error { var aggs map[string]json.RawMessage if err := json.Unmarshal(data, &aggs); err != nil { return err } if v, ok := aggs["key"]; ok && v != nil { json.Unmarshal(v, &a.Key) } if v, ok := aggs["key_as_string"]; ok && v != nil { json.Unmarshal(v, &a.KeyAsString) } if v, ok := aggs["doc_count"]; ok && v != nil { json.Unmarshal(v, &a.DocCount) } a.Aggregations = aggs return nil }
go
func (a *AggregationBucketHistogramItem) UnmarshalJSON(data []byte) error { var aggs map[string]json.RawMessage if err := json.Unmarshal(data, &aggs); err != nil { return err } if v, ok := aggs["key"]; ok && v != nil { json.Unmarshal(v, &a.Key) } if v, ok := aggs["key_as_string"]; ok && v != nil { json.Unmarshal(v, &a.KeyAsString) } if v, ok := aggs["doc_count"]; ok && v != nil { json.Unmarshal(v, &a.DocCount) } a.Aggregations = aggs return nil }
[ "func", "(", "a", "*", "AggregationBucketHistogramItem", ")", "UnmarshalJSON", "(", "data", "[", "]", "byte", ")", "error", "{", "var", "aggs", "map", "[", "string", "]", "json", ".", "RawMessage", "\n", "if", "err", ":=", "json", ".", "Unmarshal", "(", "data", ",", "&", "aggs", ")", ";", "err", "!=", "nil", "{", "return", "err", "\n", "}", "\n", "if", "v", ",", "ok", ":=", "aggs", "[", "\"", "\"", "]", ";", "ok", "&&", "v", "!=", "nil", "{", "json", ".", "Unmarshal", "(", "v", ",", "&", "a", ".", "Key", ")", "\n", "}", "\n", "if", "v", ",", "ok", ":=", "aggs", "[", "\"", "\"", "]", ";", "ok", "&&", "v", "!=", "nil", "{", "json", ".", "Unmarshal", "(", "v", ",", "&", "a", ".", "KeyAsString", ")", "\n", "}", "\n", "if", "v", ",", "ok", ":=", "aggs", "[", "\"", "\"", "]", ";", "ok", "&&", "v", "!=", "nil", "{", "json", ".", "Unmarshal", "(", "v", ",", "&", "a", ".", "DocCount", ")", "\n", "}", "\n", "a", ".", "Aggregations", "=", "aggs", "\n", "return", "nil", "\n", "}" ]
// UnmarshalJSON decodes JSON data and initializes an AggregationBucketHistogramItem structure.
[ "UnmarshalJSON", "decodes", "JSON", "data", "and", "initializes", "an", "AggregationBucketHistogramItem", "structure", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/search_aggs.go#L1433-L1449
train
olivere/elastic
search_aggs.go
UnmarshalJSON
func (a *AggregationPipelineDerivative) UnmarshalJSON(data []byte) error { var aggs map[string]json.RawMessage if err := json.Unmarshal(data, &aggs); err != nil { return err } if v, ok := aggs["value"]; ok && v != nil { json.Unmarshal(v, &a.Value) } if v, ok := aggs["value_as_string"]; ok && v != nil { json.Unmarshal(v, &a.ValueAsString) } if v, ok := aggs["normalized_value"]; ok && v != nil { json.Unmarshal(v, &a.NormalizedValue) } if v, ok := aggs["normalized_value_as_string"]; ok && v != nil { json.Unmarshal(v, &a.NormalizedValueAsString) } if v, ok := aggs["meta"]; ok && v != nil { json.Unmarshal(v, &a.Meta) } a.Aggregations = aggs return nil }
go
func (a *AggregationPipelineDerivative) UnmarshalJSON(data []byte) error { var aggs map[string]json.RawMessage if err := json.Unmarshal(data, &aggs); err != nil { return err } if v, ok := aggs["value"]; ok && v != nil { json.Unmarshal(v, &a.Value) } if v, ok := aggs["value_as_string"]; ok && v != nil { json.Unmarshal(v, &a.ValueAsString) } if v, ok := aggs["normalized_value"]; ok && v != nil { json.Unmarshal(v, &a.NormalizedValue) } if v, ok := aggs["normalized_value_as_string"]; ok && v != nil { json.Unmarshal(v, &a.NormalizedValueAsString) } if v, ok := aggs["meta"]; ok && v != nil { json.Unmarshal(v, &a.Meta) } a.Aggregations = aggs return nil }
[ "func", "(", "a", "*", "AggregationPipelineDerivative", ")", "UnmarshalJSON", "(", "data", "[", "]", "byte", ")", "error", "{", "var", "aggs", "map", "[", "string", "]", "json", ".", "RawMessage", "\n", "if", "err", ":=", "json", ".", "Unmarshal", "(", "data", ",", "&", "aggs", ")", ";", "err", "!=", "nil", "{", "return", "err", "\n", "}", "\n", "if", "v", ",", "ok", ":=", "aggs", "[", "\"", "\"", "]", ";", "ok", "&&", "v", "!=", "nil", "{", "json", ".", "Unmarshal", "(", "v", ",", "&", "a", ".", "Value", ")", "\n", "}", "\n", "if", "v", ",", "ok", ":=", "aggs", "[", "\"", "\"", "]", ";", "ok", "&&", "v", "!=", "nil", "{", "json", ".", "Unmarshal", "(", "v", ",", "&", "a", ".", "ValueAsString", ")", "\n", "}", "\n", "if", "v", ",", "ok", ":=", "aggs", "[", "\"", "\"", "]", ";", "ok", "&&", "v", "!=", "nil", "{", "json", ".", "Unmarshal", "(", "v", ",", "&", "a", ".", "NormalizedValue", ")", "\n", "}", "\n", "if", "v", ",", "ok", ":=", "aggs", "[", "\"", "\"", "]", ";", "ok", "&&", "v", "!=", "nil", "{", "json", ".", "Unmarshal", "(", "v", ",", "&", "a", ".", "NormalizedValueAsString", ")", "\n", "}", "\n", "if", "v", ",", "ok", ":=", "aggs", "[", "\"", "\"", "]", ";", "ok", "&&", "v", "!=", "nil", "{", "json", ".", "Unmarshal", "(", "v", ",", "&", "a", ".", "Meta", ")", "\n", "}", "\n", "a", ".", "Aggregations", "=", "aggs", "\n", "return", "nil", "\n", "}" ]
// UnmarshalJSON decodes JSON data and initializes an AggregationPipelineDerivative structure.
[ "UnmarshalJSON", "decodes", "JSON", "data", "and", "initializes", "an", "AggregationPipelineDerivative", "structure", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/search_aggs.go#L1532-L1554
train
olivere/elastic
search_aggs.go
UnmarshalJSON
func (a *AggregationPipelineStatsMetric) UnmarshalJSON(data []byte) error { var aggs map[string]json.RawMessage if err := json.Unmarshal(data, &aggs); err != nil { return err } if v, ok := aggs["count"]; ok && v != nil { json.Unmarshal(v, &a.Count) } if v, ok := aggs["count_as_string"]; ok && v != nil { json.Unmarshal(v, &a.CountAsString) } if v, ok := aggs["min"]; ok && v != nil { json.Unmarshal(v, &a.Min) } if v, ok := aggs["min_as_string"]; ok && v != nil { json.Unmarshal(v, &a.MinAsString) } if v, ok := aggs["max"]; ok && v != nil { json.Unmarshal(v, &a.Max) } if v, ok := aggs["max_as_string"]; ok && v != nil { json.Unmarshal(v, &a.MaxAsString) } if v, ok := aggs["avg"]; ok && v != nil { json.Unmarshal(v, &a.Avg) } if v, ok := aggs["avg_as_string"]; ok && v != nil { json.Unmarshal(v, &a.AvgAsString) } if v, ok := aggs["sum"]; ok && v != nil { json.Unmarshal(v, &a.Sum) } if v, ok := aggs["sum_as_string"]; ok && v != nil { json.Unmarshal(v, &a.SumAsString) } if v, ok := aggs["meta"]; ok && v != nil { json.Unmarshal(v, &a.Meta) } a.Aggregations = aggs return nil }
go
func (a *AggregationPipelineStatsMetric) UnmarshalJSON(data []byte) error { var aggs map[string]json.RawMessage if err := json.Unmarshal(data, &aggs); err != nil { return err } if v, ok := aggs["count"]; ok && v != nil { json.Unmarshal(v, &a.Count) } if v, ok := aggs["count_as_string"]; ok && v != nil { json.Unmarshal(v, &a.CountAsString) } if v, ok := aggs["min"]; ok && v != nil { json.Unmarshal(v, &a.Min) } if v, ok := aggs["min_as_string"]; ok && v != nil { json.Unmarshal(v, &a.MinAsString) } if v, ok := aggs["max"]; ok && v != nil { json.Unmarshal(v, &a.Max) } if v, ok := aggs["max_as_string"]; ok && v != nil { json.Unmarshal(v, &a.MaxAsString) } if v, ok := aggs["avg"]; ok && v != nil { json.Unmarshal(v, &a.Avg) } if v, ok := aggs["avg_as_string"]; ok && v != nil { json.Unmarshal(v, &a.AvgAsString) } if v, ok := aggs["sum"]; ok && v != nil { json.Unmarshal(v, &a.Sum) } if v, ok := aggs["sum_as_string"]; ok && v != nil { json.Unmarshal(v, &a.SumAsString) } if v, ok := aggs["meta"]; ok && v != nil { json.Unmarshal(v, &a.Meta) } a.Aggregations = aggs return nil }
[ "func", "(", "a", "*", "AggregationPipelineStatsMetric", ")", "UnmarshalJSON", "(", "data", "[", "]", "byte", ")", "error", "{", "var", "aggs", "map", "[", "string", "]", "json", ".", "RawMessage", "\n", "if", "err", ":=", "json", ".", "Unmarshal", "(", "data", ",", "&", "aggs", ")", ";", "err", "!=", "nil", "{", "return", "err", "\n", "}", "\n", "if", "v", ",", "ok", ":=", "aggs", "[", "\"", "\"", "]", ";", "ok", "&&", "v", "!=", "nil", "{", "json", ".", "Unmarshal", "(", "v", ",", "&", "a", ".", "Count", ")", "\n", "}", "\n", "if", "v", ",", "ok", ":=", "aggs", "[", "\"", "\"", "]", ";", "ok", "&&", "v", "!=", "nil", "{", "json", ".", "Unmarshal", "(", "v", ",", "&", "a", ".", "CountAsString", ")", "\n", "}", "\n", "if", "v", ",", "ok", ":=", "aggs", "[", "\"", "\"", "]", ";", "ok", "&&", "v", "!=", "nil", "{", "json", ".", "Unmarshal", "(", "v", ",", "&", "a", ".", "Min", ")", "\n", "}", "\n", "if", "v", ",", "ok", ":=", "aggs", "[", "\"", "\"", "]", ";", "ok", "&&", "v", "!=", "nil", "{", "json", ".", "Unmarshal", "(", "v", ",", "&", "a", ".", "MinAsString", ")", "\n", "}", "\n", "if", "v", ",", "ok", ":=", "aggs", "[", "\"", "\"", "]", ";", "ok", "&&", "v", "!=", "nil", "{", "json", ".", "Unmarshal", "(", "v", ",", "&", "a", ".", "Max", ")", "\n", "}", "\n", "if", "v", ",", "ok", ":=", "aggs", "[", "\"", "\"", "]", ";", "ok", "&&", "v", "!=", "nil", "{", "json", ".", "Unmarshal", "(", "v", ",", "&", "a", ".", "MaxAsString", ")", "\n", "}", "\n", "if", "v", ",", "ok", ":=", "aggs", "[", "\"", "\"", "]", ";", "ok", "&&", "v", "!=", "nil", "{", "json", ".", "Unmarshal", "(", "v", ",", "&", "a", ".", "Avg", ")", "\n", "}", "\n", "if", "v", ",", "ok", ":=", "aggs", "[", "\"", "\"", "]", ";", "ok", "&&", "v", "!=", "nil", "{", "json", ".", "Unmarshal", "(", "v", ",", "&", "a", ".", "AvgAsString", ")", "\n", "}", "\n", "if", "v", ",", "ok", ":=", "aggs", "[", "\"", "\"", "]", ";", "ok", "&&", "v", "!=", "nil", "{", "json", ".", "Unmarshal", "(", "v", ",", "&", "a", ".", "Sum", ")", "\n", "}", "\n", "if", "v", ",", "ok", ":=", "aggs", "[", "\"", "\"", "]", ";", "ok", "&&", "v", "!=", "nil", "{", "json", ".", "Unmarshal", "(", "v", ",", "&", "a", ".", "SumAsString", ")", "\n", "}", "\n", "if", "v", ",", "ok", ":=", "aggs", "[", "\"", "\"", "]", ";", "ok", "&&", "v", "!=", "nil", "{", "json", ".", "Unmarshal", "(", "v", ",", "&", "a", ".", "Meta", ")", "\n", "}", "\n", "a", ".", "Aggregations", "=", "aggs", "\n", "return", "nil", "\n", "}" ]
// UnmarshalJSON decodes JSON data and initializes an AggregationPipelineStatsMetric structure.
[ "UnmarshalJSON", "decodes", "JSON", "data", "and", "initializes", "an", "AggregationPipelineStatsMetric", "structure", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/search_aggs.go#L1578-L1618
train
olivere/elastic
ingest_get_pipeline.go
Id
func (s *IngestGetPipelineService) Id(id ...string) *IngestGetPipelineService { s.id = append(s.id, id...) return s }
go
func (s *IngestGetPipelineService) Id(id ...string) *IngestGetPipelineService { s.id = append(s.id, id...) return s }
[ "func", "(", "s", "*", "IngestGetPipelineService", ")", "Id", "(", "id", "...", "string", ")", "*", "IngestGetPipelineService", "{", "s", ".", "id", "=", "append", "(", "s", ".", "id", ",", "id", "...", ")", "\n", "return", "s", "\n", "}" ]
// Id is a list of pipeline ids. Wildcards supported.
[ "Id", "is", "a", "list", "of", "pipeline", "ids", ".", "Wildcards", "supported", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/ingest_get_pipeline.go#L34-L37
train
olivere/elastic
search_aggs_pipeline_percentiles_bucket.go
Format
func (p *PercentilesBucketAggregation) Format(format string) *PercentilesBucketAggregation { p.format = format return p }
go
func (p *PercentilesBucketAggregation) Format(format string) *PercentilesBucketAggregation { p.format = format return p }
[ "func", "(", "p", "*", "PercentilesBucketAggregation", ")", "Format", "(", "format", "string", ")", "*", "PercentilesBucketAggregation", "{", "p", ".", "format", "=", "format", "\n", "return", "p", "\n", "}" ]
// Format to apply the output value of this aggregation.
[ "Format", "to", "apply", "the", "output", "value", "of", "this", "aggregation", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/search_aggs_pipeline_percentiles_bucket.go#L29-L32
train
olivere/elastic
search_aggs_pipeline_percentiles_bucket.go
Percents
func (p *PercentilesBucketAggregation) Percents(percents ...float64) *PercentilesBucketAggregation { p.percents = percents return p }
go
func (p *PercentilesBucketAggregation) Percents(percents ...float64) *PercentilesBucketAggregation { p.percents = percents return p }
[ "func", "(", "p", "*", "PercentilesBucketAggregation", ")", "Percents", "(", "percents", "...", "float64", ")", "*", "PercentilesBucketAggregation", "{", "p", ".", "percents", "=", "percents", "\n", "return", "p", "\n", "}" ]
// Percents to calculate percentiles for in this aggregation.
[ "Percents", "to", "calculate", "percentiles", "for", "in", "this", "aggregation", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/search_aggs_pipeline_percentiles_bucket.go#L35-L38
train
olivere/elastic
search_queries_ids.go
Ids
func (q *IdsQuery) Ids(ids ...string) *IdsQuery { q.values = append(q.values, ids...) return q }
go
func (q *IdsQuery) Ids(ids ...string) *IdsQuery { q.values = append(q.values, ids...) return q }
[ "func", "(", "q", "*", "IdsQuery", ")", "Ids", "(", "ids", "...", "string", ")", "*", "IdsQuery", "{", "q", ".", "values", "=", "append", "(", "q", ".", "values", ",", "ids", "...", ")", "\n", "return", "q", "\n", "}" ]
// Ids adds ids to the filter.
[ "Ids", "adds", "ids", "to", "the", "filter", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/search_queries_ids.go#L30-L33
train
olivere/elastic
search_queries_ids.go
QueryName
func (q *IdsQuery) QueryName(queryName string) *IdsQuery { q.queryName = queryName return q }
go
func (q *IdsQuery) QueryName(queryName string) *IdsQuery { q.queryName = queryName return q }
[ "func", "(", "q", "*", "IdsQuery", ")", "QueryName", "(", "queryName", "string", ")", "*", "IdsQuery", "{", "q", ".", "queryName", "=", "queryName", "\n", "return", "q", "\n", "}" ]
// QueryName sets the query name for the filter.
[ "QueryName", "sets", "the", "query", "name", "for", "the", "filter", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/search_queries_ids.go#L42-L45
train
olivere/elastic
snapshot_create_repository.go
Verify
func (s *SnapshotCreateRepositoryService) Verify(verify bool) *SnapshotCreateRepositoryService { s.verify = &verify return s }
go
func (s *SnapshotCreateRepositoryService) Verify(verify bool) *SnapshotCreateRepositoryService { s.verify = &verify return s }
[ "func", "(", "s", "*", "SnapshotCreateRepositoryService", ")", "Verify", "(", "verify", "bool", ")", "*", "SnapshotCreateRepositoryService", "{", "s", ".", "verify", "=", "&", "verify", "\n", "return", "s", "\n", "}" ]
// Verify indicates whether to verify the repository after creation.
[ "Verify", "indicates", "whether", "to", "verify", "the", "repository", "after", "creation", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/snapshot_create_repository.go#L58-L61
train
olivere/elastic
snapshot_create_repository.go
Settings
func (s *SnapshotCreateRepositoryService) Settings(settings map[string]interface{}) *SnapshotCreateRepositoryService { s.settings = settings return s }
go
func (s *SnapshotCreateRepositoryService) Settings(settings map[string]interface{}) *SnapshotCreateRepositoryService { s.settings = settings return s }
[ "func", "(", "s", "*", "SnapshotCreateRepositoryService", ")", "Settings", "(", "settings", "map", "[", "string", "]", "interface", "{", "}", ")", "*", "SnapshotCreateRepositoryService", "{", "s", ".", "settings", "=", "settings", "\n", "return", "s", "\n", "}" ]
// Settings sets all settings of the snapshot repository.
[ "Settings", "sets", "all", "settings", "of", "the", "snapshot", "repository", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/snapshot_create_repository.go#L76-L79
train
olivere/elastic
snapshot_create_repository.go
Setting
func (s *SnapshotCreateRepositoryService) Setting(name string, value interface{}) *SnapshotCreateRepositoryService { if s.settings == nil { s.settings = make(map[string]interface{}) } s.settings[name] = value return s }
go
func (s *SnapshotCreateRepositoryService) Setting(name string, value interface{}) *SnapshotCreateRepositoryService { if s.settings == nil { s.settings = make(map[string]interface{}) } s.settings[name] = value return s }
[ "func", "(", "s", "*", "SnapshotCreateRepositoryService", ")", "Setting", "(", "name", "string", ",", "value", "interface", "{", "}", ")", "*", "SnapshotCreateRepositoryService", "{", "if", "s", ".", "settings", "==", "nil", "{", "s", ".", "settings", "=", "make", "(", "map", "[", "string", "]", "interface", "{", "}", ")", "\n", "}", "\n", "s", ".", "settings", "[", "name", "]", "=", "value", "\n", "return", "s", "\n", "}" ]
// Setting sets a single settings of the snapshot repository.
[ "Setting", "sets", "a", "single", "settings", "of", "the", "snapshot", "repository", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/snapshot_create_repository.go#L82-L88
train
olivere/elastic
snapshot_create_repository.go
buildBody
func (s *SnapshotCreateRepositoryService) buildBody() (interface{}, error) { if s.bodyJson != nil { return s.bodyJson, nil } if s.bodyString != "" { return s.bodyString, nil } body := map[string]interface{}{ "type": s.typ, } if len(s.settings) > 0 { body["settings"] = s.settings } return body, nil }
go
func (s *SnapshotCreateRepositoryService) buildBody() (interface{}, error) { if s.bodyJson != nil { return s.bodyJson, nil } if s.bodyString != "" { return s.bodyString, nil } body := map[string]interface{}{ "type": s.typ, } if len(s.settings) > 0 { body["settings"] = s.settings } return body, nil }
[ "func", "(", "s", "*", "SnapshotCreateRepositoryService", ")", "buildBody", "(", ")", "(", "interface", "{", "}", ",", "error", ")", "{", "if", "s", ".", "bodyJson", "!=", "nil", "{", "return", "s", ".", "bodyJson", ",", "nil", "\n", "}", "\n", "if", "s", ".", "bodyString", "!=", "\"", "\"", "{", "return", "s", ".", "bodyString", ",", "nil", "\n", "}", "\n\n", "body", ":=", "map", "[", "string", "]", "interface", "{", "}", "{", "\"", "\"", ":", "s", ".", "typ", ",", "}", "\n", "if", "len", "(", "s", ".", "settings", ")", ">", "0", "{", "body", "[", "\"", "\"", "]", "=", "s", ".", "settings", "\n", "}", "\n", "return", "body", ",", "nil", "\n", "}" ]
// buildBody builds the body for the operation.
[ "buildBody", "builds", "the", "body", "for", "the", "operation", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/snapshot_create_repository.go#L130-L145
train
olivere/elastic
indices_exists.go
NewIndicesExistsService
func NewIndicesExistsService(client *Client) *IndicesExistsService { return &IndicesExistsService{ client: client, index: make([]string, 0), } }
go
func NewIndicesExistsService(client *Client) *IndicesExistsService { return &IndicesExistsService{ client: client, index: make([]string, 0), } }
[ "func", "NewIndicesExistsService", "(", "client", "*", "Client", ")", "*", "IndicesExistsService", "{", "return", "&", "IndicesExistsService", "{", "client", ":", "client", ",", "index", ":", "make", "(", "[", "]", "string", ",", "0", ")", ",", "}", "\n", "}" ]
// NewIndicesExistsService creates and initializes a new IndicesExistsService.
[ "NewIndicesExistsService", "creates", "and", "initializes", "a", "new", "IndicesExistsService", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/indices_exists.go#L32-L37
train
olivere/elastic
indices_exists.go
Index
func (s *IndicesExistsService) Index(index []string) *IndicesExistsService { s.index = index return s }
go
func (s *IndicesExistsService) Index(index []string) *IndicesExistsService { s.index = index return s }
[ "func", "(", "s", "*", "IndicesExistsService", ")", "Index", "(", "index", "[", "]", "string", ")", "*", "IndicesExistsService", "{", "s", ".", "index", "=", "index", "\n", "return", "s", "\n", "}" ]
// Index is a list of one or more indices to check.
[ "Index", "is", "a", "list", "of", "one", "or", "more", "indices", "to", "check", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/indices_exists.go#L40-L43
train
olivere/elastic
xpack_security_put_role.go
Name
func (s *XPackSecurityPutRoleService) Name(name string) *XPackSecurityPutRoleService { s.name = name return s }
go
func (s *XPackSecurityPutRoleService) Name(name string) *XPackSecurityPutRoleService { s.name = name return s }
[ "func", "(", "s", "*", "XPackSecurityPutRoleService", ")", "Name", "(", "name", "string", ")", "*", "XPackSecurityPutRoleService", "{", "s", ".", "name", "=", "name", "\n", "return", "s", "\n", "}" ]
// Name is name of the role to create.
[ "Name", "is", "name", "of", "the", "role", "to", "create", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/xpack_security_put_role.go#L33-L36
train
olivere/elastic
search_queries_range.go
Gt
func (q *RangeQuery) Gt(from interface{}) *RangeQuery { q.from = from q.includeLower = false return q }
go
func (q *RangeQuery) Gt(from interface{}) *RangeQuery { q.from = from q.includeLower = false return q }
[ "func", "(", "q", "*", "RangeQuery", ")", "Gt", "(", "from", "interface", "{", "}", ")", "*", "RangeQuery", "{", "q", ".", "from", "=", "from", "\n", "q", ".", "includeLower", "=", "false", "\n", "return", "q", "\n", "}" ]
// Gt indicates a greater-than value for the from part. // Use nil to indicate an unbounded from part.
[ "Gt", "indicates", "a", "greater", "-", "than", "value", "for", "the", "from", "part", ".", "Use", "nil", "to", "indicate", "an", "unbounded", "from", "part", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/search_queries_range.go#L38-L42
train
olivere/elastic
search_queries_range.go
Gte
func (q *RangeQuery) Gte(from interface{}) *RangeQuery { q.from = from q.includeLower = true return q }
go
func (q *RangeQuery) Gte(from interface{}) *RangeQuery { q.from = from q.includeLower = true return q }
[ "func", "(", "q", "*", "RangeQuery", ")", "Gte", "(", "from", "interface", "{", "}", ")", "*", "RangeQuery", "{", "q", ".", "from", "=", "from", "\n", "q", ".", "includeLower", "=", "true", "\n", "return", "q", "\n", "}" ]
// Gte indicates a greater-than-or-equal value for the from part. // Use nil to indicate an unbounded from part.
[ "Gte", "indicates", "a", "greater", "-", "than", "-", "or", "-", "equal", "value", "for", "the", "from", "part", ".", "Use", "nil", "to", "indicate", "an", "unbounded", "from", "part", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/search_queries_range.go#L46-L50
train
olivere/elastic
search_queries_range.go
Format
func (q *RangeQuery) Format(format string) *RangeQuery { q.format = format return q }
go
func (q *RangeQuery) Format(format string) *RangeQuery { q.format = format return q }
[ "func", "(", "q", "*", "RangeQuery", ")", "Format", "(", "format", "string", ")", "*", "RangeQuery", "{", "q", ".", "format", "=", "format", "\n", "return", "q", "\n", "}" ]
// Format is used for date fields. In that case, we can set the format // to be used instead of the mapper format.
[ "Format", "is", "used", "for", "date", "fields", ".", "In", "that", "case", "we", "can", "set", "the", "format", "to", "be", "used", "instead", "of", "the", "mapper", "format", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/search_queries_range.go#L111-L114
train
olivere/elastic
nodes_stats.go
Metric
func (s *NodesStatsService) Metric(metric ...string) *NodesStatsService { s.metric = append(s.metric, metric...) return s }
go
func (s *NodesStatsService) Metric(metric ...string) *NodesStatsService { s.metric = append(s.metric, metric...) return s }
[ "func", "(", "s", "*", "NodesStatsService", ")", "Metric", "(", "metric", "...", "string", ")", "*", "NodesStatsService", "{", "s", ".", "metric", "=", "append", "(", "s", ".", "metric", ",", "metric", "...", ")", "\n", "return", "s", "\n", "}" ]
// Metric limits the information returned to the specified metrics.
[ "Metric", "limits", "the", "information", "returned", "to", "the", "specified", "metrics", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/nodes_stats.go#L44-L47
train
olivere/elastic
nodes_stats.go
Level
func (s *NodesStatsService) Level(level string) *NodesStatsService { s.level = level return s }
go
func (s *NodesStatsService) Level(level string) *NodesStatsService { s.level = level return s }
[ "func", "(", "s", "*", "NodesStatsService", ")", "Level", "(", "level", "string", ")", "*", "NodesStatsService", "{", "s", ".", "level", "=", "level", "\n", "return", "s", "\n", "}" ]
// Level specifies whether to return indices stats aggregated at node, index or shard level.
[ "Level", "specifies", "whether", "to", "return", "indices", "stats", "aggregated", "at", "node", "index", "or", "shard", "level", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/nodes_stats.go#L97-L100
train
olivere/elastic
nodes_stats.go
Types
func (s *NodesStatsService) Types(types ...string) *NodesStatsService { s.types = append(s.types, types...) return s }
go
func (s *NodesStatsService) Types(types ...string) *NodesStatsService { s.types = append(s.types, types...) return s }
[ "func", "(", "s", "*", "NodesStatsService", ")", "Types", "(", "types", "...", "string", ")", "*", "NodesStatsService", "{", "s", ".", "types", "=", "append", "(", "s", ".", "types", ",", "types", "...", ")", "\n", "return", "s", "\n", "}" ]
// Types a list of document types for the `indexing` index metric.
[ "Types", "a", "list", "of", "document", "types", "for", "the", "indexing", "index", "metric", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/nodes_stats.go#L109-L112
train
olivere/elastic
search_queries_raw_string.go
Source
func (q RawStringQuery) Source() (interface{}, error) { var f interface{} err := json.Unmarshal([]byte(q), &f) return f, err }
go
func (q RawStringQuery) Source() (interface{}, error) { var f interface{} err := json.Unmarshal([]byte(q), &f) return f, err }
[ "func", "(", "q", "RawStringQuery", ")", "Source", "(", ")", "(", "interface", "{", "}", ",", "error", ")", "{", "var", "f", "interface", "{", "}", "\n", "err", ":=", "json", ".", "Unmarshal", "(", "[", "]", "byte", "(", "q", ")", ",", "&", "f", ")", "\n", "return", "f", ",", "err", "\n", "}" ]
// Source returns the JSON encoded body
[ "Source", "returns", "the", "JSON", "encoded", "body" ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/search_queries_raw_string.go#L22-L26
train
olivere/elastic
search_queries_parent_id.go
NewParentIdQuery
func NewParentIdQuery(typ, id string) *ParentIdQuery { return &ParentIdQuery{ typ: typ, id: id, } }
go
func NewParentIdQuery(typ, id string) *ParentIdQuery { return &ParentIdQuery{ typ: typ, id: id, } }
[ "func", "NewParentIdQuery", "(", "typ", ",", "id", "string", ")", "*", "ParentIdQuery", "{", "return", "&", "ParentIdQuery", "{", "typ", ":", "typ", ",", "id", ":", "id", ",", "}", "\n", "}" ]
// NewParentIdQuery creates and initializes a new parent_id query.
[ "NewParentIdQuery", "creates", "and", "initializes", "a", "new", "parent_id", "query", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/search_queries_parent_id.go#L22-L27
train
olivere/elastic
search_queries_parent_id.go
Type
func (q *ParentIdQuery) Type(typ string) *ParentIdQuery { q.typ = typ return q }
go
func (q *ParentIdQuery) Type(typ string) *ParentIdQuery { q.typ = typ return q }
[ "func", "(", "q", "*", "ParentIdQuery", ")", "Type", "(", "typ", "string", ")", "*", "ParentIdQuery", "{", "q", ".", "typ", "=", "typ", "\n", "return", "q", "\n", "}" ]
// Type sets the parent type.
[ "Type", "sets", "the", "parent", "type", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/search_queries_parent_id.go#L30-L33
train
olivere/elastic
search_queries_parent_id.go
Id
func (q *ParentIdQuery) Id(id string) *ParentIdQuery { q.id = id return q }
go
func (q *ParentIdQuery) Id(id string) *ParentIdQuery { q.id = id return q }
[ "func", "(", "q", "*", "ParentIdQuery", ")", "Id", "(", "id", "string", ")", "*", "ParentIdQuery", "{", "q", ".", "id", "=", "id", "\n", "return", "q", "\n", "}" ]
// Id sets the id.
[ "Id", "sets", "the", "id", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/search_queries_parent_id.go#L36-L39
train
olivere/elastic
search_queries_parent_id.go
QueryName
func (q *ParentIdQuery) QueryName(queryName string) *ParentIdQuery { q.queryName = queryName return q }
go
func (q *ParentIdQuery) QueryName(queryName string) *ParentIdQuery { q.queryName = queryName return q }
[ "func", "(", "q", "*", "ParentIdQuery", ")", "QueryName", "(", "queryName", "string", ")", "*", "ParentIdQuery", "{", "q", ".", "queryName", "=", "queryName", "\n", "return", "q", "\n", "}" ]
// QueryName specifies the query name for the filter that can be used when // searching for matched filters per hit.
[ "QueryName", "specifies", "the", "query", "name", "for", "the", "filter", "that", "can", "be", "used", "when", "searching", "for", "matched", "filters", "per", "hit", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/search_queries_parent_id.go#L56-L59
train
olivere/elastic
search_queries_parent_id.go
InnerHit
func (q *ParentIdQuery) InnerHit(innerHit *InnerHit) *ParentIdQuery { q.innerHit = innerHit return q }
go
func (q *ParentIdQuery) InnerHit(innerHit *InnerHit) *ParentIdQuery { q.innerHit = innerHit return q }
[ "func", "(", "q", "*", "ParentIdQuery", ")", "InnerHit", "(", "innerHit", "*", "InnerHit", ")", "*", "ParentIdQuery", "{", "q", ".", "innerHit", "=", "innerHit", "\n", "return", "q", "\n", "}" ]
// InnerHit sets the inner hit definition in the scope of this query and // reusing the defined type and query.
[ "InnerHit", "sets", "the", "inner", "hit", "definition", "in", "the", "scope", "of", "this", "query", "and", "reusing", "the", "defined", "type", "and", "query", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/search_queries_parent_id.go#L63-L66
train
olivere/elastic
search_queries_parent_id.go
Source
func (q *ParentIdQuery) Source() (interface{}, error) { // { // "parent_id" : { // "type" : "blog", // "id" : "1" // } // } source := make(map[string]interface{}) query := make(map[string]interface{}) source["parent_id"] = query query["type"] = q.typ query["id"] = q.id if q.boost != nil { query["boost"] = *q.boost } if q.ignoreUnmapped != nil { query["ignore_unmapped"] = *q.ignoreUnmapped } if q.queryName != "" { query["_name"] = q.queryName } if q.innerHit != nil { src, err := q.innerHit.Source() if err != nil { return nil, err } query["inner_hits"] = src } return source, nil }
go
func (q *ParentIdQuery) Source() (interface{}, error) { // { // "parent_id" : { // "type" : "blog", // "id" : "1" // } // } source := make(map[string]interface{}) query := make(map[string]interface{}) source["parent_id"] = query query["type"] = q.typ query["id"] = q.id if q.boost != nil { query["boost"] = *q.boost } if q.ignoreUnmapped != nil { query["ignore_unmapped"] = *q.ignoreUnmapped } if q.queryName != "" { query["_name"] = q.queryName } if q.innerHit != nil { src, err := q.innerHit.Source() if err != nil { return nil, err } query["inner_hits"] = src } return source, nil }
[ "func", "(", "q", "*", "ParentIdQuery", ")", "Source", "(", ")", "(", "interface", "{", "}", ",", "error", ")", "{", "// {", "// \"parent_id\" : {", "// \"type\" : \"blog\",", "// \"id\" : \"1\"", "// }", "// }", "source", ":=", "make", "(", "map", "[", "string", "]", "interface", "{", "}", ")", "\n", "query", ":=", "make", "(", "map", "[", "string", "]", "interface", "{", "}", ")", "\n", "source", "[", "\"", "\"", "]", "=", "query", "\n\n", "query", "[", "\"", "\"", "]", "=", "q", ".", "typ", "\n", "query", "[", "\"", "\"", "]", "=", "q", ".", "id", "\n", "if", "q", ".", "boost", "!=", "nil", "{", "query", "[", "\"", "\"", "]", "=", "*", "q", ".", "boost", "\n", "}", "\n", "if", "q", ".", "ignoreUnmapped", "!=", "nil", "{", "query", "[", "\"", "\"", "]", "=", "*", "q", ".", "ignoreUnmapped", "\n", "}", "\n", "if", "q", ".", "queryName", "!=", "\"", "\"", "{", "query", "[", "\"", "\"", "]", "=", "q", ".", "queryName", "\n", "}", "\n", "if", "q", ".", "innerHit", "!=", "nil", "{", "src", ",", "err", ":=", "q", ".", "innerHit", ".", "Source", "(", ")", "\n", "if", "err", "!=", "nil", "{", "return", "nil", ",", "err", "\n", "}", "\n", "query", "[", "\"", "\"", "]", "=", "src", "\n", "}", "\n", "return", "source", ",", "nil", "\n", "}" ]
// Source returns JSON for the parent_id query.
[ "Source", "returns", "JSON", "for", "the", "parent_id", "query", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/search_queries_parent_id.go#L69-L99
train
olivere/elastic
tasks_list.go
Actions
func (s *TasksListService) Actions(actions ...string) *TasksListService { s.actions = append(s.actions, actions...) return s }
go
func (s *TasksListService) Actions(actions ...string) *TasksListService { s.actions = append(s.actions, actions...) return s }
[ "func", "(", "s", "*", "TasksListService", ")", "Actions", "(", "actions", "...", "string", ")", "*", "TasksListService", "{", "s", ".", "actions", "=", "append", "(", "s", ".", "actions", ",", "actions", "...", ")", "\n", "return", "s", "\n", "}" ]
// Actions is a list of actions that should be returned. Leave empty to return all.
[ "Actions", "is", "a", "list", "of", "actions", "that", "should", "be", "returned", ".", "Leave", "empty", "to", "return", "all", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/tasks_list.go#L52-L55
train
olivere/elastic
mget.go
Add
func (s *MgetService) Add(items ...*MultiGetItem) *MgetService { s.items = append(s.items, items...) return s }
go
func (s *MgetService) Add(items ...*MultiGetItem) *MgetService { s.items = append(s.items, items...) return s }
[ "func", "(", "s", "*", "MgetService", ")", "Add", "(", "items", "...", "*", "MultiGetItem", ")", "*", "MgetService", "{", "s", ".", "items", "=", "append", "(", "s", ".", "items", ",", "items", "...", ")", "\n", "return", "s", "\n", "}" ]
// Add an item to the request.
[ "Add", "an", "item", "to", "the", "request", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/mget.go#L81-L84
train
olivere/elastic
mget.go
Source
func (s *MgetService) Source() (interface{}, error) { source := make(map[string]interface{}) items := make([]interface{}, len(s.items)) for i, item := range s.items { src, err := item.Source() if err != nil { return nil, err } items[i] = src } source["docs"] = items return source, nil }
go
func (s *MgetService) Source() (interface{}, error) { source := make(map[string]interface{}) items := make([]interface{}, len(s.items)) for i, item := range s.items { src, err := item.Source() if err != nil { return nil, err } items[i] = src } source["docs"] = items return source, nil }
[ "func", "(", "s", "*", "MgetService", ")", "Source", "(", ")", "(", "interface", "{", "}", ",", "error", ")", "{", "source", ":=", "make", "(", "map", "[", "string", "]", "interface", "{", "}", ")", "\n", "items", ":=", "make", "(", "[", "]", "interface", "{", "}", ",", "len", "(", "s", ".", "items", ")", ")", "\n", "for", "i", ",", "item", ":=", "range", "s", ".", "items", "{", "src", ",", "err", ":=", "item", ".", "Source", "(", ")", "\n", "if", "err", "!=", "nil", "{", "return", "nil", ",", "err", "\n", "}", "\n", "items", "[", "i", "]", "=", "src", "\n", "}", "\n", "source", "[", "\"", "\"", "]", "=", "items", "\n", "return", "source", ",", "nil", "\n", "}" ]
// Source returns the request body, which will be serialized into JSON.
[ "Source", "returns", "the", "request", "body", "which", "will", "be", "serialized", "into", "JSON", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/mget.go#L87-L99
train
olivere/elastic
mget.go
Do
func (s *MgetService) Do(ctx context.Context) (*MgetResponse, error) { // Build url path := "/_mget" params := make(url.Values) if s.realtime != nil { params.Add("realtime", fmt.Sprintf("%v", *s.realtime)) } if s.preference != "" { params.Add("preference", s.preference) } if s.refresh != "" { params.Add("refresh", s.refresh) } if s.routing != "" { params.Set("routing", s.routing) } if len(s.storedFields) > 0 { params.Set("stored_fields", strings.Join(s.storedFields, ",")) } // Set body body, err := s.Source() if err != nil { return nil, err } // Get response res, err := s.client.PerformRequest(ctx, PerformRequestOptions{ Method: "GET", Path: path, Params: params, Body: body, }) if err != nil { return nil, err } // Return result ret := new(MgetResponse) if err := s.client.decoder.Decode(res.Body, ret); err != nil { return nil, err } return ret, nil }
go
func (s *MgetService) Do(ctx context.Context) (*MgetResponse, error) { // Build url path := "/_mget" params := make(url.Values) if s.realtime != nil { params.Add("realtime", fmt.Sprintf("%v", *s.realtime)) } if s.preference != "" { params.Add("preference", s.preference) } if s.refresh != "" { params.Add("refresh", s.refresh) } if s.routing != "" { params.Set("routing", s.routing) } if len(s.storedFields) > 0 { params.Set("stored_fields", strings.Join(s.storedFields, ",")) } // Set body body, err := s.Source() if err != nil { return nil, err } // Get response res, err := s.client.PerformRequest(ctx, PerformRequestOptions{ Method: "GET", Path: path, Params: params, Body: body, }) if err != nil { return nil, err } // Return result ret := new(MgetResponse) if err := s.client.decoder.Decode(res.Body, ret); err != nil { return nil, err } return ret, nil }
[ "func", "(", "s", "*", "MgetService", ")", "Do", "(", "ctx", "context", ".", "Context", ")", "(", "*", "MgetResponse", ",", "error", ")", "{", "// Build url", "path", ":=", "\"", "\"", "\n\n", "params", ":=", "make", "(", "url", ".", "Values", ")", "\n", "if", "s", ".", "realtime", "!=", "nil", "{", "params", ".", "Add", "(", "\"", "\"", ",", "fmt", ".", "Sprintf", "(", "\"", "\"", ",", "*", "s", ".", "realtime", ")", ")", "\n", "}", "\n", "if", "s", ".", "preference", "!=", "\"", "\"", "{", "params", ".", "Add", "(", "\"", "\"", ",", "s", ".", "preference", ")", "\n", "}", "\n", "if", "s", ".", "refresh", "!=", "\"", "\"", "{", "params", ".", "Add", "(", "\"", "\"", ",", "s", ".", "refresh", ")", "\n", "}", "\n", "if", "s", ".", "routing", "!=", "\"", "\"", "{", "params", ".", "Set", "(", "\"", "\"", ",", "s", ".", "routing", ")", "\n", "}", "\n", "if", "len", "(", "s", ".", "storedFields", ")", ">", "0", "{", "params", ".", "Set", "(", "\"", "\"", ",", "strings", ".", "Join", "(", "s", ".", "storedFields", ",", "\"", "\"", ")", ")", "\n", "}", "\n\n", "// Set body", "body", ",", "err", ":=", "s", ".", "Source", "(", ")", "\n", "if", "err", "!=", "nil", "{", "return", "nil", ",", "err", "\n", "}", "\n\n", "// Get response", "res", ",", "err", ":=", "s", ".", "client", ".", "PerformRequest", "(", "ctx", ",", "PerformRequestOptions", "{", "Method", ":", "\"", "\"", ",", "Path", ":", "path", ",", "Params", ":", "params", ",", "Body", ":", "body", ",", "}", ")", "\n", "if", "err", "!=", "nil", "{", "return", "nil", ",", "err", "\n", "}", "\n\n", "// Return result", "ret", ":=", "new", "(", "MgetResponse", ")", "\n", "if", "err", ":=", "s", ".", "client", ".", "decoder", ".", "Decode", "(", "res", ".", "Body", ",", "ret", ")", ";", "err", "!=", "nil", "{", "return", "nil", ",", "err", "\n", "}", "\n", "return", "ret", ",", "nil", "\n", "}" ]
// Do executes the request.
[ "Do", "executes", "the", "request", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/mget.go#L102-L146
train
olivere/elastic
mget.go
Index
func (item *MultiGetItem) Index(index string) *MultiGetItem { item.index = index return item }
go
func (item *MultiGetItem) Index(index string) *MultiGetItem { item.index = index return item }
[ "func", "(", "item", "*", "MultiGetItem", ")", "Index", "(", "index", "string", ")", "*", "MultiGetItem", "{", "item", ".", "index", "=", "index", "\n", "return", "item", "\n", "}" ]
// Index specifies the index name.
[ "Index", "specifies", "the", "index", "name", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/mget.go#L168-L171
train
olivere/elastic
mget.go
Type
func (item *MultiGetItem) Type(typ string) *MultiGetItem { item.typ = typ return item }
go
func (item *MultiGetItem) Type(typ string) *MultiGetItem { item.typ = typ return item }
[ "func", "(", "item", "*", "MultiGetItem", ")", "Type", "(", "typ", "string", ")", "*", "MultiGetItem", "{", "item", ".", "typ", "=", "typ", "\n", "return", "item", "\n", "}" ]
// Type specifies the type name.
[ "Type", "specifies", "the", "type", "name", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/mget.go#L174-L177
train
olivere/elastic
mget.go
Id
func (item *MultiGetItem) Id(id string) *MultiGetItem { item.id = id return item }
go
func (item *MultiGetItem) Id(id string) *MultiGetItem { item.id = id return item }
[ "func", "(", "item", "*", "MultiGetItem", ")", "Id", "(", "id", "string", ")", "*", "MultiGetItem", "{", "item", ".", "id", "=", "id", "\n", "return", "item", "\n", "}" ]
// Id specifies the identifier of the document.
[ "Id", "specifies", "the", "identifier", "of", "the", "document", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/mget.go#L180-L183
train
olivere/elastic
mget.go
VersionType
func (item *MultiGetItem) VersionType(versionType string) *MultiGetItem { item.versionType = versionType return item }
go
func (item *MultiGetItem) VersionType(versionType string) *MultiGetItem { item.versionType = versionType return item }
[ "func", "(", "item", "*", "MultiGetItem", ")", "VersionType", "(", "versionType", "string", ")", "*", "MultiGetItem", "{", "item", ".", "versionType", "=", "versionType", "\n", "return", "item", "\n", "}" ]
// VersionType can be "internal", "external", "external_gt", or "external_gte". // See org.elasticsearch.index.VersionType in Elasticsearch source. // It is "internal" by default.
[ "VersionType", "can", "be", "internal", "external", "external_gt", "or", "external_gte", ".", "See", "org", ".", "elasticsearch", ".", "index", ".", "VersionType", "in", "Elasticsearch", "source", ".", "It", "is", "internal", "by", "default", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/mget.go#L208-L211
train
olivere/elastic
mget.go
FetchSource
func (item *MultiGetItem) FetchSource(fetchSourceContext *FetchSourceContext) *MultiGetItem { item.fsc = fetchSourceContext return item }
go
func (item *MultiGetItem) FetchSource(fetchSourceContext *FetchSourceContext) *MultiGetItem { item.fsc = fetchSourceContext return item }
[ "func", "(", "item", "*", "MultiGetItem", ")", "FetchSource", "(", "fetchSourceContext", "*", "FetchSourceContext", ")", "*", "MultiGetItem", "{", "item", ".", "fsc", "=", "fetchSourceContext", "\n", "return", "item", "\n", "}" ]
// FetchSource allows to specify source filtering.
[ "FetchSource", "allows", "to", "specify", "source", "filtering", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/mget.go#L214-L217
train
olivere/elastic
docvalue_field.go
Source
func (d DocvalueField) Source() (interface{}, error) { if d.Format == "" { return d.Field, nil } return map[string]interface{}{ "field": d.Field, "format": d.Format, }, nil }
go
func (d DocvalueField) Source() (interface{}, error) { if d.Format == "" { return d.Field, nil } return map[string]interface{}{ "field": d.Field, "format": d.Format, }, nil }
[ "func", "(", "d", "DocvalueField", ")", "Source", "(", ")", "(", "interface", "{", "}", ",", "error", ")", "{", "if", "d", ".", "Format", "==", "\"", "\"", "{", "return", "d", ".", "Field", ",", "nil", "\n", "}", "\n", "return", "map", "[", "string", "]", "interface", "{", "}", "{", "\"", "\"", ":", "d", ".", "Field", ",", "\"", "\"", ":", "d", ".", "Format", ",", "}", ",", "nil", "\n", "}" ]
// Source serializes the DocvalueField into JSON.
[ "Source", "serializes", "the", "DocvalueField", "into", "JSON", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/docvalue_field.go#L15-L23
train
olivere/elastic
docvalue_field.go
Source
func (d DocvalueFields) Source() (interface{}, error) { if d == nil { return nil, nil } v := make([]interface{}, 0) for _, f := range d { src, err := f.Source() if err != nil { return nil, err } v = append(v, src) } return v, nil }
go
func (d DocvalueFields) Source() (interface{}, error) { if d == nil { return nil, nil } v := make([]interface{}, 0) for _, f := range d { src, err := f.Source() if err != nil { return nil, err } v = append(v, src) } return v, nil }
[ "func", "(", "d", "DocvalueFields", ")", "Source", "(", ")", "(", "interface", "{", "}", ",", "error", ")", "{", "if", "d", "==", "nil", "{", "return", "nil", ",", "nil", "\n", "}", "\n", "v", ":=", "make", "(", "[", "]", "interface", "{", "}", ",", "0", ")", "\n", "for", "_", ",", "f", ":=", "range", "d", "{", "src", ",", "err", ":=", "f", ".", "Source", "(", ")", "\n", "if", "err", "!=", "nil", "{", "return", "nil", ",", "err", "\n", "}", "\n", "v", "=", "append", "(", "v", ",", "src", ")", "\n", "}", "\n", "return", "v", ",", "nil", "\n", "}" ]
// Source serializes the DocvalueFields into JSON.
[ "Source", "serializes", "the", "DocvalueFields", "into", "JSON", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/docvalue_field.go#L29-L42
train
olivere/elastic
indices_forcemerge.go
NewIndicesForcemergeService
func NewIndicesForcemergeService(client *Client) *IndicesForcemergeService { return &IndicesForcemergeService{ client: client, index: make([]string, 0), } }
go
func NewIndicesForcemergeService(client *Client) *IndicesForcemergeService { return &IndicesForcemergeService{ client: client, index: make([]string, 0), } }
[ "func", "NewIndicesForcemergeService", "(", "client", "*", "Client", ")", "*", "IndicesForcemergeService", "{", "return", "&", "IndicesForcemergeService", "{", "client", ":", "client", ",", "index", ":", "make", "(", "[", "]", "string", ",", "0", ")", ",", "}", "\n", "}" ]
// NewIndicesForcemergeService creates a new IndicesForcemergeService.
[ "NewIndicesForcemergeService", "creates", "a", "new", "IndicesForcemergeService", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/indices_forcemerge.go#L36-L41
train
olivere/elastic
indices_forcemerge.go
OnlyExpungeDeletes
func (s *IndicesForcemergeService) OnlyExpungeDeletes(onlyExpungeDeletes bool) *IndicesForcemergeService { s.onlyExpungeDeletes = &onlyExpungeDeletes return s }
go
func (s *IndicesForcemergeService) OnlyExpungeDeletes(onlyExpungeDeletes bool) *IndicesForcemergeService { s.onlyExpungeDeletes = &onlyExpungeDeletes return s }
[ "func", "(", "s", "*", "IndicesForcemergeService", ")", "OnlyExpungeDeletes", "(", "onlyExpungeDeletes", "bool", ")", "*", "IndicesForcemergeService", "{", "s", ".", "onlyExpungeDeletes", "=", "&", "onlyExpungeDeletes", "\n", "return", "s", "\n", "}" ]
// OnlyExpungeDeletes specifies whether the operation should only expunge // deleted documents.
[ "OnlyExpungeDeletes", "specifies", "whether", "the", "operation", "should", "only", "expunge", "deleted", "documents", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/indices_forcemerge.go#L91-L94
train
olivere/elastic
search_queries_wildcard.go
NewWildcardQuery
func NewWildcardQuery(name, wildcard string) *WildcardQuery { return &WildcardQuery{ name: name, wildcard: wildcard, } }
go
func NewWildcardQuery(name, wildcard string) *WildcardQuery { return &WildcardQuery{ name: name, wildcard: wildcard, } }
[ "func", "NewWildcardQuery", "(", "name", ",", "wildcard", "string", ")", "*", "WildcardQuery", "{", "return", "&", "WildcardQuery", "{", "name", ":", "name", ",", "wildcard", ":", "wildcard", ",", "}", "\n", "}" ]
// NewWildcardQuery creates and initializes a new WildcardQuery.
[ "NewWildcardQuery", "creates", "and", "initializes", "a", "new", "WildcardQuery", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/search_queries_wildcard.go#L26-L31
train
olivere/elastic
tasks_cancel.go
Actions
func (s *TasksCancelService) Actions(actions ...string) *TasksCancelService { s.actions = append(s.actions, actions...) return s }
go
func (s *TasksCancelService) Actions(actions ...string) *TasksCancelService { s.actions = append(s.actions, actions...) return s }
[ "func", "(", "s", "*", "TasksCancelService", ")", "Actions", "(", "actions", "...", "string", ")", "*", "TasksCancelService", "{", "s", ".", "actions", "=", "append", "(", "s", ".", "actions", ",", "actions", "...", ")", "\n", "return", "s", "\n", "}" ]
// Actions is a list of actions that should be cancelled. Leave empty to cancel all.
[ "Actions", "is", "a", "list", "of", "actions", "that", "should", "be", "cancelled", ".", "Leave", "empty", "to", "cancel", "all", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/tasks_cancel.go#L54-L57
train
olivere/elastic
tasks_cancel.go
ParentTaskIdFromNodeAndId
func (s *TasksCancelService) ParentTaskIdFromNodeAndId(nodeId string, id int64) *TasksCancelService { if id != -1 { s.parentTaskId = fmt.Sprintf("%s:%d", nodeId, id) } return s }
go
func (s *TasksCancelService) ParentTaskIdFromNodeAndId(nodeId string, id int64) *TasksCancelService { if id != -1 { s.parentTaskId = fmt.Sprintf("%s:%d", nodeId, id) } return s }
[ "func", "(", "s", "*", "TasksCancelService", ")", "ParentTaskIdFromNodeAndId", "(", "nodeId", "string", ",", "id", "int64", ")", "*", "TasksCancelService", "{", "if", "id", "!=", "-", "1", "{", "s", ".", "parentTaskId", "=", "fmt", ".", "Sprintf", "(", "\"", "\"", ",", "nodeId", ",", "id", ")", "\n", "}", "\n", "return", "s", "\n", "}" ]
// ParentTaskIdFromNodeAndId specifies to cancel tasks with specified parent task id.
[ "ParentTaskIdFromNodeAndId", "specifies", "to", "cancel", "tasks", "with", "specified", "parent", "task", "id", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/tasks_cancel.go#L76-L81
train
olivere/elastic
search_source.go
Query
func (s *SearchSource) Query(query Query) *SearchSource { s.query = query return s }
go
func (s *SearchSource) Query(query Query) *SearchSource { s.query = query return s }
[ "func", "(", "s", "*", "SearchSource", ")", "Query", "(", "query", "Query", ")", "*", "SearchSource", "{", "s", ".", "query", "=", "query", "\n", "return", "s", "\n", "}" ]
// Query sets the query to use with this search source.
[ "Query", "sets", "the", "query", "to", "use", "with", "this", "search", "source", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/search_source.go#L58-L61
train
olivere/elastic
search_source.go
TimeoutInMillis
func (s *SearchSource) TimeoutInMillis(timeoutInMillis int) *SearchSource { s.timeout = fmt.Sprintf("%dms", timeoutInMillis) return s }
go
func (s *SearchSource) TimeoutInMillis(timeoutInMillis int) *SearchSource { s.timeout = fmt.Sprintf("%dms", timeoutInMillis) return s }
[ "func", "(", "s", "*", "SearchSource", ")", "TimeoutInMillis", "(", "timeoutInMillis", "int", ")", "*", "SearchSource", "{", "s", ".", "timeout", "=", "fmt", ".", "Sprintf", "(", "\"", "\"", ",", "timeoutInMillis", ")", "\n", "return", "s", "\n", "}" ]
// TimeoutInMillis controls how many milliseconds a search is allowed // to take before it is canceled.
[ "TimeoutInMillis", "controls", "how", "many", "milliseconds", "a", "search", "is", "allowed", "to", "take", "before", "it", "is", "canceled", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/search_source.go#L128-L131
train
olivere/elastic
search_source.go
DefaultRescoreWindowSize
func (s *SearchSource) DefaultRescoreWindowSize(defaultRescoreWindowSize int) *SearchSource { s.defaultRescoreWindowSize = &defaultRescoreWindowSize return s }
go
func (s *SearchSource) DefaultRescoreWindowSize(defaultRescoreWindowSize int) *SearchSource { s.defaultRescoreWindowSize = &defaultRescoreWindowSize return s }
[ "func", "(", "s", "*", "SearchSource", ")", "DefaultRescoreWindowSize", "(", "defaultRescoreWindowSize", "int", ")", "*", "SearchSource", "{", "s", ".", "defaultRescoreWindowSize", "=", "&", "defaultRescoreWindowSize", "\n", "return", "s", "\n", "}" ]
// DefaultRescoreWindowSize sets the rescore window size for rescores // that don't specify their window.
[ "DefaultRescoreWindowSize", "sets", "the", "rescore", "window", "size", "for", "rescores", "that", "don", "t", "specify", "their", "window", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/search_source.go#L196-L199
train
olivere/elastic
search_source.go
Highlighter
func (s *SearchSource) Highlighter() *Highlight { if s.highlight == nil { s.highlight = NewHighlight() } return s.highlight }
go
func (s *SearchSource) Highlighter() *Highlight { if s.highlight == nil { s.highlight = NewHighlight() } return s.highlight }
[ "func", "(", "s", "*", "SearchSource", ")", "Highlighter", "(", ")", "*", "Highlight", "{", "if", "s", ".", "highlight", "==", "nil", "{", "s", ".", "highlight", "=", "NewHighlight", "(", ")", "\n", "}", "\n", "return", "s", ".", "highlight", "\n", "}" ]
// Highlighter returns the highlighter.
[ "Highlighter", "returns", "the", "highlighter", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/search_source.go#L208-L213
train
olivere/elastic
search_source.go
DocvalueField
func (s *SearchSource) DocvalueField(fieldDataField string) *SearchSource { s.docvalueFields = append(s.docvalueFields, DocvalueField{Field: fieldDataField}) return s }
go
func (s *SearchSource) DocvalueField(fieldDataField string) *SearchSource { s.docvalueFields = append(s.docvalueFields, DocvalueField{Field: fieldDataField}) return s }
[ "func", "(", "s", "*", "SearchSource", ")", "DocvalueField", "(", "fieldDataField", "string", ")", "*", "SearchSource", "{", "s", ".", "docvalueFields", "=", "append", "(", "s", ".", "docvalueFields", ",", "DocvalueField", "{", "Field", ":", "fieldDataField", "}", ")", "\n", "return", "s", "\n", "}" ]
// DocvalueField adds a single field to load from the field data cache // and return as part of the search request.
[ "DocvalueField", "adds", "a", "single", "field", "to", "load", "from", "the", "field", "data", "cache", "and", "return", "as", "part", "of", "the", "search", "request", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/search_source.go#L292-L295
train
olivere/elastic
search_source.go
DocvalueFieldWithFormat
func (s *SearchSource) DocvalueFieldWithFormat(fieldDataFieldWithFormat DocvalueField) *SearchSource { s.docvalueFields = append(s.docvalueFields, fieldDataFieldWithFormat) return s }
go
func (s *SearchSource) DocvalueFieldWithFormat(fieldDataFieldWithFormat DocvalueField) *SearchSource { s.docvalueFields = append(s.docvalueFields, fieldDataFieldWithFormat) return s }
[ "func", "(", "s", "*", "SearchSource", ")", "DocvalueFieldWithFormat", "(", "fieldDataFieldWithFormat", "DocvalueField", ")", "*", "SearchSource", "{", "s", ".", "docvalueFields", "=", "append", "(", "s", ".", "docvalueFields", ",", "fieldDataFieldWithFormat", ")", "\n", "return", "s", "\n", "}" ]
// DocvalueField adds a single docvalue field to load from the field data cache // and return as part of the search request.
[ "DocvalueField", "adds", "a", "single", "docvalue", "field", "to", "load", "from", "the", "field", "data", "cache", "and", "return", "as", "part", "of", "the", "search", "request", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/search_source.go#L299-L302
train
olivere/elastic
search_source.go
DocvalueFields
func (s *SearchSource) DocvalueFields(docvalueFields ...string) *SearchSource { for _, f := range docvalueFields { s.docvalueFields = append(s.docvalueFields, DocvalueField{Field: f}) } return s }
go
func (s *SearchSource) DocvalueFields(docvalueFields ...string) *SearchSource { for _, f := range docvalueFields { s.docvalueFields = append(s.docvalueFields, DocvalueField{Field: f}) } return s }
[ "func", "(", "s", "*", "SearchSource", ")", "DocvalueFields", "(", "docvalueFields", "...", "string", ")", "*", "SearchSource", "{", "for", "_", ",", "f", ":=", "range", "docvalueFields", "{", "s", ".", "docvalueFields", "=", "append", "(", "s", ".", "docvalueFields", ",", "DocvalueField", "{", "Field", ":", "f", "}", ")", "\n", "}", "\n", "return", "s", "\n", "}" ]
// DocvalueFields adds one or more fields to load from the field data cache // and return as part of the search request.
[ "DocvalueFields", "adds", "one", "or", "more", "fields", "to", "load", "from", "the", "field", "data", "cache", "and", "return", "as", "part", "of", "the", "search", "request", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/search_source.go#L306-L311
train
olivere/elastic
search_source.go
DocvalueFieldsWithFormat
func (s *SearchSource) DocvalueFieldsWithFormat(docvalueFields ...DocvalueField) *SearchSource { s.docvalueFields = append(s.docvalueFields, docvalueFields...) return s }
go
func (s *SearchSource) DocvalueFieldsWithFormat(docvalueFields ...DocvalueField) *SearchSource { s.docvalueFields = append(s.docvalueFields, docvalueFields...) return s }
[ "func", "(", "s", "*", "SearchSource", ")", "DocvalueFieldsWithFormat", "(", "docvalueFields", "...", "DocvalueField", ")", "*", "SearchSource", "{", "s", ".", "docvalueFields", "=", "append", "(", "s", ".", "docvalueFields", ",", "docvalueFields", "...", ")", "\n", "return", "s", "\n", "}" ]
// DocvalueFields adds one or more docvalue fields to load from the field data cache // and return as part of the search request.
[ "DocvalueFields", "adds", "one", "or", "more", "docvalue", "fields", "to", "load", "from", "the", "field", "data", "cache", "and", "return", "as", "part", "of", "the", "search", "request", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/search_source.go#L315-L318
train
olivere/elastic
search_source.go
ScriptField
func (s *SearchSource) ScriptField(scriptField *ScriptField) *SearchSource { s.scriptFields = append(s.scriptFields, scriptField) return s }
go
func (s *SearchSource) ScriptField(scriptField *ScriptField) *SearchSource { s.scriptFields = append(s.scriptFields, scriptField) return s }
[ "func", "(", "s", "*", "SearchSource", ")", "ScriptField", "(", "scriptField", "*", "ScriptField", ")", "*", "SearchSource", "{", "s", ".", "scriptFields", "=", "append", "(", "s", ".", "scriptFields", ",", "scriptField", ")", "\n", "return", "s", "\n", "}" ]
// ScriptField adds a single script field with the provided script.
[ "ScriptField", "adds", "a", "single", "script", "field", "with", "the", "provided", "script", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/search_source.go#L321-L324
train
olivere/elastic
search_source.go
ScriptFields
func (s *SearchSource) ScriptFields(scriptFields ...*ScriptField) *SearchSource { s.scriptFields = append(s.scriptFields, scriptFields...) return s }
go
func (s *SearchSource) ScriptFields(scriptFields ...*ScriptField) *SearchSource { s.scriptFields = append(s.scriptFields, scriptFields...) return s }
[ "func", "(", "s", "*", "SearchSource", ")", "ScriptFields", "(", "scriptFields", "...", "*", "ScriptField", ")", "*", "SearchSource", "{", "s", ".", "scriptFields", "=", "append", "(", "s", ".", "scriptFields", ",", "scriptFields", "...", ")", "\n", "return", "s", "\n", "}" ]
// ScriptFields adds one or more script fields with the provided scripts.
[ "ScriptFields", "adds", "one", "or", "more", "script", "fields", "with", "the", "provided", "scripts", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/search_source.go#L327-L330
train
olivere/elastic
search_source.go
IndexBoost
func (s *SearchSource) IndexBoost(index string, boost float64) *SearchSource { s.indexBoosts[index] = boost return s }
go
func (s *SearchSource) IndexBoost(index string, boost float64) *SearchSource { s.indexBoosts[index] = boost return s }
[ "func", "(", "s", "*", "SearchSource", ")", "IndexBoost", "(", "index", "string", ",", "boost", "float64", ")", "*", "SearchSource", "{", "s", ".", "indexBoosts", "[", "index", "]", "=", "boost", "\n", "return", "s", "\n", "}" ]
// IndexBoost sets the boost that a specific index will receive when the // query is executed against it.
[ "IndexBoost", "sets", "the", "boost", "that", "a", "specific", "index", "will", "receive", "when", "the", "query", "is", "executed", "against", "it", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/search_source.go#L334-L337
train
olivere/elastic
search_source.go
Stats
func (s *SearchSource) Stats(statsGroup ...string) *SearchSource { s.stats = append(s.stats, statsGroup...) return s }
go
func (s *SearchSource) Stats(statsGroup ...string) *SearchSource { s.stats = append(s.stats, statsGroup...) return s }
[ "func", "(", "s", "*", "SearchSource", ")", "Stats", "(", "statsGroup", "...", "string", ")", "*", "SearchSource", "{", "s", ".", "stats", "=", "append", "(", "s", ".", "stats", ",", "statsGroup", "...", ")", "\n", "return", "s", "\n", "}" ]
// Stats group this request will be aggregated under.
[ "Stats", "group", "this", "request", "will", "be", "aggregated", "under", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/search_source.go#L340-L343
train
olivere/elastic
search_source.go
InnerHit
func (s *SearchSource) InnerHit(name string, innerHit *InnerHit) *SearchSource { s.innerHits[name] = innerHit return s }
go
func (s *SearchSource) InnerHit(name string, innerHit *InnerHit) *SearchSource { s.innerHits[name] = innerHit return s }
[ "func", "(", "s", "*", "SearchSource", ")", "InnerHit", "(", "name", "string", ",", "innerHit", "*", "InnerHit", ")", "*", "SearchSource", "{", "s", ".", "innerHits", "[", "name", "]", "=", "innerHit", "\n", "return", "s", "\n", "}" ]
// InnerHit adds an inner hit to return with the result.
[ "InnerHit", "adds", "an", "inner", "hit", "to", "return", "with", "the", "result", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/search_source.go#L346-L349
train
olivere/elastic
indices_analyze.go
NewIndicesAnalyzeService
func NewIndicesAnalyzeService(client *Client) *IndicesAnalyzeService { return &IndicesAnalyzeService{ client: client, request: new(IndicesAnalyzeRequest), } }
go
func NewIndicesAnalyzeService(client *Client) *IndicesAnalyzeService { return &IndicesAnalyzeService{ client: client, request: new(IndicesAnalyzeRequest), } }
[ "func", "NewIndicesAnalyzeService", "(", "client", "*", "Client", ")", "*", "IndicesAnalyzeService", "{", "return", "&", "IndicesAnalyzeService", "{", "client", ":", "client", ",", "request", ":", "new", "(", "IndicesAnalyzeRequest", ")", ",", "}", "\n", "}" ]
// NewIndicesAnalyzeService creates a new IndicesAnalyzeService.
[ "NewIndicesAnalyzeService", "creates", "a", "new", "IndicesAnalyzeService", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/indices_analyze.go#L32-L37
train
olivere/elastic
indices_analyze.go
Index
func (s *IndicesAnalyzeService) Index(index string) *IndicesAnalyzeService { s.index = index return s }
go
func (s *IndicesAnalyzeService) Index(index string) *IndicesAnalyzeService { s.index = index return s }
[ "func", "(", "s", "*", "IndicesAnalyzeService", ")", "Index", "(", "index", "string", ")", "*", "IndicesAnalyzeService", "{", "s", ".", "index", "=", "index", "\n", "return", "s", "\n", "}" ]
// Index is the name of the index to scope the operation.
[ "Index", "is", "the", "name", "of", "the", "index", "to", "scope", "the", "operation", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/indices_analyze.go#L40-L43
train
olivere/elastic
indices_analyze.go
Format
func (s *IndicesAnalyzeService) Format(format string) *IndicesAnalyzeService { s.format = format return s }
go
func (s *IndicesAnalyzeService) Format(format string) *IndicesAnalyzeService { s.format = format return s }
[ "func", "(", "s", "*", "IndicesAnalyzeService", ")", "Format", "(", "format", "string", ")", "*", "IndicesAnalyzeService", "{", "s", ".", "format", "=", "format", "\n", "return", "s", "\n", "}" ]
// Format of the output.
[ "Format", "of", "the", "output", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/indices_analyze.go#L46-L49
train
olivere/elastic
indices_analyze.go
Request
func (s *IndicesAnalyzeService) Request(request *IndicesAnalyzeRequest) *IndicesAnalyzeService { if request == nil { s.request = new(IndicesAnalyzeRequest) } else { s.request = request } return s }
go
func (s *IndicesAnalyzeService) Request(request *IndicesAnalyzeRequest) *IndicesAnalyzeService { if request == nil { s.request = new(IndicesAnalyzeRequest) } else { s.request = request } return s }
[ "func", "(", "s", "*", "IndicesAnalyzeService", ")", "Request", "(", "request", "*", "IndicesAnalyzeRequest", ")", "*", "IndicesAnalyzeService", "{", "if", "request", "==", "nil", "{", "s", ".", "request", "=", "new", "(", "IndicesAnalyzeRequest", ")", "\n", "}", "else", "{", "s", ".", "request", "=", "request", "\n", "}", "\n", "return", "s", "\n", "}" ]
// Request passes the analyze request to use.
[ "Request", "passes", "the", "analyze", "request", "to", "use", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/indices_analyze.go#L59-L66
train
olivere/elastic
indices_analyze.go
Analyzer
func (s *IndicesAnalyzeService) Analyzer(analyzer string) *IndicesAnalyzeService { s.request.Analyzer = analyzer return s }
go
func (s *IndicesAnalyzeService) Analyzer(analyzer string) *IndicesAnalyzeService { s.request.Analyzer = analyzer return s }
[ "func", "(", "s", "*", "IndicesAnalyzeService", ")", "Analyzer", "(", "analyzer", "string", ")", "*", "IndicesAnalyzeService", "{", "s", ".", "request", ".", "Analyzer", "=", "analyzer", "\n", "return", "s", "\n", "}" ]
// Analyzer is the name of the analyzer to use.
[ "Analyzer", "is", "the", "name", "of", "the", "analyzer", "to", "use", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/indices_analyze.go#L69-L72
train
olivere/elastic
indices_analyze.go
Attributes
func (s *IndicesAnalyzeService) Attributes(attributes ...string) *IndicesAnalyzeService { s.request.Attributes = attributes return s }
go
func (s *IndicesAnalyzeService) Attributes(attributes ...string) *IndicesAnalyzeService { s.request.Attributes = attributes return s }
[ "func", "(", "s", "*", "IndicesAnalyzeService", ")", "Attributes", "(", "attributes", "...", "string", ")", "*", "IndicesAnalyzeService", "{", "s", ".", "request", ".", "Attributes", "=", "attributes", "\n", "return", "s", "\n", "}" ]
// Attributes is a list of token attributes to output; this parameter works // only with explain=true.
[ "Attributes", "is", "a", "list", "of", "token", "attributes", "to", "output", ";", "this", "parameter", "works", "only", "with", "explain", "=", "true", "." ]
0534a7b1bf47b1ccf57e905491a641709f8a623d
https://github.com/olivere/elastic/blob/0534a7b1bf47b1ccf57e905491a641709f8a623d/indices_analyze.go#L76-L79
train