repo
stringlengths
6
47
file_url
stringlengths
77
269
file_path
stringlengths
5
186
content
stringlengths
0
32.8k
language
stringclasses
1 value
license
stringclasses
7 values
commit_sha
stringlengths
40
40
retrieved_at
stringdate
2026-01-07 08:35:43
2026-01-07 08:55:24
truncated
bool
2 classes
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/gogo/protobuf/proto/encode.go
vendor/github.com/gogo/protobuf/proto/encode.go
// Go support for Protocol Buffers - Google's data interchange format // // Copyright 2010 The Go Authors. All rights reserved. // https://github.com/golang/protobuf // // Redistribution and use in source and binary forms, with or without // modification, are permitted provided that the following conditions are // met: // // * Redistributions of source code must retain the above copyright // notice, this list of conditions and the following disclaimer. // * Redistributions in binary form must reproduce the above // copyright notice, this list of conditions and the following disclaimer // in the documentation and/or other materials provided with the // distribution. // * Neither the name of Google Inc. nor the names of its // contributors may be used to endorse or promote products derived from // this software without specific prior written permission. // // THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS // "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT // LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR // A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT // OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, // SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT // LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, // DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY // THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT // (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE // OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. package proto /* * Routines for encoding data into the wire format for protocol buffers. */ import ( "errors" "reflect" ) var ( // errRepeatedHasNil is the error returned if Marshal is called with // a struct with a repeated field containing a nil element. errRepeatedHasNil = errors.New("proto: repeated field has nil element") // errOneofHasNil is the error returned if Marshal is called with // a struct with a oneof field containing a nil element. errOneofHasNil = errors.New("proto: oneof field has nil value") // ErrNil is the error returned if Marshal is called with nil. ErrNil = errors.New("proto: Marshal called with nil") // ErrTooLarge is the error returned if Marshal is called with a // message that encodes to >2GB. ErrTooLarge = errors.New("proto: message encodes to over 2 GB") ) // The fundamental encoders that put bytes on the wire. // Those that take integer types all accept uint64 and are // therefore of type valueEncoder. const maxVarintBytes = 10 // maximum length of a varint // EncodeVarint returns the varint encoding of x. // This is the format for the // int32, int64, uint32, uint64, bool, and enum // protocol buffer types. // Not used by the package itself, but helpful to clients // wishing to use the same encoding. func EncodeVarint(x uint64) []byte { var buf [maxVarintBytes]byte var n int for n = 0; x > 127; n++ { buf[n] = 0x80 | uint8(x&0x7F) x >>= 7 } buf[n] = uint8(x) n++ return buf[0:n] } // EncodeVarint writes a varint-encoded integer to the Buffer. // This is the format for the // int32, int64, uint32, uint64, bool, and enum // protocol buffer types. func (p *Buffer) EncodeVarint(x uint64) error { for x >= 1<<7 { p.buf = append(p.buf, uint8(x&0x7f|0x80)) x >>= 7 } p.buf = append(p.buf, uint8(x)) return nil } // SizeVarint returns the varint encoding size of an integer. func SizeVarint(x uint64) int { switch { case x < 1<<7: return 1 case x < 1<<14: return 2 case x < 1<<21: return 3 case x < 1<<28: return 4 case x < 1<<35: return 5 case x < 1<<42: return 6 case x < 1<<49: return 7 case x < 1<<56: return 8 case x < 1<<63: return 9 } return 10 } // EncodeFixed64 writes a 64-bit integer to the Buffer. // This is the format for the // fixed64, sfixed64, and double protocol buffer types. func (p *Buffer) EncodeFixed64(x uint64) error { p.buf = append(p.buf, uint8(x), uint8(x>>8), uint8(x>>16), uint8(x>>24), uint8(x>>32), uint8(x>>40), uint8(x>>48), uint8(x>>56)) return nil } // EncodeFixed32 writes a 32-bit integer to the Buffer. // This is the format for the // fixed32, sfixed32, and float protocol buffer types. func (p *Buffer) EncodeFixed32(x uint64) error { p.buf = append(p.buf, uint8(x), uint8(x>>8), uint8(x>>16), uint8(x>>24)) return nil } // EncodeZigzag64 writes a zigzag-encoded 64-bit integer // to the Buffer. // This is the format used for the sint64 protocol buffer type. func (p *Buffer) EncodeZigzag64(x uint64) error { // use signed number to get arithmetic right shift. return p.EncodeVarint(uint64((x << 1) ^ uint64((int64(x) >> 63)))) } // EncodeZigzag32 writes a zigzag-encoded 32-bit integer // to the Buffer. // This is the format used for the sint32 protocol buffer type. func (p *Buffer) EncodeZigzag32(x uint64) error { // use signed number to get arithmetic right shift. return p.EncodeVarint(uint64((uint32(x) << 1) ^ uint32((int32(x) >> 31)))) } // EncodeRawBytes writes a count-delimited byte buffer to the Buffer. // This is the format used for the bytes protocol buffer // type and for embedded messages. func (p *Buffer) EncodeRawBytes(b []byte) error { p.EncodeVarint(uint64(len(b))) p.buf = append(p.buf, b...) return nil } // EncodeStringBytes writes an encoded string to the Buffer. // This is the format used for the proto2 string type. func (p *Buffer) EncodeStringBytes(s string) error { p.EncodeVarint(uint64(len(s))) p.buf = append(p.buf, s...) return nil } // Marshaler is the interface representing objects that can marshal themselves. type Marshaler interface { Marshal() ([]byte, error) } // EncodeMessage writes the protocol buffer to the Buffer, // prefixed by a varint-encoded length. func (p *Buffer) EncodeMessage(pb Message) error { siz := Size(pb) sizVar := SizeVarint(uint64(siz)) p.grow(siz + sizVar) p.EncodeVarint(uint64(siz)) return p.Marshal(pb) } // All protocol buffer fields are nillable, but be careful. func isNil(v reflect.Value) bool { switch v.Kind() { case reflect.Interface, reflect.Map, reflect.Ptr, reflect.Slice: return v.IsNil() } return false }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/gogo/protobuf/proto/extensions.go
vendor/github.com/gogo/protobuf/proto/extensions.go
// Go support for Protocol Buffers - Google's data interchange format // // Copyright 2010 The Go Authors. All rights reserved. // https://github.com/golang/protobuf // // Redistribution and use in source and binary forms, with or without // modification, are permitted provided that the following conditions are // met: // // * Redistributions of source code must retain the above copyright // notice, this list of conditions and the following disclaimer. // * Redistributions in binary form must reproduce the above // copyright notice, this list of conditions and the following disclaimer // in the documentation and/or other materials provided with the // distribution. // * Neither the name of Google Inc. nor the names of its // contributors may be used to endorse or promote products derived from // this software without specific prior written permission. // // THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS // "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT // LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR // A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT // OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, // SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT // LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, // DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY // THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT // (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE // OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. package proto /* * Types and routines for supporting protocol buffer extensions. */ import ( "errors" "fmt" "io" "reflect" "strconv" "sync" ) // ErrMissingExtension is the error returned by GetExtension if the named extension is not in the message. var ErrMissingExtension = errors.New("proto: missing extension") // ExtensionRange represents a range of message extensions for a protocol buffer. // Used in code generated by the protocol compiler. type ExtensionRange struct { Start, End int32 // both inclusive } // extendableProto is an interface implemented by any protocol buffer generated by the current // proto compiler that may be extended. type extendableProto interface { Message ExtensionRangeArray() []ExtensionRange extensionsWrite() map[int32]Extension extensionsRead() (map[int32]Extension, sync.Locker) } // extendableProtoV1 is an interface implemented by a protocol buffer generated by the previous // version of the proto compiler that may be extended. type extendableProtoV1 interface { Message ExtensionRangeArray() []ExtensionRange ExtensionMap() map[int32]Extension } // extensionAdapter is a wrapper around extendableProtoV1 that implements extendableProto. type extensionAdapter struct { extendableProtoV1 } func (e extensionAdapter) extensionsWrite() map[int32]Extension { return e.ExtensionMap() } func (e extensionAdapter) extensionsRead() (map[int32]Extension, sync.Locker) { return e.ExtensionMap(), notLocker{} } // notLocker is a sync.Locker whose Lock and Unlock methods are nops. type notLocker struct{} func (n notLocker) Lock() {} func (n notLocker) Unlock() {} // extendable returns the extendableProto interface for the given generated proto message. // If the proto message has the old extension format, it returns a wrapper that implements // the extendableProto interface. func extendable(p interface{}) (extendableProto, error) { switch p := p.(type) { case extendableProto: if isNilPtr(p) { return nil, fmt.Errorf("proto: nil %T is not extendable", p) } return p, nil case extendableProtoV1: if isNilPtr(p) { return nil, fmt.Errorf("proto: nil %T is not extendable", p) } return extensionAdapter{p}, nil case extensionsBytes: return slowExtensionAdapter{p}, nil } // Don't allocate a specific error containing %T: // this is the hot path for Clone and MarshalText. return nil, errNotExtendable } var errNotExtendable = errors.New("proto: not an extendable proto.Message") func isNilPtr(x interface{}) bool { v := reflect.ValueOf(x) return v.Kind() == reflect.Ptr && v.IsNil() } // XXX_InternalExtensions is an internal representation of proto extensions. // // Each generated message struct type embeds an anonymous XXX_InternalExtensions field, // thus gaining the unexported 'extensions' method, which can be called only from the proto package. // // The methods of XXX_InternalExtensions are not concurrency safe in general, // but calls to logically read-only methods such as has and get may be executed concurrently. type XXX_InternalExtensions struct { // The struct must be indirect so that if a user inadvertently copies a // generated message and its embedded XXX_InternalExtensions, they // avoid the mayhem of a copied mutex. // // The mutex serializes all logically read-only operations to p.extensionMap. // It is up to the client to ensure that write operations to p.extensionMap are // mutually exclusive with other accesses. p *struct { mu sync.Mutex extensionMap map[int32]Extension } } // extensionsWrite returns the extension map, creating it on first use. func (e *XXX_InternalExtensions) extensionsWrite() map[int32]Extension { if e.p == nil { e.p = new(struct { mu sync.Mutex extensionMap map[int32]Extension }) e.p.extensionMap = make(map[int32]Extension) } return e.p.extensionMap } // extensionsRead returns the extensions map for read-only use. It may be nil. // The caller must hold the returned mutex's lock when accessing Elements within the map. func (e *XXX_InternalExtensions) extensionsRead() (map[int32]Extension, sync.Locker) { if e.p == nil { return nil, nil } return e.p.extensionMap, &e.p.mu } // ExtensionDesc represents an extension specification. // Used in generated code from the protocol compiler. type ExtensionDesc struct { ExtendedType Message // nil pointer to the type that is being extended ExtensionType interface{} // nil pointer to the extension type Field int32 // field number Name string // fully-qualified name of extension, for text formatting Tag string // protobuf tag style Filename string // name of the file in which the extension is defined } func (ed *ExtensionDesc) repeated() bool { t := reflect.TypeOf(ed.ExtensionType) return t.Kind() == reflect.Slice && t.Elem().Kind() != reflect.Uint8 } // Extension represents an extension in a message. type Extension struct { // When an extension is stored in a message using SetExtension // only desc and value are set. When the message is marshaled // enc will be set to the encoded form of the message. // // When a message is unmarshaled and contains extensions, each // extension will have only enc set. When such an extension is // accessed using GetExtension (or GetExtensions) desc and value // will be set. desc *ExtensionDesc value interface{} enc []byte } // SetRawExtension is for testing only. func SetRawExtension(base Message, id int32, b []byte) { if ebase, ok := base.(extensionsBytes); ok { clearExtension(base, id) ext := ebase.GetExtensions() *ext = append(*ext, b...) return } epb, err := extendable(base) if err != nil { return } extmap := epb.extensionsWrite() extmap[id] = Extension{enc: b} } // isExtensionField returns true iff the given field number is in an extension range. func isExtensionField(pb extendableProto, field int32) bool { for _, er := range pb.ExtensionRangeArray() { if er.Start <= field && field <= er.End { return true } } return false } // checkExtensionTypes checks that the given extension is valid for pb. func checkExtensionTypes(pb extendableProto, extension *ExtensionDesc) error { var pbi interface{} = pb // Check the extended type. if ea, ok := pbi.(extensionAdapter); ok { pbi = ea.extendableProtoV1 } if ea, ok := pbi.(slowExtensionAdapter); ok { pbi = ea.extensionsBytes } if a, b := reflect.TypeOf(pbi), reflect.TypeOf(extension.ExtendedType); a != b { return fmt.Errorf("proto: bad extended type; %v does not extend %v", b, a) } // Check the range. if !isExtensionField(pb, extension.Field) { return errors.New("proto: bad extension number; not in declared ranges") } return nil } // extPropKey is sufficient to uniquely identify an extension. type extPropKey struct { base reflect.Type field int32 } var extProp = struct { sync.RWMutex m map[extPropKey]*Properties }{ m: make(map[extPropKey]*Properties), } func extensionProperties(ed *ExtensionDesc) *Properties { key := extPropKey{base: reflect.TypeOf(ed.ExtendedType), field: ed.Field} extProp.RLock() if prop, ok := extProp.m[key]; ok { extProp.RUnlock() return prop } extProp.RUnlock() extProp.Lock() defer extProp.Unlock() // Check again. if prop, ok := extProp.m[key]; ok { return prop } prop := new(Properties) prop.Init(reflect.TypeOf(ed.ExtensionType), "unknown_name", ed.Tag, nil) extProp.m[key] = prop return prop } // HasExtension returns whether the given extension is present in pb. func HasExtension(pb Message, extension *ExtensionDesc) bool { if epb, doki := pb.(extensionsBytes); doki { ext := epb.GetExtensions() buf := *ext o := 0 for o < len(buf) { tag, n := DecodeVarint(buf[o:]) fieldNum := int32(tag >> 3) if int32(fieldNum) == extension.Field { return true } wireType := int(tag & 0x7) o += n l, err := size(buf[o:], wireType) if err != nil { return false } o += l } return false } // TODO: Check types, field numbers, etc.? epb, err := extendable(pb) if err != nil { return false } extmap, mu := epb.extensionsRead() if extmap == nil { return false } mu.Lock() _, ok := extmap[extension.Field] mu.Unlock() return ok } // ClearExtension removes the given extension from pb. func ClearExtension(pb Message, extension *ExtensionDesc) { clearExtension(pb, extension.Field) } func clearExtension(pb Message, fieldNum int32) { if epb, ok := pb.(extensionsBytes); ok { offset := 0 for offset != -1 { offset = deleteExtension(epb, fieldNum, offset) } return } epb, err := extendable(pb) if err != nil { return } // TODO: Check types, field numbers, etc.? extmap := epb.extensionsWrite() delete(extmap, fieldNum) } // GetExtension retrieves a proto2 extended field from pb. // // If the descriptor is type complete (i.e., ExtensionDesc.ExtensionType is non-nil), // then GetExtension parses the encoded field and returns a Go value of the specified type. // If the field is not present, then the default value is returned (if one is specified), // otherwise ErrMissingExtension is reported. // // If the descriptor is not type complete (i.e., ExtensionDesc.ExtensionType is nil), // then GetExtension returns the raw encoded bytes of the field extension. func GetExtension(pb Message, extension *ExtensionDesc) (interface{}, error) { if epb, doki := pb.(extensionsBytes); doki { ext := epb.GetExtensions() return decodeExtensionFromBytes(extension, *ext) } epb, err := extendable(pb) if err != nil { return nil, err } if extension.ExtendedType != nil { // can only check type if this is a complete descriptor if cerr := checkExtensionTypes(epb, extension); cerr != nil { return nil, cerr } } emap, mu := epb.extensionsRead() if emap == nil { return defaultExtensionValue(extension) } mu.Lock() defer mu.Unlock() e, ok := emap[extension.Field] if !ok { // defaultExtensionValue returns the default value or // ErrMissingExtension if there is no default. return defaultExtensionValue(extension) } if e.value != nil { // Already decoded. Check the descriptor, though. if e.desc != extension { // This shouldn't happen. If it does, it means that // GetExtension was called twice with two different // descriptors with the same field number. return nil, errors.New("proto: descriptor conflict") } return e.value, nil } if extension.ExtensionType == nil { // incomplete descriptor return e.enc, nil } v, err := decodeExtension(e.enc, extension) if err != nil { return nil, err } // Remember the decoded version and drop the encoded version. // That way it is safe to mutate what we return. e.value = v e.desc = extension e.enc = nil emap[extension.Field] = e return e.value, nil } // defaultExtensionValue returns the default value for extension. // If no default for an extension is defined ErrMissingExtension is returned. func defaultExtensionValue(extension *ExtensionDesc) (interface{}, error) { if extension.ExtensionType == nil { // incomplete descriptor, so no default return nil, ErrMissingExtension } t := reflect.TypeOf(extension.ExtensionType) props := extensionProperties(extension) sf, _, err := fieldDefault(t, props) if err != nil { return nil, err } if sf == nil || sf.value == nil { // There is no default value. return nil, ErrMissingExtension } if t.Kind() != reflect.Ptr { // We do not need to return a Ptr, we can directly return sf.value. return sf.value, nil } // We need to return an interface{} that is a pointer to sf.value. value := reflect.New(t).Elem() value.Set(reflect.New(value.Type().Elem())) if sf.kind == reflect.Int32 { // We may have an int32 or an enum, but the underlying data is int32. // Since we can't set an int32 into a non int32 reflect.value directly // set it as a int32. value.Elem().SetInt(int64(sf.value.(int32))) } else { value.Elem().Set(reflect.ValueOf(sf.value)) } return value.Interface(), nil } // decodeExtension decodes an extension encoded in b. func decodeExtension(b []byte, extension *ExtensionDesc) (interface{}, error) { t := reflect.TypeOf(extension.ExtensionType) unmarshal := typeUnmarshaler(t, extension.Tag) // t is a pointer to a struct, pointer to basic type or a slice. // Allocate space to store the pointer/slice. value := reflect.New(t).Elem() var err error for { x, n := decodeVarint(b) if n == 0 { return nil, io.ErrUnexpectedEOF } b = b[n:] wire := int(x) & 7 b, err = unmarshal(b, valToPointer(value.Addr()), wire) if err != nil { return nil, err } if len(b) == 0 { break } } return value.Interface(), nil } // GetExtensions returns a slice of the extensions present in pb that are also listed in es. // The returned slice has the same length as es; missing extensions will appear as nil elements. func GetExtensions(pb Message, es []*ExtensionDesc) (extensions []interface{}, err error) { epb, err := extendable(pb) if err != nil { return nil, err } extensions = make([]interface{}, len(es)) for i, e := range es { extensions[i], err = GetExtension(epb, e) if err == ErrMissingExtension { err = nil } if err != nil { return } } return } // ExtensionDescs returns a new slice containing pb's extension descriptors, in undefined order. // For non-registered extensions, ExtensionDescs returns an incomplete descriptor containing // just the Field field, which defines the extension's field number. func ExtensionDescs(pb Message) ([]*ExtensionDesc, error) { epb, err := extendable(pb) if err != nil { return nil, err } registeredExtensions := RegisteredExtensions(pb) emap, mu := epb.extensionsRead() if emap == nil { return nil, nil } mu.Lock() defer mu.Unlock() extensions := make([]*ExtensionDesc, 0, len(emap)) for extid, e := range emap { desc := e.desc if desc == nil { desc = registeredExtensions[extid] if desc == nil { desc = &ExtensionDesc{Field: extid} } } extensions = append(extensions, desc) } return extensions, nil } // SetExtension sets the specified extension of pb to the specified value. func SetExtension(pb Message, extension *ExtensionDesc, value interface{}) error { if epb, ok := pb.(extensionsBytes); ok { ClearExtension(pb, extension) newb, err := encodeExtension(extension, value) if err != nil { return err } bb := epb.GetExtensions() *bb = append(*bb, newb...) return nil } epb, err := extendable(pb) if err != nil { return err } if err := checkExtensionTypes(epb, extension); err != nil { return err } typ := reflect.TypeOf(extension.ExtensionType) if typ != reflect.TypeOf(value) { return fmt.Errorf("proto: bad extension value type. got: %T, want: %T", value, extension.ExtensionType) } // nil extension values need to be caught early, because the // encoder can't distinguish an ErrNil due to a nil extension // from an ErrNil due to a missing field. Extensions are // always optional, so the encoder would just swallow the error // and drop all the extensions from the encoded message. if reflect.ValueOf(value).IsNil() { return fmt.Errorf("proto: SetExtension called with nil value of type %T", value) } extmap := epb.extensionsWrite() extmap[extension.Field] = Extension{desc: extension, value: value} return nil } // ClearAllExtensions clears all extensions from pb. func ClearAllExtensions(pb Message) { if epb, doki := pb.(extensionsBytes); doki { ext := epb.GetExtensions() *ext = []byte{} return } epb, err := extendable(pb) if err != nil { return } m := epb.extensionsWrite() for k := range m { delete(m, k) } } // A global registry of extensions. // The generated code will register the generated descriptors by calling RegisterExtension. var extensionMaps = make(map[reflect.Type]map[int32]*ExtensionDesc) // RegisterExtension is called from the generated code. func RegisterExtension(desc *ExtensionDesc) { st := reflect.TypeOf(desc.ExtendedType).Elem() m := extensionMaps[st] if m == nil { m = make(map[int32]*ExtensionDesc) extensionMaps[st] = m } if _, ok := m[desc.Field]; ok { panic("proto: duplicate extension registered: " + st.String() + " " + strconv.Itoa(int(desc.Field))) } m[desc.Field] = desc } // RegisteredExtensions returns a map of the registered extensions of a // protocol buffer struct, indexed by the extension number. // The argument pb should be a nil pointer to the struct type. func RegisteredExtensions(pb Message) map[int32]*ExtensionDesc { return extensionMaps[reflect.TypeOf(pb).Elem()] }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/gogo/protobuf/proto/text_gogo.go
vendor/github.com/gogo/protobuf/proto/text_gogo.go
// Protocol Buffers for Go with Gadgets // // Copyright (c) 2013, The GoGo Authors. All rights reserved. // http://github.com/gogo/protobuf // // Redistribution and use in source and binary forms, with or without // modification, are permitted provided that the following conditions are // met: // // * Redistributions of source code must retain the above copyright // notice, this list of conditions and the following disclaimer. // * Redistributions in binary form must reproduce the above // copyright notice, this list of conditions and the following disclaimer // in the documentation and/or other materials provided with the // distribution. // // THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS // "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT // LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR // A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT // OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, // SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT // LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, // DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY // THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT // (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE // OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. package proto import ( "fmt" "reflect" ) func (tm *TextMarshaler) writeEnum(w *textWriter, v reflect.Value, props *Properties) error { m, ok := enumStringMaps[props.Enum] if !ok { if err := tm.writeAny(w, v, props); err != nil { return err } } key := int32(0) if v.Kind() == reflect.Ptr { key = int32(v.Elem().Int()) } else { key = int32(v.Int()) } s, ok := m[key] if !ok { if err := tm.writeAny(w, v, props); err != nil { return err } } _, err := fmt.Fprint(w, s) return err }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/gogo/protobuf/proto/properties.go
vendor/github.com/gogo/protobuf/proto/properties.go
// Protocol Buffers for Go with Gadgets // // Copyright (c) 2013, The GoGo Authors. All rights reserved. // http://github.com/gogo/protobuf // // Go support for Protocol Buffers - Google's data interchange format // // Copyright 2010 The Go Authors. All rights reserved. // https://github.com/golang/protobuf // // Redistribution and use in source and binary forms, with or without // modification, are permitted provided that the following conditions are // met: // // * Redistributions of source code must retain the above copyright // notice, this list of conditions and the following disclaimer. // * Redistributions in binary form must reproduce the above // copyright notice, this list of conditions and the following disclaimer // in the documentation and/or other materials provided with the // distribution. // * Neither the name of Google Inc. nor the names of its // contributors may be used to endorse or promote products derived from // this software without specific prior written permission. // // THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS // "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT // LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR // A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT // OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, // SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT // LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, // DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY // THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT // (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE // OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. package proto /* * Routines for encoding data into the wire format for protocol buffers. */ import ( "fmt" "log" "reflect" "sort" "strconv" "strings" "sync" ) const debug bool = false // Constants that identify the encoding of a value on the wire. const ( WireVarint = 0 WireFixed64 = 1 WireBytes = 2 WireStartGroup = 3 WireEndGroup = 4 WireFixed32 = 5 ) // tagMap is an optimization over map[int]int for typical protocol buffer // use-cases. Encoded protocol buffers are often in tag order with small tag // numbers. type tagMap struct { fastTags []int slowTags map[int]int } // tagMapFastLimit is the upper bound on the tag number that will be stored in // the tagMap slice rather than its map. const tagMapFastLimit = 1024 func (p *tagMap) get(t int) (int, bool) { if t > 0 && t < tagMapFastLimit { if t >= len(p.fastTags) { return 0, false } fi := p.fastTags[t] return fi, fi >= 0 } fi, ok := p.slowTags[t] return fi, ok } func (p *tagMap) put(t int, fi int) { if t > 0 && t < tagMapFastLimit { for len(p.fastTags) < t+1 { p.fastTags = append(p.fastTags, -1) } p.fastTags[t] = fi return } if p.slowTags == nil { p.slowTags = make(map[int]int) } p.slowTags[t] = fi } // StructProperties represents properties for all the fields of a struct. // decoderTags and decoderOrigNames should only be used by the decoder. type StructProperties struct { Prop []*Properties // properties for each field reqCount int // required count decoderTags tagMap // map from proto tag to struct field number decoderOrigNames map[string]int // map from original name to struct field number order []int // list of struct field numbers in tag order // OneofTypes contains information about the oneof fields in this message. // It is keyed by the original name of a field. OneofTypes map[string]*OneofProperties } // OneofProperties represents information about a specific field in a oneof. type OneofProperties struct { Type reflect.Type // pointer to generated struct type for this oneof field Field int // struct field number of the containing oneof in the message Prop *Properties } // Implement the sorting interface so we can sort the fields in tag order, as recommended by the spec. // See encode.go, (*Buffer).enc_struct. func (sp *StructProperties) Len() int { return len(sp.order) } func (sp *StructProperties) Less(i, j int) bool { return sp.Prop[sp.order[i]].Tag < sp.Prop[sp.order[j]].Tag } func (sp *StructProperties) Swap(i, j int) { sp.order[i], sp.order[j] = sp.order[j], sp.order[i] } // Properties represents the protocol-specific behavior of a single struct field. type Properties struct { Name string // name of the field, for error messages OrigName string // original name before protocol compiler (always set) JSONName string // name to use for JSON; determined by protoc Wire string WireType int Tag int Required bool Optional bool Repeated bool Packed bool // relevant for repeated primitives only Enum string // set for enum types only proto3 bool // whether this is known to be a proto3 field oneof bool // whether this is a oneof field Default string // default value HasDefault bool // whether an explicit default was provided CustomType string CastType string StdTime bool StdDuration bool WktPointer bool stype reflect.Type // set for struct types only ctype reflect.Type // set for custom types only sprop *StructProperties // set for struct types only mtype reflect.Type // set for map types only MapKeyProp *Properties // set for map types only MapValProp *Properties // set for map types only } // String formats the properties in the protobuf struct field tag style. func (p *Properties) String() string { s := p.Wire s += "," s += strconv.Itoa(p.Tag) if p.Required { s += ",req" } if p.Optional { s += ",opt" } if p.Repeated { s += ",rep" } if p.Packed { s += ",packed" } s += ",name=" + p.OrigName if p.JSONName != p.OrigName { s += ",json=" + p.JSONName } if p.proto3 { s += ",proto3" } if p.oneof { s += ",oneof" } if len(p.Enum) > 0 { s += ",enum=" + p.Enum } if p.HasDefault { s += ",def=" + p.Default } return s } // Parse populates p by parsing a string in the protobuf struct field tag style. func (p *Properties) Parse(s string) { // "bytes,49,opt,name=foo,def=hello!" fields := strings.Split(s, ",") // breaks def=, but handled below. if len(fields) < 2 { log.Printf("proto: tag has too few fields: %q", s) return } p.Wire = fields[0] switch p.Wire { case "varint": p.WireType = WireVarint case "fixed32": p.WireType = WireFixed32 case "fixed64": p.WireType = WireFixed64 case "zigzag32": p.WireType = WireVarint case "zigzag64": p.WireType = WireVarint case "bytes", "group": p.WireType = WireBytes // no numeric converter for non-numeric types default: log.Printf("proto: tag has unknown wire type: %q", s) return } var err error p.Tag, err = strconv.Atoi(fields[1]) if err != nil { return } outer: for i := 2; i < len(fields); i++ { f := fields[i] switch { case f == "req": p.Required = true case f == "opt": p.Optional = true case f == "rep": p.Repeated = true case f == "packed": p.Packed = true case strings.HasPrefix(f, "name="): p.OrigName = f[5:] case strings.HasPrefix(f, "json="): p.JSONName = f[5:] case strings.HasPrefix(f, "enum="): p.Enum = f[5:] case f == "proto3": p.proto3 = true case f == "oneof": p.oneof = true case strings.HasPrefix(f, "def="): p.HasDefault = true p.Default = f[4:] // rest of string if i+1 < len(fields) { // Commas aren't escaped, and def is always last. p.Default += "," + strings.Join(fields[i+1:], ",") break outer } case strings.HasPrefix(f, "embedded="): p.OrigName = strings.Split(f, "=")[1] case strings.HasPrefix(f, "customtype="): p.CustomType = strings.Split(f, "=")[1] case strings.HasPrefix(f, "casttype="): p.CastType = strings.Split(f, "=")[1] case f == "stdtime": p.StdTime = true case f == "stdduration": p.StdDuration = true case f == "wktptr": p.WktPointer = true } } } var protoMessageType = reflect.TypeOf((*Message)(nil)).Elem() // setFieldProps initializes the field properties for submessages and maps. func (p *Properties) setFieldProps(typ reflect.Type, f *reflect.StructField, lockGetProp bool) { isMap := typ.Kind() == reflect.Map if len(p.CustomType) > 0 && !isMap { p.ctype = typ p.setTag(lockGetProp) return } if p.StdTime && !isMap { p.setTag(lockGetProp) return } if p.StdDuration && !isMap { p.setTag(lockGetProp) return } if p.WktPointer && !isMap { p.setTag(lockGetProp) return } switch t1 := typ; t1.Kind() { case reflect.Struct: p.stype = typ case reflect.Ptr: if t1.Elem().Kind() == reflect.Struct { p.stype = t1.Elem() } case reflect.Slice: switch t2 := t1.Elem(); t2.Kind() { case reflect.Ptr: switch t3 := t2.Elem(); t3.Kind() { case reflect.Struct: p.stype = t3 } case reflect.Struct: p.stype = t2 } case reflect.Map: p.mtype = t1 p.MapKeyProp = &Properties{} p.MapKeyProp.init(reflect.PtrTo(p.mtype.Key()), "Key", f.Tag.Get("protobuf_key"), nil, lockGetProp) p.MapValProp = &Properties{} vtype := p.mtype.Elem() if vtype.Kind() != reflect.Ptr && vtype.Kind() != reflect.Slice { // The value type is not a message (*T) or bytes ([]byte), // so we need encoders for the pointer to this type. vtype = reflect.PtrTo(vtype) } p.MapValProp.CustomType = p.CustomType p.MapValProp.StdDuration = p.StdDuration p.MapValProp.StdTime = p.StdTime p.MapValProp.WktPointer = p.WktPointer p.MapValProp.init(vtype, "Value", f.Tag.Get("protobuf_val"), nil, lockGetProp) } p.setTag(lockGetProp) } func (p *Properties) setTag(lockGetProp bool) { if p.stype != nil { if lockGetProp { p.sprop = GetProperties(p.stype) } else { p.sprop = getPropertiesLocked(p.stype) } } } var ( marshalerType = reflect.TypeOf((*Marshaler)(nil)).Elem() ) // Init populates the properties from a protocol buffer struct tag. func (p *Properties) Init(typ reflect.Type, name, tag string, f *reflect.StructField) { p.init(typ, name, tag, f, true) } func (p *Properties) init(typ reflect.Type, name, tag string, f *reflect.StructField, lockGetProp bool) { // "bytes,49,opt,def=hello!" p.Name = name p.OrigName = name if tag == "" { return } p.Parse(tag) p.setFieldProps(typ, f, lockGetProp) } var ( propertiesMu sync.RWMutex propertiesMap = make(map[reflect.Type]*StructProperties) ) // GetProperties returns the list of properties for the type represented by t. // t must represent a generated struct type of a protocol message. func GetProperties(t reflect.Type) *StructProperties { if t.Kind() != reflect.Struct { panic("proto: type must have kind struct") } // Most calls to GetProperties in a long-running program will be // retrieving details for types we have seen before. propertiesMu.RLock() sprop, ok := propertiesMap[t] propertiesMu.RUnlock() if ok { return sprop } propertiesMu.Lock() sprop = getPropertiesLocked(t) propertiesMu.Unlock() return sprop } type ( oneofFuncsIface interface { XXX_OneofFuncs() (func(Message, *Buffer) error, func(Message, int, int, *Buffer) (bool, error), func(Message) int, []interface{}) } oneofWrappersIface interface { XXX_OneofWrappers() []interface{} } ) // getPropertiesLocked requires that propertiesMu is held. func getPropertiesLocked(t reflect.Type) *StructProperties { if prop, ok := propertiesMap[t]; ok { return prop } prop := new(StructProperties) // in case of recursive protos, fill this in now. propertiesMap[t] = prop // build properties prop.Prop = make([]*Properties, t.NumField()) prop.order = make([]int, t.NumField()) isOneofMessage := false for i := 0; i < t.NumField(); i++ { f := t.Field(i) p := new(Properties) name := f.Name p.init(f.Type, name, f.Tag.Get("protobuf"), &f, false) oneof := f.Tag.Get("protobuf_oneof") // special case if oneof != "" { isOneofMessage = true // Oneof fields don't use the traditional protobuf tag. p.OrigName = oneof } prop.Prop[i] = p prop.order[i] = i if debug { print(i, " ", f.Name, " ", t.String(), " ") if p.Tag > 0 { print(p.String()) } print("\n") } } // Re-order prop.order. sort.Sort(prop) if isOneofMessage { var oots []interface{} switch m := reflect.Zero(reflect.PtrTo(t)).Interface().(type) { case oneofFuncsIface: _, _, _, oots = m.XXX_OneofFuncs() case oneofWrappersIface: oots = m.XXX_OneofWrappers() } if len(oots) > 0 { // Interpret oneof metadata. prop.OneofTypes = make(map[string]*OneofProperties) for _, oot := range oots { oop := &OneofProperties{ Type: reflect.ValueOf(oot).Type(), // *T Prop: new(Properties), } sft := oop.Type.Elem().Field(0) oop.Prop.Name = sft.Name oop.Prop.Parse(sft.Tag.Get("protobuf")) // There will be exactly one interface field that // this new value is assignable to. for i := 0; i < t.NumField(); i++ { f := t.Field(i) if f.Type.Kind() != reflect.Interface { continue } if !oop.Type.AssignableTo(f.Type) { continue } oop.Field = i break } prop.OneofTypes[oop.Prop.OrigName] = oop } } } // build required counts // build tags reqCount := 0 prop.decoderOrigNames = make(map[string]int) for i, p := range prop.Prop { if strings.HasPrefix(p.Name, "XXX_") { // Internal fields should not appear in tags/origNames maps. // They are handled specially when encoding and decoding. continue } if p.Required { reqCount++ } prop.decoderTags.put(p.Tag, i) prop.decoderOrigNames[p.OrigName] = i } prop.reqCount = reqCount return prop } // A global registry of enum types. // The generated code will register the generated maps by calling RegisterEnum. var enumValueMaps = make(map[string]map[string]int32) var enumStringMaps = make(map[string]map[int32]string) // RegisterEnum is called from the generated code to install the enum descriptor // maps into the global table to aid parsing text format protocol buffers. func RegisterEnum(typeName string, unusedNameMap map[int32]string, valueMap map[string]int32) { if _, ok := enumValueMaps[typeName]; ok { panic("proto: duplicate enum registered: " + typeName) } enumValueMaps[typeName] = valueMap if _, ok := enumStringMaps[typeName]; ok { panic("proto: duplicate enum registered: " + typeName) } enumStringMaps[typeName] = unusedNameMap } // EnumValueMap returns the mapping from names to integers of the // enum type enumType, or a nil if not found. func EnumValueMap(enumType string) map[string]int32 { return enumValueMaps[enumType] } // A registry of all linked message types. // The string is a fully-qualified proto name ("pkg.Message"). var ( protoTypedNils = make(map[string]Message) // a map from proto names to typed nil pointers protoMapTypes = make(map[string]reflect.Type) // a map from proto names to map types revProtoTypes = make(map[reflect.Type]string) ) // RegisterType is called from generated code and maps from the fully qualified // proto name to the type (pointer to struct) of the protocol buffer. func RegisterType(x Message, name string) { if _, ok := protoTypedNils[name]; ok { // TODO: Some day, make this a panic. log.Printf("proto: duplicate proto type registered: %s", name) return } t := reflect.TypeOf(x) if v := reflect.ValueOf(x); v.Kind() == reflect.Ptr && v.Pointer() == 0 { // Generated code always calls RegisterType with nil x. // This check is just for extra safety. protoTypedNils[name] = x } else { protoTypedNils[name] = reflect.Zero(t).Interface().(Message) } revProtoTypes[t] = name } // RegisterMapType is called from generated code and maps from the fully qualified // proto name to the native map type of the proto map definition. func RegisterMapType(x interface{}, name string) { if reflect.TypeOf(x).Kind() != reflect.Map { panic(fmt.Sprintf("RegisterMapType(%T, %q); want map", x, name)) } if _, ok := protoMapTypes[name]; ok { log.Printf("proto: duplicate proto type registered: %s", name) return } t := reflect.TypeOf(x) protoMapTypes[name] = t revProtoTypes[t] = name } // MessageName returns the fully-qualified proto name for the given message type. func MessageName(x Message) string { type xname interface { XXX_MessageName() string } if m, ok := x.(xname); ok { return m.XXX_MessageName() } return revProtoTypes[reflect.TypeOf(x)] } // MessageType returns the message type (pointer to struct) for a named message. // The type is not guaranteed to implement proto.Message if the name refers to a // map entry. func MessageType(name string) reflect.Type { if t, ok := protoTypedNils[name]; ok { return reflect.TypeOf(t) } return protoMapTypes[name] } // A registry of all linked proto files. var ( protoFiles = make(map[string][]byte) // file name => fileDescriptor ) // RegisterFile is called from generated code and maps from the // full file name of a .proto file to its compressed FileDescriptorProto. func RegisterFile(filename string, fileDescriptor []byte) { protoFiles[filename] = fileDescriptor } // FileDescriptor returns the compressed FileDescriptorProto for a .proto file. func FileDescriptor(filename string) []byte { return protoFiles[filename] }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/gogo/protobuf/proto/skip_gogo.go
vendor/github.com/gogo/protobuf/proto/skip_gogo.go
// Protocol Buffers for Go with Gadgets // // Copyright (c) 2013, The GoGo Authors. All rights reserved. // http://github.com/gogo/protobuf // // Redistribution and use in source and binary forms, with or without // modification, are permitted provided that the following conditions are // met: // // * Redistributions of source code must retain the above copyright // notice, this list of conditions and the following disclaimer. // * Redistributions in binary form must reproduce the above // copyright notice, this list of conditions and the following disclaimer // in the documentation and/or other materials provided with the // distribution. // // THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS // "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT // LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR // A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT // OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, // SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT // LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, // DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY // THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT // (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE // OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. package proto import ( "fmt" "io" ) func Skip(data []byte) (n int, err error) { l := len(data) index := 0 for index < l { var wire uint64 for shift := uint(0); ; shift += 7 { if index >= l { return 0, io.ErrUnexpectedEOF } b := data[index] index++ wire |= (uint64(b) & 0x7F) << shift if b < 0x80 { break } } wireType := int(wire & 0x7) switch wireType { case 0: for { if index >= l { return 0, io.ErrUnexpectedEOF } index++ if data[index-1] < 0x80 { break } } return index, nil case 1: index += 8 return index, nil case 2: var length int for shift := uint(0); ; shift += 7 { if index >= l { return 0, io.ErrUnexpectedEOF } b := data[index] index++ length |= (int(b) & 0x7F) << shift if b < 0x80 { break } } index += length return index, nil case 3: for { var innerWire uint64 var start int = index for shift := uint(0); ; shift += 7 { if index >= l { return 0, io.ErrUnexpectedEOF } b := data[index] index++ innerWire |= (uint64(b) & 0x7F) << shift if b < 0x80 { break } } innerWireType := int(innerWire & 0x7) if innerWireType == 4 { break } next, err := Skip(data[start:]) if err != nil { return 0, err } index = start + next } return index, nil case 4: return index, nil case 5: index += 4 return index, nil default: return 0, fmt.Errorf("proto: illegal wireType %d", wireType) } } panic("unreachable") }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/gogo/protobuf/proto/lib.go
vendor/github.com/gogo/protobuf/proto/lib.go
// Go support for Protocol Buffers - Google's data interchange format // // Copyright 2010 The Go Authors. All rights reserved. // https://github.com/golang/protobuf // // Redistribution and use in source and binary forms, with or without // modification, are permitted provided that the following conditions are // met: // // * Redistributions of source code must retain the above copyright // notice, this list of conditions and the following disclaimer. // * Redistributions in binary form must reproduce the above // copyright notice, this list of conditions and the following disclaimer // in the documentation and/or other materials provided with the // distribution. // * Neither the name of Google Inc. nor the names of its // contributors may be used to endorse or promote products derived from // this software without specific prior written permission. // // THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS // "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT // LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR // A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT // OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, // SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT // LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, // DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY // THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT // (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE // OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. /* Package proto converts data structures to and from the wire format of protocol buffers. It works in concert with the Go source code generated for .proto files by the protocol compiler. A summary of the properties of the protocol buffer interface for a protocol buffer variable v: - Names are turned from camel_case to CamelCase for export. - There are no methods on v to set fields; just treat them as structure fields. - There are getters that return a field's value if set, and return the field's default value if unset. The getters work even if the receiver is a nil message. - The zero value for a struct is its correct initialization state. All desired fields must be set before marshaling. - A Reset() method will restore a protobuf struct to its zero state. - Non-repeated fields are pointers to the values; nil means unset. That is, optional or required field int32 f becomes F *int32. - Repeated fields are slices. - Helper functions are available to aid the setting of fields. msg.Foo = proto.String("hello") // set field - Constants are defined to hold the default values of all fields that have them. They have the form Default_StructName_FieldName. Because the getter methods handle defaulted values, direct use of these constants should be rare. - Enums are given type names and maps from names to values. Enum values are prefixed by the enclosing message's name, or by the enum's type name if it is a top-level enum. Enum types have a String method, and a Enum method to assist in message construction. - Nested messages, groups and enums have type names prefixed with the name of the surrounding message type. - Extensions are given descriptor names that start with E_, followed by an underscore-delimited list of the nested messages that contain it (if any) followed by the CamelCased name of the extension field itself. HasExtension, ClearExtension, GetExtension and SetExtension are functions for manipulating extensions. - Oneof field sets are given a single field in their message, with distinguished wrapper types for each possible field value. - Marshal and Unmarshal are functions to encode and decode the wire format. When the .proto file specifies `syntax="proto3"`, there are some differences: - Non-repeated fields of non-message type are values instead of pointers. - Enum types do not get an Enum method. The simplest way to describe this is to see an example. Given file test.proto, containing package example; enum FOO { X = 17; } message Test { required string label = 1; optional int32 type = 2 [default=77]; repeated int64 reps = 3; optional group OptionalGroup = 4 { required string RequiredField = 5; } oneof union { int32 number = 6; string name = 7; } } The resulting file, test.pb.go, is: package example import proto "github.com/gogo/protobuf/proto" import math "math" type FOO int32 const ( FOO_X FOO = 17 ) var FOO_name = map[int32]string{ 17: "X", } var FOO_value = map[string]int32{ "X": 17, } func (x FOO) Enum() *FOO { p := new(FOO) *p = x return p } func (x FOO) String() string { return proto.EnumName(FOO_name, int32(x)) } func (x *FOO) UnmarshalJSON(data []byte) error { value, err := proto.UnmarshalJSONEnum(FOO_value, data) if err != nil { return err } *x = FOO(value) return nil } type Test struct { Label *string `protobuf:"bytes,1,req,name=label" json:"label,omitempty"` Type *int32 `protobuf:"varint,2,opt,name=type,def=77" json:"type,omitempty"` Reps []int64 `protobuf:"varint,3,rep,name=reps" json:"reps,omitempty"` Optionalgroup *Test_OptionalGroup `protobuf:"group,4,opt,name=OptionalGroup" json:"optionalgroup,omitempty"` // Types that are valid to be assigned to Union: // *Test_Number // *Test_Name Union isTest_Union `protobuf_oneof:"union"` XXX_unrecognized []byte `json:"-"` } func (m *Test) Reset() { *m = Test{} } func (m *Test) String() string { return proto.CompactTextString(m) } func (*Test) ProtoMessage() {} type isTest_Union interface { isTest_Union() } type Test_Number struct { Number int32 `protobuf:"varint,6,opt,name=number"` } type Test_Name struct { Name string `protobuf:"bytes,7,opt,name=name"` } func (*Test_Number) isTest_Union() {} func (*Test_Name) isTest_Union() {} func (m *Test) GetUnion() isTest_Union { if m != nil { return m.Union } return nil } const Default_Test_Type int32 = 77 func (m *Test) GetLabel() string { if m != nil && m.Label != nil { return *m.Label } return "" } func (m *Test) GetType() int32 { if m != nil && m.Type != nil { return *m.Type } return Default_Test_Type } func (m *Test) GetOptionalgroup() *Test_OptionalGroup { if m != nil { return m.Optionalgroup } return nil } type Test_OptionalGroup struct { RequiredField *string `protobuf:"bytes,5,req" json:"RequiredField,omitempty"` } func (m *Test_OptionalGroup) Reset() { *m = Test_OptionalGroup{} } func (m *Test_OptionalGroup) String() string { return proto.CompactTextString(m) } func (m *Test_OptionalGroup) GetRequiredField() string { if m != nil && m.RequiredField != nil { return *m.RequiredField } return "" } func (m *Test) GetNumber() int32 { if x, ok := m.GetUnion().(*Test_Number); ok { return x.Number } return 0 } func (m *Test) GetName() string { if x, ok := m.GetUnion().(*Test_Name); ok { return x.Name } return "" } func init() { proto.RegisterEnum("example.FOO", FOO_name, FOO_value) } To create and play with a Test object: package main import ( "log" "github.com/gogo/protobuf/proto" pb "./example.pb" ) func main() { test := &pb.Test{ Label: proto.String("hello"), Type: proto.Int32(17), Reps: []int64{1, 2, 3}, Optionalgroup: &pb.Test_OptionalGroup{ RequiredField: proto.String("good bye"), }, Union: &pb.Test_Name{"fred"}, } data, err := proto.Marshal(test) if err != nil { log.Fatal("marshaling error: ", err) } newTest := &pb.Test{} err = proto.Unmarshal(data, newTest) if err != nil { log.Fatal("unmarshaling error: ", err) } // Now test and newTest contain the same data. if test.GetLabel() != newTest.GetLabel() { log.Fatalf("data mismatch %q != %q", test.GetLabel(), newTest.GetLabel()) } // Use a type switch to determine which oneof was set. switch u := test.Union.(type) { case *pb.Test_Number: // u.Number contains the number. case *pb.Test_Name: // u.Name contains the string. } // etc. } */ package proto import ( "encoding/json" "fmt" "log" "reflect" "sort" "strconv" "sync" ) // RequiredNotSetError is an error type returned by either Marshal or Unmarshal. // Marshal reports this when a required field is not initialized. // Unmarshal reports this when a required field is missing from the wire data. type RequiredNotSetError struct{ field string } func (e *RequiredNotSetError) Error() string { if e.field == "" { return fmt.Sprintf("proto: required field not set") } return fmt.Sprintf("proto: required field %q not set", e.field) } func (e *RequiredNotSetError) RequiredNotSet() bool { return true } type invalidUTF8Error struct{ field string } func (e *invalidUTF8Error) Error() string { if e.field == "" { return "proto: invalid UTF-8 detected" } return fmt.Sprintf("proto: field %q contains invalid UTF-8", e.field) } func (e *invalidUTF8Error) InvalidUTF8() bool { return true } // errInvalidUTF8 is a sentinel error to identify fields with invalid UTF-8. // This error should not be exposed to the external API as such errors should // be recreated with the field information. var errInvalidUTF8 = &invalidUTF8Error{} // isNonFatal reports whether the error is either a RequiredNotSet error // or a InvalidUTF8 error. func isNonFatal(err error) bool { if re, ok := err.(interface{ RequiredNotSet() bool }); ok && re.RequiredNotSet() { return true } if re, ok := err.(interface{ InvalidUTF8() bool }); ok && re.InvalidUTF8() { return true } return false } type nonFatal struct{ E error } // Merge merges err into nf and reports whether it was successful. // Otherwise it returns false for any fatal non-nil errors. func (nf *nonFatal) Merge(err error) (ok bool) { if err == nil { return true // not an error } if !isNonFatal(err) { return false // fatal error } if nf.E == nil { nf.E = err // store first instance of non-fatal error } return true } // Message is implemented by generated protocol buffer messages. type Message interface { Reset() String() string ProtoMessage() } // A Buffer is a buffer manager for marshaling and unmarshaling // protocol buffers. It may be reused between invocations to // reduce memory usage. It is not necessary to use a Buffer; // the global functions Marshal and Unmarshal create a // temporary Buffer and are fine for most applications. type Buffer struct { buf []byte // encode/decode byte stream index int // read point deterministic bool } // NewBuffer allocates a new Buffer and initializes its internal data to // the contents of the argument slice. func NewBuffer(e []byte) *Buffer { return &Buffer{buf: e} } // Reset resets the Buffer, ready for marshaling a new protocol buffer. func (p *Buffer) Reset() { p.buf = p.buf[0:0] // for reading/writing p.index = 0 // for reading } // SetBuf replaces the internal buffer with the slice, // ready for unmarshaling the contents of the slice. func (p *Buffer) SetBuf(s []byte) { p.buf = s p.index = 0 } // Bytes returns the contents of the Buffer. func (p *Buffer) Bytes() []byte { return p.buf } // SetDeterministic sets whether to use deterministic serialization. // // Deterministic serialization guarantees that for a given binary, equal // messages will always be serialized to the same bytes. This implies: // // - Repeated serialization of a message will return the same bytes. // - Different processes of the same binary (which may be executing on // different machines) will serialize equal messages to the same bytes. // // Note that the deterministic serialization is NOT canonical across // languages. It is not guaranteed to remain stable over time. It is unstable // across different builds with schema changes due to unknown fields. // Users who need canonical serialization (e.g., persistent storage in a // canonical form, fingerprinting, etc.) should define their own // canonicalization specification and implement their own serializer rather // than relying on this API. // // If deterministic serialization is requested, map entries will be sorted // by keys in lexographical order. This is an implementation detail and // subject to change. func (p *Buffer) SetDeterministic(deterministic bool) { p.deterministic = deterministic } /* * Helper routines for simplifying the creation of optional fields of basic type. */ // Bool is a helper routine that allocates a new bool value // to store v and returns a pointer to it. func Bool(v bool) *bool { return &v } // Int32 is a helper routine that allocates a new int32 value // to store v and returns a pointer to it. func Int32(v int32) *int32 { return &v } // Int is a helper routine that allocates a new int32 value // to store v and returns a pointer to it, but unlike Int32 // its argument value is an int. func Int(v int) *int32 { p := new(int32) *p = int32(v) return p } // Int64 is a helper routine that allocates a new int64 value // to store v and returns a pointer to it. func Int64(v int64) *int64 { return &v } // Float32 is a helper routine that allocates a new float32 value // to store v and returns a pointer to it. func Float32(v float32) *float32 { return &v } // Float64 is a helper routine that allocates a new float64 value // to store v and returns a pointer to it. func Float64(v float64) *float64 { return &v } // Uint32 is a helper routine that allocates a new uint32 value // to store v and returns a pointer to it. func Uint32(v uint32) *uint32 { return &v } // Uint64 is a helper routine that allocates a new uint64 value // to store v and returns a pointer to it. func Uint64(v uint64) *uint64 { return &v } // String is a helper routine that allocates a new string value // to store v and returns a pointer to it. func String(v string) *string { return &v } // EnumName is a helper function to simplify printing protocol buffer enums // by name. Given an enum map and a value, it returns a useful string. func EnumName(m map[int32]string, v int32) string { s, ok := m[v] if ok { return s } return strconv.Itoa(int(v)) } // UnmarshalJSONEnum is a helper function to simplify recovering enum int values // from their JSON-encoded representation. Given a map from the enum's symbolic // names to its int values, and a byte buffer containing the JSON-encoded // value, it returns an int32 that can be cast to the enum type by the caller. // // The function can deal with both JSON representations, numeric and symbolic. func UnmarshalJSONEnum(m map[string]int32, data []byte, enumName string) (int32, error) { if data[0] == '"' { // New style: enums are strings. var repr string if err := json.Unmarshal(data, &repr); err != nil { return -1, err } val, ok := m[repr] if !ok { return 0, fmt.Errorf("unrecognized enum %s value %q", enumName, repr) } return val, nil } // Old style: enums are ints. var val int32 if err := json.Unmarshal(data, &val); err != nil { return 0, fmt.Errorf("cannot unmarshal %#q into enum %s", data, enumName) } return val, nil } // DebugPrint dumps the encoded data in b in a debugging format with a header // including the string s. Used in testing but made available for general debugging. func (p *Buffer) DebugPrint(s string, b []byte) { var u uint64 obuf := p.buf sindex := p.index p.buf = b p.index = 0 depth := 0 fmt.Printf("\n--- %s ---\n", s) out: for { for i := 0; i < depth; i++ { fmt.Print(" ") } index := p.index if index == len(p.buf) { break } op, err := p.DecodeVarint() if err != nil { fmt.Printf("%3d: fetching op err %v\n", index, err) break out } tag := op >> 3 wire := op & 7 switch wire { default: fmt.Printf("%3d: t=%3d unknown wire=%d\n", index, tag, wire) break out case WireBytes: var r []byte r, err = p.DecodeRawBytes(false) if err != nil { break out } fmt.Printf("%3d: t=%3d bytes [%d]", index, tag, len(r)) if len(r) <= 6 { for i := 0; i < len(r); i++ { fmt.Printf(" %.2x", r[i]) } } else { for i := 0; i < 3; i++ { fmt.Printf(" %.2x", r[i]) } fmt.Printf(" ..") for i := len(r) - 3; i < len(r); i++ { fmt.Printf(" %.2x", r[i]) } } fmt.Printf("\n") case WireFixed32: u, err = p.DecodeFixed32() if err != nil { fmt.Printf("%3d: t=%3d fix32 err %v\n", index, tag, err) break out } fmt.Printf("%3d: t=%3d fix32 %d\n", index, tag, u) case WireFixed64: u, err = p.DecodeFixed64() if err != nil { fmt.Printf("%3d: t=%3d fix64 err %v\n", index, tag, err) break out } fmt.Printf("%3d: t=%3d fix64 %d\n", index, tag, u) case WireVarint: u, err = p.DecodeVarint() if err != nil { fmt.Printf("%3d: t=%3d varint err %v\n", index, tag, err) break out } fmt.Printf("%3d: t=%3d varint %d\n", index, tag, u) case WireStartGroup: fmt.Printf("%3d: t=%3d start\n", index, tag) depth++ case WireEndGroup: depth-- fmt.Printf("%3d: t=%3d end\n", index, tag) } } if depth != 0 { fmt.Printf("%3d: start-end not balanced %d\n", p.index, depth) } fmt.Printf("\n") p.buf = obuf p.index = sindex } // SetDefaults sets unset protocol buffer fields to their default values. // It only modifies fields that are both unset and have defined defaults. // It recursively sets default values in any non-nil sub-messages. func SetDefaults(pb Message) { setDefaults(reflect.ValueOf(pb), true, false) } // v is a struct. func setDefaults(v reflect.Value, recur, zeros bool) { if v.Kind() == reflect.Ptr { v = v.Elem() } defaultMu.RLock() dm, ok := defaults[v.Type()] defaultMu.RUnlock() if !ok { dm = buildDefaultMessage(v.Type()) defaultMu.Lock() defaults[v.Type()] = dm defaultMu.Unlock() } for _, sf := range dm.scalars { f := v.Field(sf.index) if !f.IsNil() { // field already set continue } dv := sf.value if dv == nil && !zeros { // no explicit default, and don't want to set zeros continue } fptr := f.Addr().Interface() // **T // TODO: Consider batching the allocations we do here. switch sf.kind { case reflect.Bool: b := new(bool) if dv != nil { *b = dv.(bool) } *(fptr.(**bool)) = b case reflect.Float32: f := new(float32) if dv != nil { *f = dv.(float32) } *(fptr.(**float32)) = f case reflect.Float64: f := new(float64) if dv != nil { *f = dv.(float64) } *(fptr.(**float64)) = f case reflect.Int32: // might be an enum if ft := f.Type(); ft != int32PtrType { // enum f.Set(reflect.New(ft.Elem())) if dv != nil { f.Elem().SetInt(int64(dv.(int32))) } } else { // int32 field i := new(int32) if dv != nil { *i = dv.(int32) } *(fptr.(**int32)) = i } case reflect.Int64: i := new(int64) if dv != nil { *i = dv.(int64) } *(fptr.(**int64)) = i case reflect.String: s := new(string) if dv != nil { *s = dv.(string) } *(fptr.(**string)) = s case reflect.Uint8: // exceptional case: []byte var b []byte if dv != nil { db := dv.([]byte) b = make([]byte, len(db)) copy(b, db) } else { b = []byte{} } *(fptr.(*[]byte)) = b case reflect.Uint32: u := new(uint32) if dv != nil { *u = dv.(uint32) } *(fptr.(**uint32)) = u case reflect.Uint64: u := new(uint64) if dv != nil { *u = dv.(uint64) } *(fptr.(**uint64)) = u default: log.Printf("proto: can't set default for field %v (sf.kind=%v)", f, sf.kind) } } for _, ni := range dm.nested { f := v.Field(ni) // f is *T or T or []*T or []T switch f.Kind() { case reflect.Struct: setDefaults(f, recur, zeros) case reflect.Ptr: if f.IsNil() { continue } setDefaults(f, recur, zeros) case reflect.Slice: for i := 0; i < f.Len(); i++ { e := f.Index(i) if e.Kind() == reflect.Ptr && e.IsNil() { continue } setDefaults(e, recur, zeros) } case reflect.Map: for _, k := range f.MapKeys() { e := f.MapIndex(k) if e.IsNil() { continue } setDefaults(e, recur, zeros) } } } } var ( // defaults maps a protocol buffer struct type to a slice of the fields, // with its scalar fields set to their proto-declared non-zero default values. defaultMu sync.RWMutex defaults = make(map[reflect.Type]defaultMessage) int32PtrType = reflect.TypeOf((*int32)(nil)) ) // defaultMessage represents information about the default values of a message. type defaultMessage struct { scalars []scalarField nested []int // struct field index of nested messages } type scalarField struct { index int // struct field index kind reflect.Kind // element type (the T in *T or []T) value interface{} // the proto-declared default value, or nil } // t is a struct type. func buildDefaultMessage(t reflect.Type) (dm defaultMessage) { sprop := GetProperties(t) for _, prop := range sprop.Prop { fi, ok := sprop.decoderTags.get(prop.Tag) if !ok { // XXX_unrecognized continue } ft := t.Field(fi).Type sf, nested, err := fieldDefault(ft, prop) switch { case err != nil: log.Print(err) case nested: dm.nested = append(dm.nested, fi) case sf != nil: sf.index = fi dm.scalars = append(dm.scalars, *sf) } } return dm } // fieldDefault returns the scalarField for field type ft. // sf will be nil if the field can not have a default. // nestedMessage will be true if this is a nested message. // Note that sf.index is not set on return. func fieldDefault(ft reflect.Type, prop *Properties) (sf *scalarField, nestedMessage bool, err error) { var canHaveDefault bool switch ft.Kind() { case reflect.Struct: nestedMessage = true // non-nullable case reflect.Ptr: if ft.Elem().Kind() == reflect.Struct { nestedMessage = true } else { canHaveDefault = true // proto2 scalar field } case reflect.Slice: switch ft.Elem().Kind() { case reflect.Ptr, reflect.Struct: nestedMessage = true // repeated message case reflect.Uint8: canHaveDefault = true // bytes field } case reflect.Map: if ft.Elem().Kind() == reflect.Ptr { nestedMessage = true // map with message values } } if !canHaveDefault { if nestedMessage { return nil, true, nil } return nil, false, nil } // We now know that ft is a pointer or slice. sf = &scalarField{kind: ft.Elem().Kind()} // scalar fields without defaults if !prop.HasDefault { return sf, false, nil } // a scalar field: either *T or []byte switch ft.Elem().Kind() { case reflect.Bool: x, err := strconv.ParseBool(prop.Default) if err != nil { return nil, false, fmt.Errorf("proto: bad default bool %q: %v", prop.Default, err) } sf.value = x case reflect.Float32: x, err := strconv.ParseFloat(prop.Default, 32) if err != nil { return nil, false, fmt.Errorf("proto: bad default float32 %q: %v", prop.Default, err) } sf.value = float32(x) case reflect.Float64: x, err := strconv.ParseFloat(prop.Default, 64) if err != nil { return nil, false, fmt.Errorf("proto: bad default float64 %q: %v", prop.Default, err) } sf.value = x case reflect.Int32: x, err := strconv.ParseInt(prop.Default, 10, 32) if err != nil { return nil, false, fmt.Errorf("proto: bad default int32 %q: %v", prop.Default, err) } sf.value = int32(x) case reflect.Int64: x, err := strconv.ParseInt(prop.Default, 10, 64) if err != nil { return nil, false, fmt.Errorf("proto: bad default int64 %q: %v", prop.Default, err) } sf.value = x case reflect.String: sf.value = prop.Default case reflect.Uint8: // []byte (not *uint8) sf.value = []byte(prop.Default) case reflect.Uint32: x, err := strconv.ParseUint(prop.Default, 10, 32) if err != nil { return nil, false, fmt.Errorf("proto: bad default uint32 %q: %v", prop.Default, err) } sf.value = uint32(x) case reflect.Uint64: x, err := strconv.ParseUint(prop.Default, 10, 64) if err != nil { return nil, false, fmt.Errorf("proto: bad default uint64 %q: %v", prop.Default, err) } sf.value = x default: return nil, false, fmt.Errorf("proto: unhandled def kind %v", ft.Elem().Kind()) } return sf, false, nil } // mapKeys returns a sort.Interface to be used for sorting the map keys. // Map fields may have key types of non-float scalars, strings and enums. func mapKeys(vs []reflect.Value) sort.Interface { s := mapKeySorter{vs: vs} // Type specialization per https://developers.google.com/protocol-buffers/docs/proto#maps. if len(vs) == 0 { return s } switch vs[0].Kind() { case reflect.Int32, reflect.Int64: s.less = func(a, b reflect.Value) bool { return a.Int() < b.Int() } case reflect.Uint32, reflect.Uint64: s.less = func(a, b reflect.Value) bool { return a.Uint() < b.Uint() } case reflect.Bool: s.less = func(a, b reflect.Value) bool { return !a.Bool() && b.Bool() } // false < true case reflect.String: s.less = func(a, b reflect.Value) bool { return a.String() < b.String() } default: panic(fmt.Sprintf("unsupported map key type: %v", vs[0].Kind())) } return s } type mapKeySorter struct { vs []reflect.Value less func(a, b reflect.Value) bool } func (s mapKeySorter) Len() int { return len(s.vs) } func (s mapKeySorter) Swap(i, j int) { s.vs[i], s.vs[j] = s.vs[j], s.vs[i] } func (s mapKeySorter) Less(i, j int) bool { return s.less(s.vs[i], s.vs[j]) } // isProto3Zero reports whether v is a zero proto3 value. func isProto3Zero(v reflect.Value) bool { switch v.Kind() { case reflect.Bool: return !v.Bool() case reflect.Int32, reflect.Int64: return v.Int() == 0 case reflect.Uint32, reflect.Uint64: return v.Uint() == 0 case reflect.Float32, reflect.Float64: return v.Float() == 0 case reflect.String: return v.String() == "" } return false } const ( // ProtoPackageIsVersion3 is referenced from generated protocol buffer files // to assert that that code is compatible with this version of the proto package. GoGoProtoPackageIsVersion3 = true // ProtoPackageIsVersion2 is referenced from generated protocol buffer files // to assert that that code is compatible with this version of the proto package. GoGoProtoPackageIsVersion2 = true // ProtoPackageIsVersion1 is referenced from generated protocol buffer files // to assert that that code is compatible with this version of the proto package. GoGoProtoPackageIsVersion1 = true ) // InternalMessageInfo is a type used internally by generated .pb.go files. // This type is not intended to be used by non-generated code. // This type is not subject to any compatibility guarantee. type InternalMessageInfo struct { marshal *marshalInfo unmarshal *unmarshalInfo merge *mergeInfo discard *discardInfo }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/gogo/protobuf/proto/extensions_gogo.go
vendor/github.com/gogo/protobuf/proto/extensions_gogo.go
// Protocol Buffers for Go with Gadgets // // Copyright (c) 2013, The GoGo Authors. All rights reserved. // http://github.com/gogo/protobuf // // Redistribution and use in source and binary forms, with or without // modification, are permitted provided that the following conditions are // met: // // * Redistributions of source code must retain the above copyright // notice, this list of conditions and the following disclaimer. // * Redistributions in binary form must reproduce the above // copyright notice, this list of conditions and the following disclaimer // in the documentation and/or other materials provided with the // distribution. // // THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS // "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT // LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR // A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT // OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, // SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT // LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, // DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY // THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT // (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE // OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. package proto import ( "bytes" "errors" "fmt" "io" "reflect" "sort" "strings" "sync" ) type extensionsBytes interface { Message ExtensionRangeArray() []ExtensionRange GetExtensions() *[]byte } type slowExtensionAdapter struct { extensionsBytes } func (s slowExtensionAdapter) extensionsWrite() map[int32]Extension { panic("Please report a bug to github.com/gogo/protobuf if you see this message: Writing extensions is not supported for extensions stored in a byte slice field.") } func (s slowExtensionAdapter) extensionsRead() (map[int32]Extension, sync.Locker) { b := s.GetExtensions() m, err := BytesToExtensionsMap(*b) if err != nil { panic(err) } return m, notLocker{} } func GetBoolExtension(pb Message, extension *ExtensionDesc, ifnotset bool) bool { if reflect.ValueOf(pb).IsNil() { return ifnotset } value, err := GetExtension(pb, extension) if err != nil { return ifnotset } if value == nil { return ifnotset } if value.(*bool) == nil { return ifnotset } return *(value.(*bool)) } func (this *Extension) Equal(that *Extension) bool { if err := this.Encode(); err != nil { return false } if err := that.Encode(); err != nil { return false } return bytes.Equal(this.enc, that.enc) } func (this *Extension) Compare(that *Extension) int { if err := this.Encode(); err != nil { return 1 } if err := that.Encode(); err != nil { return -1 } return bytes.Compare(this.enc, that.enc) } func SizeOfInternalExtension(m extendableProto) (n int) { info := getMarshalInfo(reflect.TypeOf(m)) return info.sizeV1Extensions(m.extensionsWrite()) } type sortableMapElem struct { field int32 ext Extension } func newSortableExtensionsFromMap(m map[int32]Extension) sortableExtensions { s := make(sortableExtensions, 0, len(m)) for k, v := range m { s = append(s, &sortableMapElem{field: k, ext: v}) } return s } type sortableExtensions []*sortableMapElem func (this sortableExtensions) Len() int { return len(this) } func (this sortableExtensions) Swap(i, j int) { this[i], this[j] = this[j], this[i] } func (this sortableExtensions) Less(i, j int) bool { return this[i].field < this[j].field } func (this sortableExtensions) String() string { sort.Sort(this) ss := make([]string, len(this)) for i := range this { ss[i] = fmt.Sprintf("%d: %v", this[i].field, this[i].ext) } return "map[" + strings.Join(ss, ",") + "]" } func StringFromInternalExtension(m extendableProto) string { return StringFromExtensionsMap(m.extensionsWrite()) } func StringFromExtensionsMap(m map[int32]Extension) string { return newSortableExtensionsFromMap(m).String() } func StringFromExtensionsBytes(ext []byte) string { m, err := BytesToExtensionsMap(ext) if err != nil { panic(err) } return StringFromExtensionsMap(m) } func EncodeInternalExtension(m extendableProto, data []byte) (n int, err error) { return EncodeExtensionMap(m.extensionsWrite(), data) } func EncodeInternalExtensionBackwards(m extendableProto, data []byte) (n int, err error) { return EncodeExtensionMapBackwards(m.extensionsWrite(), data) } func EncodeExtensionMap(m map[int32]Extension, data []byte) (n int, err error) { o := 0 for _, e := range m { if err := e.Encode(); err != nil { return 0, err } n := copy(data[o:], e.enc) if n != len(e.enc) { return 0, io.ErrShortBuffer } o += n } return o, nil } func EncodeExtensionMapBackwards(m map[int32]Extension, data []byte) (n int, err error) { o := 0 end := len(data) for _, e := range m { if err := e.Encode(); err != nil { return 0, err } n := copy(data[end-len(e.enc):], e.enc) if n != len(e.enc) { return 0, io.ErrShortBuffer } end -= n o += n } return o, nil } func GetRawExtension(m map[int32]Extension, id int32) ([]byte, error) { e := m[id] if err := e.Encode(); err != nil { return nil, err } return e.enc, nil } func size(buf []byte, wire int) (int, error) { switch wire { case WireVarint: _, n := DecodeVarint(buf) return n, nil case WireFixed64: return 8, nil case WireBytes: v, n := DecodeVarint(buf) return int(v) + n, nil case WireFixed32: return 4, nil case WireStartGroup: offset := 0 for { u, n := DecodeVarint(buf[offset:]) fwire := int(u & 0x7) offset += n if fwire == WireEndGroup { return offset, nil } s, err := size(buf[offset:], wire) if err != nil { return 0, err } offset += s } } return 0, fmt.Errorf("proto: can't get size for unknown wire type %d", wire) } func BytesToExtensionsMap(buf []byte) (map[int32]Extension, error) { m := make(map[int32]Extension) i := 0 for i < len(buf) { tag, n := DecodeVarint(buf[i:]) if n <= 0 { return nil, fmt.Errorf("unable to decode varint") } fieldNum := int32(tag >> 3) wireType := int(tag & 0x7) l, err := size(buf[i+n:], wireType) if err != nil { return nil, err } end := i + int(l) + n m[int32(fieldNum)] = Extension{enc: buf[i:end]} i = end } return m, nil } func NewExtension(e []byte) Extension { ee := Extension{enc: make([]byte, len(e))} copy(ee.enc, e) return ee } func AppendExtension(e Message, tag int32, buf []byte) { if ee, eok := e.(extensionsBytes); eok { ext := ee.GetExtensions() *ext = append(*ext, buf...) return } if ee, eok := e.(extendableProto); eok { m := ee.extensionsWrite() ext := m[int32(tag)] // may be missing ext.enc = append(ext.enc, buf...) m[int32(tag)] = ext } } func encodeExtension(extension *ExtensionDesc, value interface{}) ([]byte, error) { u := getMarshalInfo(reflect.TypeOf(extension.ExtendedType)) ei := u.getExtElemInfo(extension) v := value p := toAddrPointer(&v, ei.isptr) siz := ei.sizer(p, SizeVarint(ei.wiretag)) buf := make([]byte, 0, siz) return ei.marshaler(buf, p, ei.wiretag, false) } func decodeExtensionFromBytes(extension *ExtensionDesc, buf []byte) (interface{}, error) { o := 0 for o < len(buf) { tag, n := DecodeVarint((buf)[o:]) fieldNum := int32(tag >> 3) wireType := int(tag & 0x7) if o+n > len(buf) { return nil, fmt.Errorf("unable to decode extension") } l, err := size((buf)[o+n:], wireType) if err != nil { return nil, err } if int32(fieldNum) == extension.Field { if o+n+l > len(buf) { return nil, fmt.Errorf("unable to decode extension") } v, err := decodeExtension((buf)[o:o+n+l], extension) if err != nil { return nil, err } return v, nil } o += n + l } return defaultExtensionValue(extension) } func (this *Extension) Encode() error { if this.enc == nil { var err error this.enc, err = encodeExtension(this.desc, this.value) if err != nil { return err } } return nil } func (this Extension) GoString() string { if err := this.Encode(); err != nil { return fmt.Sprintf("error encoding extension: %v", err) } return fmt.Sprintf("proto.NewExtension(%#v)", this.enc) } func SetUnsafeExtension(pb Message, fieldNum int32, value interface{}) error { typ := reflect.TypeOf(pb).Elem() ext, ok := extensionMaps[typ] if !ok { return fmt.Errorf("proto: bad extended type; %s is not extendable", typ.String()) } desc, ok := ext[fieldNum] if !ok { return errors.New("proto: bad extension number; not in declared ranges") } return SetExtension(pb, desc, value) } func GetUnsafeExtension(pb Message, fieldNum int32) (interface{}, error) { typ := reflect.TypeOf(pb).Elem() ext, ok := extensionMaps[typ] if !ok { return nil, fmt.Errorf("proto: bad extended type; %s is not extendable", typ.String()) } desc, ok := ext[fieldNum] if !ok { return nil, fmt.Errorf("unregistered field number %d", fieldNum) } return GetExtension(pb, desc) } func NewUnsafeXXX_InternalExtensions(m map[int32]Extension) XXX_InternalExtensions { x := &XXX_InternalExtensions{ p: new(struct { mu sync.Mutex extensionMap map[int32]Extension }), } x.p.extensionMap = m return *x } func GetUnsafeExtensionsMap(extendable Message) map[int32]Extension { pb := extendable.(extendableProto) return pb.extensionsWrite() } func deleteExtension(pb extensionsBytes, theFieldNum int32, offset int) int { ext := pb.GetExtensions() for offset < len(*ext) { tag, n1 := DecodeVarint((*ext)[offset:]) fieldNum := int32(tag >> 3) wireType := int(tag & 0x7) n2, err := size((*ext)[offset+n1:], wireType) if err != nil { panic(err) } newOffset := offset + n1 + n2 if fieldNum == theFieldNum { *ext = append((*ext)[:offset], (*ext)[newOffset:]...) return offset } offset = newOffset } return -1 }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/gogo/protobuf/proto/pointer_reflect_gogo.go
vendor/github.com/gogo/protobuf/proto/pointer_reflect_gogo.go
// Protocol Buffers for Go with Gadgets // // Copyright (c) 2018, The GoGo Authors. All rights reserved. // http://github.com/gogo/protobuf // // Redistribution and use in source and binary forms, with or without // modification, are permitted provided that the following conditions are // met: // // * Redistributions of source code must retain the above copyright // notice, this list of conditions and the following disclaimer. // * Redistributions in binary form must reproduce the above // copyright notice, this list of conditions and the following disclaimer // in the documentation and/or other materials provided with the // distribution. // // THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS // "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT // LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR // A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT // OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, // SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT // LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, // DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY // THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT // (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE // OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. // +build purego appengine js // This file contains an implementation of proto field accesses using package reflect. // It is slower than the code in pointer_unsafe.go but it avoids package unsafe and can // be used on App Engine. package proto import ( "reflect" ) // TODO: untested, so probably incorrect. func (p pointer) getRef() pointer { return pointer{v: p.v.Addr()} } func (p pointer) appendRef(v pointer, typ reflect.Type) { slice := p.getSlice(typ) elem := v.asPointerTo(typ).Elem() newSlice := reflect.Append(slice, elem) slice.Set(newSlice) } func (p pointer) getSlice(typ reflect.Type) reflect.Value { sliceTyp := reflect.SliceOf(typ) slice := p.asPointerTo(sliceTyp) slice = slice.Elem() return slice }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/gogo/protobuf/proto/table_unmarshal_gogo.go
vendor/github.com/gogo/protobuf/proto/table_unmarshal_gogo.go
// Protocol Buffers for Go with Gadgets // // Copyright (c) 2018, The GoGo Authors. All rights reserved. // http://github.com/gogo/protobuf // // Redistribution and use in source and binary forms, with or without // modification, are permitted provided that the following conditions are // met: // // * Redistributions of source code must retain the above copyright // notice, this list of conditions and the following disclaimer. // * Redistributions in binary form must reproduce the above // copyright notice, this list of conditions and the following disclaimer // in the documentation and/or other materials provided with the // distribution. // // THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS // "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT // LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR // A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT // OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, // SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT // LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, // DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY // THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT // (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE // OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. package proto import ( "io" "reflect" ) func makeUnmarshalMessage(sub *unmarshalInfo, name string) unmarshaler { return func(b []byte, f pointer, w int) ([]byte, error) { if w != WireBytes { return nil, errInternalBadWireType } x, n := decodeVarint(b) if n == 0 { return nil, io.ErrUnexpectedEOF } b = b[n:] if x > uint64(len(b)) { return nil, io.ErrUnexpectedEOF } // First read the message field to see if something is there. // The semantics of multiple submessages are weird. Instead of // the last one winning (as it is for all other fields), multiple // submessages are merged. v := f // gogo: changed from v := f.getPointer() if v.isNil() { v = valToPointer(reflect.New(sub.typ)) f.setPointer(v) } err := sub.unmarshal(v, b[:x]) if err != nil { if r, ok := err.(*RequiredNotSetError); ok { r.field = name + "." + r.field } else { return nil, err } } return b[x:], err } } func makeUnmarshalMessageSlice(sub *unmarshalInfo, name string) unmarshaler { return func(b []byte, f pointer, w int) ([]byte, error) { if w != WireBytes { return nil, errInternalBadWireType } x, n := decodeVarint(b) if n == 0 { return nil, io.ErrUnexpectedEOF } b = b[n:] if x > uint64(len(b)) { return nil, io.ErrUnexpectedEOF } v := valToPointer(reflect.New(sub.typ)) err := sub.unmarshal(v, b[:x]) if err != nil { if r, ok := err.(*RequiredNotSetError); ok { r.field = name + "." + r.field } else { return nil, err } } f.appendRef(v, sub.typ) // gogo: changed from f.appendPointer(v) return b[x:], err } } func makeUnmarshalCustomPtr(sub *unmarshalInfo, name string) unmarshaler { return func(b []byte, f pointer, w int) ([]byte, error) { if w != WireBytes { return nil, errInternalBadWireType } x, n := decodeVarint(b) if n == 0 { return nil, io.ErrUnexpectedEOF } b = b[n:] if x > uint64(len(b)) { return nil, io.ErrUnexpectedEOF } s := f.asPointerTo(reflect.PtrTo(sub.typ)).Elem() s.Set(reflect.New(sub.typ)) m := s.Interface().(custom) if err := m.Unmarshal(b[:x]); err != nil { return nil, err } return b[x:], nil } } func makeUnmarshalCustomSlice(sub *unmarshalInfo, name string) unmarshaler { return func(b []byte, f pointer, w int) ([]byte, error) { if w != WireBytes { return nil, errInternalBadWireType } x, n := decodeVarint(b) if n == 0 { return nil, io.ErrUnexpectedEOF } b = b[n:] if x > uint64(len(b)) { return nil, io.ErrUnexpectedEOF } m := reflect.New(sub.typ) c := m.Interface().(custom) if err := c.Unmarshal(b[:x]); err != nil { return nil, err } v := valToPointer(m) f.appendRef(v, sub.typ) return b[x:], nil } } func makeUnmarshalCustom(sub *unmarshalInfo, name string) unmarshaler { return func(b []byte, f pointer, w int) ([]byte, error) { if w != WireBytes { return nil, errInternalBadWireType } x, n := decodeVarint(b) if n == 0 { return nil, io.ErrUnexpectedEOF } b = b[n:] if x > uint64(len(b)) { return nil, io.ErrUnexpectedEOF } m := f.asPointerTo(sub.typ).Interface().(custom) if err := m.Unmarshal(b[:x]); err != nil { return nil, err } return b[x:], nil } } func makeUnmarshalTime(sub *unmarshalInfo, name string) unmarshaler { return func(b []byte, f pointer, w int) ([]byte, error) { if w != WireBytes { return nil, errInternalBadWireType } x, n := decodeVarint(b) if n == 0 { return nil, io.ErrUnexpectedEOF } b = b[n:] if x > uint64(len(b)) { return nil, io.ErrUnexpectedEOF } m := &timestamp{} if err := Unmarshal(b[:x], m); err != nil { return nil, err } t, err := timestampFromProto(m) if err != nil { return nil, err } s := f.asPointerTo(sub.typ).Elem() s.Set(reflect.ValueOf(t)) return b[x:], nil } } func makeUnmarshalTimePtr(sub *unmarshalInfo, name string) unmarshaler { return func(b []byte, f pointer, w int) ([]byte, error) { if w != WireBytes { return nil, errInternalBadWireType } x, n := decodeVarint(b) if n == 0 { return nil, io.ErrUnexpectedEOF } b = b[n:] if x > uint64(len(b)) { return nil, io.ErrUnexpectedEOF } m := &timestamp{} if err := Unmarshal(b[:x], m); err != nil { return nil, err } t, err := timestampFromProto(m) if err != nil { return nil, err } s := f.asPointerTo(reflect.PtrTo(sub.typ)).Elem() s.Set(reflect.ValueOf(&t)) return b[x:], nil } } func makeUnmarshalTimePtrSlice(sub *unmarshalInfo, name string) unmarshaler { return func(b []byte, f pointer, w int) ([]byte, error) { if w != WireBytes { return nil, errInternalBadWireType } x, n := decodeVarint(b) if n == 0 { return nil, io.ErrUnexpectedEOF } b = b[n:] if x > uint64(len(b)) { return nil, io.ErrUnexpectedEOF } m := &timestamp{} if err := Unmarshal(b[:x], m); err != nil { return nil, err } t, err := timestampFromProto(m) if err != nil { return nil, err } slice := f.getSlice(reflect.PtrTo(sub.typ)) newSlice := reflect.Append(slice, reflect.ValueOf(&t)) slice.Set(newSlice) return b[x:], nil } } func makeUnmarshalTimeSlice(sub *unmarshalInfo, name string) unmarshaler { return func(b []byte, f pointer, w int) ([]byte, error) { if w != WireBytes { return nil, errInternalBadWireType } x, n := decodeVarint(b) if n == 0 { return nil, io.ErrUnexpectedEOF } b = b[n:] if x > uint64(len(b)) { return nil, io.ErrUnexpectedEOF } m := &timestamp{} if err := Unmarshal(b[:x], m); err != nil { return nil, err } t, err := timestampFromProto(m) if err != nil { return nil, err } slice := f.getSlice(sub.typ) newSlice := reflect.Append(slice, reflect.ValueOf(t)) slice.Set(newSlice) return b[x:], nil } } func makeUnmarshalDurationPtr(sub *unmarshalInfo, name string) unmarshaler { return func(b []byte, f pointer, w int) ([]byte, error) { if w != WireBytes { return nil, errInternalBadWireType } x, n := decodeVarint(b) if n == 0 { return nil, io.ErrUnexpectedEOF } b = b[n:] if x > uint64(len(b)) { return nil, io.ErrUnexpectedEOF } m := &duration{} if err := Unmarshal(b[:x], m); err != nil { return nil, err } d, err := durationFromProto(m) if err != nil { return nil, err } s := f.asPointerTo(reflect.PtrTo(sub.typ)).Elem() s.Set(reflect.ValueOf(&d)) return b[x:], nil } } func makeUnmarshalDuration(sub *unmarshalInfo, name string) unmarshaler { return func(b []byte, f pointer, w int) ([]byte, error) { if w != WireBytes { return nil, errInternalBadWireType } x, n := decodeVarint(b) if n == 0 { return nil, io.ErrUnexpectedEOF } b = b[n:] if x > uint64(len(b)) { return nil, io.ErrUnexpectedEOF } m := &duration{} if err := Unmarshal(b[:x], m); err != nil { return nil, err } d, err := durationFromProto(m) if err != nil { return nil, err } s := f.asPointerTo(sub.typ).Elem() s.Set(reflect.ValueOf(d)) return b[x:], nil } } func makeUnmarshalDurationPtrSlice(sub *unmarshalInfo, name string) unmarshaler { return func(b []byte, f pointer, w int) ([]byte, error) { if w != WireBytes { return nil, errInternalBadWireType } x, n := decodeVarint(b) if n == 0 { return nil, io.ErrUnexpectedEOF } b = b[n:] if x > uint64(len(b)) { return nil, io.ErrUnexpectedEOF } m := &duration{} if err := Unmarshal(b[:x], m); err != nil { return nil, err } d, err := durationFromProto(m) if err != nil { return nil, err } slice := f.getSlice(reflect.PtrTo(sub.typ)) newSlice := reflect.Append(slice, reflect.ValueOf(&d)) slice.Set(newSlice) return b[x:], nil } } func makeUnmarshalDurationSlice(sub *unmarshalInfo, name string) unmarshaler { return func(b []byte, f pointer, w int) ([]byte, error) { if w != WireBytes { return nil, errInternalBadWireType } x, n := decodeVarint(b) if n == 0 { return nil, io.ErrUnexpectedEOF } b = b[n:] if x > uint64(len(b)) { return nil, io.ErrUnexpectedEOF } m := &duration{} if err := Unmarshal(b[:x], m); err != nil { return nil, err } d, err := durationFromProto(m) if err != nil { return nil, err } slice := f.getSlice(sub.typ) newSlice := reflect.Append(slice, reflect.ValueOf(d)) slice.Set(newSlice) return b[x:], nil } }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/gogo/protobuf/proto/table_marshal_gogo.go
vendor/github.com/gogo/protobuf/proto/table_marshal_gogo.go
// Protocol Buffers for Go with Gadgets // // Copyright (c) 2018, The GoGo Authors. All rights reserved. // http://github.com/gogo/protobuf // // Redistribution and use in source and binary forms, with or without // modification, are permitted provided that the following conditions are // met: // // * Redistributions of source code must retain the above copyright // notice, this list of conditions and the following disclaimer. // * Redistributions in binary form must reproduce the above // copyright notice, this list of conditions and the following disclaimer // in the documentation and/or other materials provided with the // distribution. // // THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS // "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT // LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR // A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT // OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, // SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT // LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, // DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY // THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT // (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE // OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. package proto import ( "reflect" "time" ) // makeMessageRefMarshaler differs a bit from makeMessageMarshaler // It marshal a message T instead of a *T func makeMessageRefMarshaler(u *marshalInfo) (sizer, marshaler) { return func(ptr pointer, tagsize int) int { siz := u.size(ptr) return siz + SizeVarint(uint64(siz)) + tagsize }, func(b []byte, ptr pointer, wiretag uint64, deterministic bool) ([]byte, error) { b = appendVarint(b, wiretag) siz := u.cachedsize(ptr) b = appendVarint(b, uint64(siz)) return u.marshal(b, ptr, deterministic) } } // makeMessageRefSliceMarshaler differs quite a lot from makeMessageSliceMarshaler // It marshals a slice of messages []T instead of []*T func makeMessageRefSliceMarshaler(u *marshalInfo) (sizer, marshaler) { return func(ptr pointer, tagsize int) int { s := ptr.getSlice(u.typ) n := 0 for i := 0; i < s.Len(); i++ { elem := s.Index(i) e := elem.Interface() v := toAddrPointer(&e, false) siz := u.size(v) n += siz + SizeVarint(uint64(siz)) + tagsize } return n }, func(b []byte, ptr pointer, wiretag uint64, deterministic bool) ([]byte, error) { s := ptr.getSlice(u.typ) var err, errreq error for i := 0; i < s.Len(); i++ { elem := s.Index(i) e := elem.Interface() v := toAddrPointer(&e, false) b = appendVarint(b, wiretag) siz := u.size(v) b = appendVarint(b, uint64(siz)) b, err = u.marshal(b, v, deterministic) if err != nil { if _, ok := err.(*RequiredNotSetError); ok { // Required field in submessage is not set. // We record the error but keep going, to give a complete marshaling. if errreq == nil { errreq = err } continue } if err == ErrNil { err = errRepeatedHasNil } return b, err } } return b, errreq } } func makeCustomPtrMarshaler(u *marshalInfo) (sizer, marshaler) { return func(ptr pointer, tagsize int) int { if ptr.isNil() { return 0 } m := ptr.asPointerTo(reflect.PtrTo(u.typ)).Elem().Interface().(custom) siz := m.Size() return tagsize + SizeVarint(uint64(siz)) + siz }, func(b []byte, ptr pointer, wiretag uint64, deterministic bool) ([]byte, error) { if ptr.isNil() { return b, nil } m := ptr.asPointerTo(reflect.PtrTo(u.typ)).Elem().Interface().(custom) siz := m.Size() buf, err := m.Marshal() if err != nil { return nil, err } b = appendVarint(b, wiretag) b = appendVarint(b, uint64(siz)) b = append(b, buf...) return b, nil } } func makeCustomMarshaler(u *marshalInfo) (sizer, marshaler) { return func(ptr pointer, tagsize int) int { m := ptr.asPointerTo(u.typ).Interface().(custom) siz := m.Size() return tagsize + SizeVarint(uint64(siz)) + siz }, func(b []byte, ptr pointer, wiretag uint64, deterministic bool) ([]byte, error) { m := ptr.asPointerTo(u.typ).Interface().(custom) siz := m.Size() buf, err := m.Marshal() if err != nil { return nil, err } b = appendVarint(b, wiretag) b = appendVarint(b, uint64(siz)) b = append(b, buf...) return b, nil } } func makeTimeMarshaler(u *marshalInfo) (sizer, marshaler) { return func(ptr pointer, tagsize int) int { t := ptr.asPointerTo(u.typ).Interface().(*time.Time) ts, err := timestampProto(*t) if err != nil { return 0 } siz := Size(ts) return tagsize + SizeVarint(uint64(siz)) + siz }, func(b []byte, ptr pointer, wiretag uint64, deterministic bool) ([]byte, error) { t := ptr.asPointerTo(u.typ).Interface().(*time.Time) ts, err := timestampProto(*t) if err != nil { return nil, err } buf, err := Marshal(ts) if err != nil { return nil, err } b = appendVarint(b, wiretag) b = appendVarint(b, uint64(len(buf))) b = append(b, buf...) return b, nil } } func makeTimePtrMarshaler(u *marshalInfo) (sizer, marshaler) { return func(ptr pointer, tagsize int) int { if ptr.isNil() { return 0 } t := ptr.asPointerTo(reflect.PtrTo(u.typ)).Elem().Interface().(*time.Time) ts, err := timestampProto(*t) if err != nil { return 0 } siz := Size(ts) return tagsize + SizeVarint(uint64(siz)) + siz }, func(b []byte, ptr pointer, wiretag uint64, deterministic bool) ([]byte, error) { if ptr.isNil() { return b, nil } t := ptr.asPointerTo(reflect.PtrTo(u.typ)).Elem().Interface().(*time.Time) ts, err := timestampProto(*t) if err != nil { return nil, err } buf, err := Marshal(ts) if err != nil { return nil, err } b = appendVarint(b, wiretag) b = appendVarint(b, uint64(len(buf))) b = append(b, buf...) return b, nil } } func makeTimeSliceMarshaler(u *marshalInfo) (sizer, marshaler) { return func(ptr pointer, tagsize int) int { s := ptr.getSlice(u.typ) n := 0 for i := 0; i < s.Len(); i++ { elem := s.Index(i) t := elem.Interface().(time.Time) ts, err := timestampProto(t) if err != nil { return 0 } siz := Size(ts) n += siz + SizeVarint(uint64(siz)) + tagsize } return n }, func(b []byte, ptr pointer, wiretag uint64, deterministic bool) ([]byte, error) { s := ptr.getSlice(u.typ) for i := 0; i < s.Len(); i++ { elem := s.Index(i) t := elem.Interface().(time.Time) ts, err := timestampProto(t) if err != nil { return nil, err } siz := Size(ts) buf, err := Marshal(ts) if err != nil { return nil, err } b = appendVarint(b, wiretag) b = appendVarint(b, uint64(siz)) b = append(b, buf...) } return b, nil } } func makeTimePtrSliceMarshaler(u *marshalInfo) (sizer, marshaler) { return func(ptr pointer, tagsize int) int { s := ptr.getSlice(reflect.PtrTo(u.typ)) n := 0 for i := 0; i < s.Len(); i++ { elem := s.Index(i) t := elem.Interface().(*time.Time) ts, err := timestampProto(*t) if err != nil { return 0 } siz := Size(ts) n += siz + SizeVarint(uint64(siz)) + tagsize } return n }, func(b []byte, ptr pointer, wiretag uint64, deterministic bool) ([]byte, error) { s := ptr.getSlice(reflect.PtrTo(u.typ)) for i := 0; i < s.Len(); i++ { elem := s.Index(i) t := elem.Interface().(*time.Time) ts, err := timestampProto(*t) if err != nil { return nil, err } siz := Size(ts) buf, err := Marshal(ts) if err != nil { return nil, err } b = appendVarint(b, wiretag) b = appendVarint(b, uint64(siz)) b = append(b, buf...) } return b, nil } } func makeDurationMarshaler(u *marshalInfo) (sizer, marshaler) { return func(ptr pointer, tagsize int) int { d := ptr.asPointerTo(u.typ).Interface().(*time.Duration) dur := durationProto(*d) siz := Size(dur) return tagsize + SizeVarint(uint64(siz)) + siz }, func(b []byte, ptr pointer, wiretag uint64, deterministic bool) ([]byte, error) { d := ptr.asPointerTo(u.typ).Interface().(*time.Duration) dur := durationProto(*d) buf, err := Marshal(dur) if err != nil { return nil, err } b = appendVarint(b, wiretag) b = appendVarint(b, uint64(len(buf))) b = append(b, buf...) return b, nil } } func makeDurationPtrMarshaler(u *marshalInfo) (sizer, marshaler) { return func(ptr pointer, tagsize int) int { if ptr.isNil() { return 0 } d := ptr.asPointerTo(reflect.PtrTo(u.typ)).Elem().Interface().(*time.Duration) dur := durationProto(*d) siz := Size(dur) return tagsize + SizeVarint(uint64(siz)) + siz }, func(b []byte, ptr pointer, wiretag uint64, deterministic bool) ([]byte, error) { if ptr.isNil() { return b, nil } d := ptr.asPointerTo(reflect.PtrTo(u.typ)).Elem().Interface().(*time.Duration) dur := durationProto(*d) buf, err := Marshal(dur) if err != nil { return nil, err } b = appendVarint(b, wiretag) b = appendVarint(b, uint64(len(buf))) b = append(b, buf...) return b, nil } } func makeDurationSliceMarshaler(u *marshalInfo) (sizer, marshaler) { return func(ptr pointer, tagsize int) int { s := ptr.getSlice(u.typ) n := 0 for i := 0; i < s.Len(); i++ { elem := s.Index(i) d := elem.Interface().(time.Duration) dur := durationProto(d) siz := Size(dur) n += siz + SizeVarint(uint64(siz)) + tagsize } return n }, func(b []byte, ptr pointer, wiretag uint64, deterministic bool) ([]byte, error) { s := ptr.getSlice(u.typ) for i := 0; i < s.Len(); i++ { elem := s.Index(i) d := elem.Interface().(time.Duration) dur := durationProto(d) siz := Size(dur) buf, err := Marshal(dur) if err != nil { return nil, err } b = appendVarint(b, wiretag) b = appendVarint(b, uint64(siz)) b = append(b, buf...) } return b, nil } } func makeDurationPtrSliceMarshaler(u *marshalInfo) (sizer, marshaler) { return func(ptr pointer, tagsize int) int { s := ptr.getSlice(reflect.PtrTo(u.typ)) n := 0 for i := 0; i < s.Len(); i++ { elem := s.Index(i) d := elem.Interface().(*time.Duration) dur := durationProto(*d) siz := Size(dur) n += siz + SizeVarint(uint64(siz)) + tagsize } return n }, func(b []byte, ptr pointer, wiretag uint64, deterministic bool) ([]byte, error) { s := ptr.getSlice(reflect.PtrTo(u.typ)) for i := 0; i < s.Len(); i++ { elem := s.Index(i) d := elem.Interface().(*time.Duration) dur := durationProto(*d) siz := Size(dur) buf, err := Marshal(dur) if err != nil { return nil, err } b = appendVarint(b, wiretag) b = appendVarint(b, uint64(siz)) b = append(b, buf...) } return b, nil } }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/gogo/protobuf/proto/pointer_reflect.go
vendor/github.com/gogo/protobuf/proto/pointer_reflect.go
// Go support for Protocol Buffers - Google's data interchange format // // Copyright 2012 The Go Authors. All rights reserved. // https://github.com/golang/protobuf // // Redistribution and use in source and binary forms, with or without // modification, are permitted provided that the following conditions are // met: // // * Redistributions of source code must retain the above copyright // notice, this list of conditions and the following disclaimer. // * Redistributions in binary form must reproduce the above // copyright notice, this list of conditions and the following disclaimer // in the documentation and/or other materials provided with the // distribution. // * Neither the name of Google Inc. nor the names of its // contributors may be used to endorse or promote products derived from // this software without specific prior written permission. // // THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS // "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT // LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR // A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT // OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, // SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT // LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, // DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY // THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT // (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE // OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. // +build purego appengine js // This file contains an implementation of proto field accesses using package reflect. // It is slower than the code in pointer_unsafe.go but it avoids package unsafe and can // be used on App Engine. package proto import ( "reflect" "sync" ) const unsafeAllowed = false // A field identifies a field in a struct, accessible from a pointer. // In this implementation, a field is identified by the sequence of field indices // passed to reflect's FieldByIndex. type field []int // toField returns a field equivalent to the given reflect field. func toField(f *reflect.StructField) field { return f.Index } // invalidField is an invalid field identifier. var invalidField = field(nil) // zeroField is a noop when calling pointer.offset. var zeroField = field([]int{}) // IsValid reports whether the field identifier is valid. func (f field) IsValid() bool { return f != nil } // The pointer type is for the table-driven decoder. // The implementation here uses a reflect.Value of pointer type to // create a generic pointer. In pointer_unsafe.go we use unsafe // instead of reflect to implement the same (but faster) interface. type pointer struct { v reflect.Value } // toPointer converts an interface of pointer type to a pointer // that points to the same target. func toPointer(i *Message) pointer { return pointer{v: reflect.ValueOf(*i)} } // toAddrPointer converts an interface to a pointer that points to // the interface data. func toAddrPointer(i *interface{}, isptr bool) pointer { v := reflect.ValueOf(*i) u := reflect.New(v.Type()) u.Elem().Set(v) return pointer{v: u} } // valToPointer converts v to a pointer. v must be of pointer type. func valToPointer(v reflect.Value) pointer { return pointer{v: v} } // offset converts from a pointer to a structure to a pointer to // one of its fields. func (p pointer) offset(f field) pointer { return pointer{v: p.v.Elem().FieldByIndex(f).Addr()} } func (p pointer) isNil() bool { return p.v.IsNil() } // grow updates the slice s in place to make it one element longer. // s must be addressable. // Returns the (addressable) new element. func grow(s reflect.Value) reflect.Value { n, m := s.Len(), s.Cap() if n < m { s.SetLen(n + 1) } else { s.Set(reflect.Append(s, reflect.Zero(s.Type().Elem()))) } return s.Index(n) } func (p pointer) toInt64() *int64 { return p.v.Interface().(*int64) } func (p pointer) toInt64Ptr() **int64 { return p.v.Interface().(**int64) } func (p pointer) toInt64Slice() *[]int64 { return p.v.Interface().(*[]int64) } var int32ptr = reflect.TypeOf((*int32)(nil)) func (p pointer) toInt32() *int32 { return p.v.Convert(int32ptr).Interface().(*int32) } // The toInt32Ptr/Slice methods don't work because of enums. // Instead, we must use set/get methods for the int32ptr/slice case. /* func (p pointer) toInt32Ptr() **int32 { return p.v.Interface().(**int32) } func (p pointer) toInt32Slice() *[]int32 { return p.v.Interface().(*[]int32) } */ func (p pointer) getInt32Ptr() *int32 { if p.v.Type().Elem().Elem() == reflect.TypeOf(int32(0)) { // raw int32 type return p.v.Elem().Interface().(*int32) } // an enum return p.v.Elem().Convert(int32PtrType).Interface().(*int32) } func (p pointer) setInt32Ptr(v int32) { // Allocate value in a *int32. Possibly convert that to a *enum. // Then assign it to a **int32 or **enum. // Note: we can convert *int32 to *enum, but we can't convert // **int32 to **enum! p.v.Elem().Set(reflect.ValueOf(&v).Convert(p.v.Type().Elem())) } // getInt32Slice copies []int32 from p as a new slice. // This behavior differs from the implementation in pointer_unsafe.go. func (p pointer) getInt32Slice() []int32 { if p.v.Type().Elem().Elem() == reflect.TypeOf(int32(0)) { // raw int32 type return p.v.Elem().Interface().([]int32) } // an enum // Allocate a []int32, then assign []enum's values into it. // Note: we can't convert []enum to []int32. slice := p.v.Elem() s := make([]int32, slice.Len()) for i := 0; i < slice.Len(); i++ { s[i] = int32(slice.Index(i).Int()) } return s } // setInt32Slice copies []int32 into p as a new slice. // This behavior differs from the implementation in pointer_unsafe.go. func (p pointer) setInt32Slice(v []int32) { if p.v.Type().Elem().Elem() == reflect.TypeOf(int32(0)) { // raw int32 type p.v.Elem().Set(reflect.ValueOf(v)) return } // an enum // Allocate a []enum, then assign []int32's values into it. // Note: we can't convert []enum to []int32. slice := reflect.MakeSlice(p.v.Type().Elem(), len(v), cap(v)) for i, x := range v { slice.Index(i).SetInt(int64(x)) } p.v.Elem().Set(slice) } func (p pointer) appendInt32Slice(v int32) { grow(p.v.Elem()).SetInt(int64(v)) } func (p pointer) toUint64() *uint64 { return p.v.Interface().(*uint64) } func (p pointer) toUint64Ptr() **uint64 { return p.v.Interface().(**uint64) } func (p pointer) toUint64Slice() *[]uint64 { return p.v.Interface().(*[]uint64) } func (p pointer) toUint32() *uint32 { return p.v.Interface().(*uint32) } func (p pointer) toUint32Ptr() **uint32 { return p.v.Interface().(**uint32) } func (p pointer) toUint32Slice() *[]uint32 { return p.v.Interface().(*[]uint32) } func (p pointer) toBool() *bool { return p.v.Interface().(*bool) } func (p pointer) toBoolPtr() **bool { return p.v.Interface().(**bool) } func (p pointer) toBoolSlice() *[]bool { return p.v.Interface().(*[]bool) } func (p pointer) toFloat64() *float64 { return p.v.Interface().(*float64) } func (p pointer) toFloat64Ptr() **float64 { return p.v.Interface().(**float64) } func (p pointer) toFloat64Slice() *[]float64 { return p.v.Interface().(*[]float64) } func (p pointer) toFloat32() *float32 { return p.v.Interface().(*float32) } func (p pointer) toFloat32Ptr() **float32 { return p.v.Interface().(**float32) } func (p pointer) toFloat32Slice() *[]float32 { return p.v.Interface().(*[]float32) } func (p pointer) toString() *string { return p.v.Interface().(*string) } func (p pointer) toStringPtr() **string { return p.v.Interface().(**string) } func (p pointer) toStringSlice() *[]string { return p.v.Interface().(*[]string) } func (p pointer) toBytes() *[]byte { return p.v.Interface().(*[]byte) } func (p pointer) toBytesSlice() *[][]byte { return p.v.Interface().(*[][]byte) } func (p pointer) toExtensions() *XXX_InternalExtensions { return p.v.Interface().(*XXX_InternalExtensions) } func (p pointer) toOldExtensions() *map[int32]Extension { return p.v.Interface().(*map[int32]Extension) } func (p pointer) getPointer() pointer { return pointer{v: p.v.Elem()} } func (p pointer) setPointer(q pointer) { p.v.Elem().Set(q.v) } func (p pointer) appendPointer(q pointer) { grow(p.v.Elem()).Set(q.v) } // getPointerSlice copies []*T from p as a new []pointer. // This behavior differs from the implementation in pointer_unsafe.go. func (p pointer) getPointerSlice() []pointer { if p.v.IsNil() { return nil } n := p.v.Elem().Len() s := make([]pointer, n) for i := 0; i < n; i++ { s[i] = pointer{v: p.v.Elem().Index(i)} } return s } // setPointerSlice copies []pointer into p as a new []*T. // This behavior differs from the implementation in pointer_unsafe.go. func (p pointer) setPointerSlice(v []pointer) { if v == nil { p.v.Elem().Set(reflect.New(p.v.Elem().Type()).Elem()) return } s := reflect.MakeSlice(p.v.Elem().Type(), 0, len(v)) for _, p := range v { s = reflect.Append(s, p.v) } p.v.Elem().Set(s) } // getInterfacePointer returns a pointer that points to the // interface data of the interface pointed by p. func (p pointer) getInterfacePointer() pointer { if p.v.Elem().IsNil() { return pointer{v: p.v.Elem()} } return pointer{v: p.v.Elem().Elem().Elem().Field(0).Addr()} // *interface -> interface -> *struct -> struct } func (p pointer) asPointerTo(t reflect.Type) reflect.Value { // TODO: check that p.v.Type().Elem() == t? return p.v } func atomicLoadUnmarshalInfo(p **unmarshalInfo) *unmarshalInfo { atomicLock.Lock() defer atomicLock.Unlock() return *p } func atomicStoreUnmarshalInfo(p **unmarshalInfo, v *unmarshalInfo) { atomicLock.Lock() defer atomicLock.Unlock() *p = v } func atomicLoadMarshalInfo(p **marshalInfo) *marshalInfo { atomicLock.Lock() defer atomicLock.Unlock() return *p } func atomicStoreMarshalInfo(p **marshalInfo, v *marshalInfo) { atomicLock.Lock() defer atomicLock.Unlock() *p = v } func atomicLoadMergeInfo(p **mergeInfo) *mergeInfo { atomicLock.Lock() defer atomicLock.Unlock() return *p } func atomicStoreMergeInfo(p **mergeInfo, v *mergeInfo) { atomicLock.Lock() defer atomicLock.Unlock() *p = v } func atomicLoadDiscardInfo(p **discardInfo) *discardInfo { atomicLock.Lock() defer atomicLock.Unlock() return *p } func atomicStoreDiscardInfo(p **discardInfo, v *discardInfo) { atomicLock.Lock() defer atomicLock.Unlock() *p = v } var atomicLock sync.Mutex
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/gogo/protobuf/proto/wrappers_gogo.go
vendor/github.com/gogo/protobuf/proto/wrappers_gogo.go
// Protocol Buffers for Go with Gadgets // // Copyright (c) 2018, The GoGo Authors. All rights reserved. // http://github.com/gogo/protobuf // // Redistribution and use in source and binary forms, with or without // modification, are permitted provided that the following conditions are // met: // // * Redistributions of source code must retain the above copyright // notice, this list of conditions and the following disclaimer. // * Redistributions in binary form must reproduce the above // copyright notice, this list of conditions and the following disclaimer // in the documentation and/or other materials provided with the // distribution. // // THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS // "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT // LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR // A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT // OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, // SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT // LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, // DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY // THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT // (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE // OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. package proto type float64Value struct { Value float64 `protobuf:"fixed64,1,opt,name=value,proto3" json:"value,omitempty"` } func (m *float64Value) Reset() { *m = float64Value{} } func (*float64Value) ProtoMessage() {} func (*float64Value) String() string { return "float64<string>" } type float32Value struct { Value float32 `protobuf:"fixed32,1,opt,name=value,proto3" json:"value,omitempty"` } func (m *float32Value) Reset() { *m = float32Value{} } func (*float32Value) ProtoMessage() {} func (*float32Value) String() string { return "float32<string>" } type int64Value struct { Value int64 `protobuf:"varint,1,opt,name=value,proto3" json:"value,omitempty"` } func (m *int64Value) Reset() { *m = int64Value{} } func (*int64Value) ProtoMessage() {} func (*int64Value) String() string { return "int64<string>" } type uint64Value struct { Value uint64 `protobuf:"varint,1,opt,name=value,proto3" json:"value,omitempty"` } func (m *uint64Value) Reset() { *m = uint64Value{} } func (*uint64Value) ProtoMessage() {} func (*uint64Value) String() string { return "uint64<string>" } type int32Value struct { Value int32 `protobuf:"varint,1,opt,name=value,proto3" json:"value,omitempty"` } func (m *int32Value) Reset() { *m = int32Value{} } func (*int32Value) ProtoMessage() {} func (*int32Value) String() string { return "int32<string>" } type uint32Value struct { Value uint32 `protobuf:"varint,1,opt,name=value,proto3" json:"value,omitempty"` } func (m *uint32Value) Reset() { *m = uint32Value{} } func (*uint32Value) ProtoMessage() {} func (*uint32Value) String() string { return "uint32<string>" } type boolValue struct { Value bool `protobuf:"varint,1,opt,name=value,proto3" json:"value,omitempty"` } func (m *boolValue) Reset() { *m = boolValue{} } func (*boolValue) ProtoMessage() {} func (*boolValue) String() string { return "bool<string>" } type stringValue struct { Value string `protobuf:"bytes,1,opt,name=value,proto3" json:"value,omitempty"` } func (m *stringValue) Reset() { *m = stringValue{} } func (*stringValue) ProtoMessage() {} func (*stringValue) String() string { return "string<string>" } type bytesValue struct { Value []byte `protobuf:"bytes,1,opt,name=value,proto3" json:"value,omitempty"` } func (m *bytesValue) Reset() { *m = bytesValue{} } func (*bytesValue) ProtoMessage() {} func (*bytesValue) String() string { return "[]byte<string>" } func init() { RegisterType((*float64Value)(nil), "gogo.protobuf.proto.DoubleValue") RegisterType((*float32Value)(nil), "gogo.protobuf.proto.FloatValue") RegisterType((*int64Value)(nil), "gogo.protobuf.proto.Int64Value") RegisterType((*uint64Value)(nil), "gogo.protobuf.proto.UInt64Value") RegisterType((*int32Value)(nil), "gogo.protobuf.proto.Int32Value") RegisterType((*uint32Value)(nil), "gogo.protobuf.proto.UInt32Value") RegisterType((*boolValue)(nil), "gogo.protobuf.proto.BoolValue") RegisterType((*stringValue)(nil), "gogo.protobuf.proto.StringValue") RegisterType((*bytesValue)(nil), "gogo.protobuf.proto.BytesValue") }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/gogo/protobuf/proto/decode.go
vendor/github.com/gogo/protobuf/proto/decode.go
// Go support for Protocol Buffers - Google's data interchange format // // Copyright 2010 The Go Authors. All rights reserved. // https://github.com/golang/protobuf // // Redistribution and use in source and binary forms, with or without // modification, are permitted provided that the following conditions are // met: // // * Redistributions of source code must retain the above copyright // notice, this list of conditions and the following disclaimer. // * Redistributions in binary form must reproduce the above // copyright notice, this list of conditions and the following disclaimer // in the documentation and/or other materials provided with the // distribution. // * Neither the name of Google Inc. nor the names of its // contributors may be used to endorse or promote products derived from // this software without specific prior written permission. // // THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS // "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT // LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR // A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT // OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, // SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT // LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, // DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY // THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT // (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE // OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. package proto /* * Routines for decoding protocol buffer data to construct in-memory representations. */ import ( "errors" "fmt" "io" ) // errOverflow is returned when an integer is too large to be represented. var errOverflow = errors.New("proto: integer overflow") // ErrInternalBadWireType is returned by generated code when an incorrect // wire type is encountered. It does not get returned to user code. var ErrInternalBadWireType = errors.New("proto: internal error: bad wiretype for oneof") // DecodeVarint reads a varint-encoded integer from the slice. // It returns the integer and the number of bytes consumed, or // zero if there is not enough. // This is the format for the // int32, int64, uint32, uint64, bool, and enum // protocol buffer types. func DecodeVarint(buf []byte) (x uint64, n int) { for shift := uint(0); shift < 64; shift += 7 { if n >= len(buf) { return 0, 0 } b := uint64(buf[n]) n++ x |= (b & 0x7F) << shift if (b & 0x80) == 0 { return x, n } } // The number is too large to represent in a 64-bit value. return 0, 0 } func (p *Buffer) decodeVarintSlow() (x uint64, err error) { i := p.index l := len(p.buf) for shift := uint(0); shift < 64; shift += 7 { if i >= l { err = io.ErrUnexpectedEOF return } b := p.buf[i] i++ x |= (uint64(b) & 0x7F) << shift if b < 0x80 { p.index = i return } } // The number is too large to represent in a 64-bit value. err = errOverflow return } // DecodeVarint reads a varint-encoded integer from the Buffer. // This is the format for the // int32, int64, uint32, uint64, bool, and enum // protocol buffer types. func (p *Buffer) DecodeVarint() (x uint64, err error) { i := p.index buf := p.buf if i >= len(buf) { return 0, io.ErrUnexpectedEOF } else if buf[i] < 0x80 { p.index++ return uint64(buf[i]), nil } else if len(buf)-i < 10 { return p.decodeVarintSlow() } var b uint64 // we already checked the first byte x = uint64(buf[i]) - 0x80 i++ b = uint64(buf[i]) i++ x += b << 7 if b&0x80 == 0 { goto done } x -= 0x80 << 7 b = uint64(buf[i]) i++ x += b << 14 if b&0x80 == 0 { goto done } x -= 0x80 << 14 b = uint64(buf[i]) i++ x += b << 21 if b&0x80 == 0 { goto done } x -= 0x80 << 21 b = uint64(buf[i]) i++ x += b << 28 if b&0x80 == 0 { goto done } x -= 0x80 << 28 b = uint64(buf[i]) i++ x += b << 35 if b&0x80 == 0 { goto done } x -= 0x80 << 35 b = uint64(buf[i]) i++ x += b << 42 if b&0x80 == 0 { goto done } x -= 0x80 << 42 b = uint64(buf[i]) i++ x += b << 49 if b&0x80 == 0 { goto done } x -= 0x80 << 49 b = uint64(buf[i]) i++ x += b << 56 if b&0x80 == 0 { goto done } x -= 0x80 << 56 b = uint64(buf[i]) i++ x += b << 63 if b&0x80 == 0 { goto done } return 0, errOverflow done: p.index = i return x, nil } // DecodeFixed64 reads a 64-bit integer from the Buffer. // This is the format for the // fixed64, sfixed64, and double protocol buffer types. func (p *Buffer) DecodeFixed64() (x uint64, err error) { // x, err already 0 i := p.index + 8 if i < 0 || i > len(p.buf) { err = io.ErrUnexpectedEOF return } p.index = i x = uint64(p.buf[i-8]) x |= uint64(p.buf[i-7]) << 8 x |= uint64(p.buf[i-6]) << 16 x |= uint64(p.buf[i-5]) << 24 x |= uint64(p.buf[i-4]) << 32 x |= uint64(p.buf[i-3]) << 40 x |= uint64(p.buf[i-2]) << 48 x |= uint64(p.buf[i-1]) << 56 return } // DecodeFixed32 reads a 32-bit integer from the Buffer. // This is the format for the // fixed32, sfixed32, and float protocol buffer types. func (p *Buffer) DecodeFixed32() (x uint64, err error) { // x, err already 0 i := p.index + 4 if i < 0 || i > len(p.buf) { err = io.ErrUnexpectedEOF return } p.index = i x = uint64(p.buf[i-4]) x |= uint64(p.buf[i-3]) << 8 x |= uint64(p.buf[i-2]) << 16 x |= uint64(p.buf[i-1]) << 24 return } // DecodeZigzag64 reads a zigzag-encoded 64-bit integer // from the Buffer. // This is the format used for the sint64 protocol buffer type. func (p *Buffer) DecodeZigzag64() (x uint64, err error) { x, err = p.DecodeVarint() if err != nil { return } x = (x >> 1) ^ uint64((int64(x&1)<<63)>>63) return } // DecodeZigzag32 reads a zigzag-encoded 32-bit integer // from the Buffer. // This is the format used for the sint32 protocol buffer type. func (p *Buffer) DecodeZigzag32() (x uint64, err error) { x, err = p.DecodeVarint() if err != nil { return } x = uint64((uint32(x) >> 1) ^ uint32((int32(x&1)<<31)>>31)) return } // DecodeRawBytes reads a count-delimited byte buffer from the Buffer. // This is the format used for the bytes protocol buffer // type and for embedded messages. func (p *Buffer) DecodeRawBytes(alloc bool) (buf []byte, err error) { n, err := p.DecodeVarint() if err != nil { return nil, err } nb := int(n) if nb < 0 { return nil, fmt.Errorf("proto: bad byte length %d", nb) } end := p.index + nb if end < p.index || end > len(p.buf) { return nil, io.ErrUnexpectedEOF } if !alloc { // todo: check if can get more uses of alloc=false buf = p.buf[p.index:end] p.index += nb return } buf = make([]byte, nb) copy(buf, p.buf[p.index:]) p.index += nb return } // DecodeStringBytes reads an encoded string from the Buffer. // This is the format used for the proto2 string type. func (p *Buffer) DecodeStringBytes() (s string, err error) { buf, err := p.DecodeRawBytes(false) if err != nil { return } return string(buf), nil } // Unmarshaler is the interface representing objects that can // unmarshal themselves. The argument points to data that may be // overwritten, so implementations should not keep references to the // buffer. // Unmarshal implementations should not clear the receiver. // Any unmarshaled data should be merged into the receiver. // Callers of Unmarshal that do not want to retain existing data // should Reset the receiver before calling Unmarshal. type Unmarshaler interface { Unmarshal([]byte) error } // newUnmarshaler is the interface representing objects that can // unmarshal themselves. The semantics are identical to Unmarshaler. // // This exists to support protoc-gen-go generated messages. // The proto package will stop type-asserting to this interface in the future. // // DO NOT DEPEND ON THIS. type newUnmarshaler interface { XXX_Unmarshal([]byte) error } // Unmarshal parses the protocol buffer representation in buf and places the // decoded result in pb. If the struct underlying pb does not match // the data in buf, the results can be unpredictable. // // Unmarshal resets pb before starting to unmarshal, so any // existing data in pb is always removed. Use UnmarshalMerge // to preserve and append to existing data. func Unmarshal(buf []byte, pb Message) error { pb.Reset() if u, ok := pb.(newUnmarshaler); ok { return u.XXX_Unmarshal(buf) } if u, ok := pb.(Unmarshaler); ok { return u.Unmarshal(buf) } return NewBuffer(buf).Unmarshal(pb) } // UnmarshalMerge parses the protocol buffer representation in buf and // writes the decoded result to pb. If the struct underlying pb does not match // the data in buf, the results can be unpredictable. // // UnmarshalMerge merges into existing data in pb. // Most code should use Unmarshal instead. func UnmarshalMerge(buf []byte, pb Message) error { if u, ok := pb.(newUnmarshaler); ok { return u.XXX_Unmarshal(buf) } if u, ok := pb.(Unmarshaler); ok { // NOTE: The history of proto have unfortunately been inconsistent // whether Unmarshaler should or should not implicitly clear itself. // Some implementations do, most do not. // Thus, calling this here may or may not do what people want. // // See https://github.com/golang/protobuf/issues/424 return u.Unmarshal(buf) } return NewBuffer(buf).Unmarshal(pb) } // DecodeMessage reads a count-delimited message from the Buffer. func (p *Buffer) DecodeMessage(pb Message) error { enc, err := p.DecodeRawBytes(false) if err != nil { return err } return NewBuffer(enc).Unmarshal(pb) } // DecodeGroup reads a tag-delimited group from the Buffer. // StartGroup tag is already consumed. This function consumes // EndGroup tag. func (p *Buffer) DecodeGroup(pb Message) error { b := p.buf[p.index:] x, y := findEndGroup(b) if x < 0 { return io.ErrUnexpectedEOF } err := Unmarshal(b[:x], pb) p.index += y return err } // Unmarshal parses the protocol buffer representation in the // Buffer and places the decoded result in pb. If the struct // underlying pb does not match the data in the buffer, the results can be // unpredictable. // // Unlike proto.Unmarshal, this does not reset pb before starting to unmarshal. func (p *Buffer) Unmarshal(pb Message) error { // If the object can unmarshal itself, let it. if u, ok := pb.(newUnmarshaler); ok { err := u.XXX_Unmarshal(p.buf[p.index:]) p.index = len(p.buf) return err } if u, ok := pb.(Unmarshaler); ok { // NOTE: The history of proto have unfortunately been inconsistent // whether Unmarshaler should or should not implicitly clear itself. // Some implementations do, most do not. // Thus, calling this here may or may not do what people want. // // See https://github.com/golang/protobuf/issues/424 err := u.Unmarshal(p.buf[p.index:]) p.index = len(p.buf) return err } // Slow workaround for messages that aren't Unmarshalers. // This includes some hand-coded .pb.go files and // bootstrap protos. // TODO: fix all of those and then add Unmarshal to // the Message interface. Then: // The cast above and code below can be deleted. // The old unmarshaler can be deleted. // Clients can call Unmarshal directly (can already do that, actually). var info InternalMessageInfo err := info.Unmarshal(pb, p.buf[p.index:]) p.index = len(p.buf) return err }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/gogo/protobuf/proto/duration.go
vendor/github.com/gogo/protobuf/proto/duration.go
// Go support for Protocol Buffers - Google's data interchange format // // Copyright 2016 The Go Authors. All rights reserved. // https://github.com/golang/protobuf // // Redistribution and use in source and binary forms, with or without // modification, are permitted provided that the following conditions are // met: // // * Redistributions of source code must retain the above copyright // notice, this list of conditions and the following disclaimer. // * Redistributions in binary form must reproduce the above // copyright notice, this list of conditions and the following disclaimer // in the documentation and/or other materials provided with the // distribution. // * Neither the name of Google Inc. nor the names of its // contributors may be used to endorse or promote products derived from // this software without specific prior written permission. // // THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS // "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT // LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR // A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT // OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, // SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT // LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, // DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY // THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT // (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE // OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. package proto // This file implements conversions between google.protobuf.Duration // and time.Duration. import ( "errors" "fmt" "time" ) const ( // Range of a Duration in seconds, as specified in // google/protobuf/duration.proto. This is about 10,000 years in seconds. maxSeconds = int64(10000 * 365.25 * 24 * 60 * 60) minSeconds = -maxSeconds ) // validateDuration determines whether the Duration is valid according to the // definition in google/protobuf/duration.proto. A valid Duration // may still be too large to fit into a time.Duration (the range of Duration // is about 10,000 years, and the range of time.Duration is about 290). func validateDuration(d *duration) error { if d == nil { return errors.New("duration: nil Duration") } if d.Seconds < minSeconds || d.Seconds > maxSeconds { return fmt.Errorf("duration: %#v: seconds out of range", d) } if d.Nanos <= -1e9 || d.Nanos >= 1e9 { return fmt.Errorf("duration: %#v: nanos out of range", d) } // Seconds and Nanos must have the same sign, unless d.Nanos is zero. if (d.Seconds < 0 && d.Nanos > 0) || (d.Seconds > 0 && d.Nanos < 0) { return fmt.Errorf("duration: %#v: seconds and nanos have different signs", d) } return nil } // DurationFromProto converts a Duration to a time.Duration. DurationFromProto // returns an error if the Duration is invalid or is too large to be // represented in a time.Duration. func durationFromProto(p *duration) (time.Duration, error) { if err := validateDuration(p); err != nil { return 0, err } d := time.Duration(p.Seconds) * time.Second if int64(d/time.Second) != p.Seconds { return 0, fmt.Errorf("duration: %#v is out of range for time.Duration", p) } if p.Nanos != 0 { d += time.Duration(p.Nanos) if (d < 0) != (p.Nanos < 0) { return 0, fmt.Errorf("duration: %#v is out of range for time.Duration", p) } } return d, nil } // DurationProto converts a time.Duration to a Duration. func durationProto(d time.Duration) *duration { nanos := d.Nanoseconds() secs := nanos / 1e9 nanos -= secs * 1e9 return &duration{ Seconds: secs, Nanos: int32(nanos), } }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/gogo/protobuf/proto/duration_gogo.go
vendor/github.com/gogo/protobuf/proto/duration_gogo.go
// Protocol Buffers for Go with Gadgets // // Copyright (c) 2016, The GoGo Authors. All rights reserved. // http://github.com/gogo/protobuf // // Redistribution and use in source and binary forms, with or without // modification, are permitted provided that the following conditions are // met: // // * Redistributions of source code must retain the above copyright // notice, this list of conditions and the following disclaimer. // * Redistributions in binary form must reproduce the above // copyright notice, this list of conditions and the following disclaimer // in the documentation and/or other materials provided with the // distribution. // // THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS // "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT // LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR // A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT // OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, // SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT // LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, // DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY // THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT // (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE // OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. package proto import ( "reflect" "time" ) var durationType = reflect.TypeOf((*time.Duration)(nil)).Elem() type duration struct { Seconds int64 `protobuf:"varint,1,opt,name=seconds,proto3" json:"seconds,omitempty"` Nanos int32 `protobuf:"varint,2,opt,name=nanos,proto3" json:"nanos,omitempty"` } func (m *duration) Reset() { *m = duration{} } func (*duration) ProtoMessage() {} func (*duration) String() string { return "duration<string>" } func init() { RegisterType((*duration)(nil), "gogo.protobuf.proto.duration") }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/gogo/protobuf/proto/wrappers.go
vendor/github.com/gogo/protobuf/proto/wrappers.go
// Protocol Buffers for Go with Gadgets // // Copyright (c) 2018, The GoGo Authors. All rights reserved. // http://github.com/gogo/protobuf // // Redistribution and use in source and binary forms, with or without // modification, are permitted provided that the following conditions are // met: // // * Redistributions of source code must retain the above copyright // notice, this list of conditions and the following disclaimer. // * Redistributions in binary form must reproduce the above // copyright notice, this list of conditions and the following disclaimer // in the documentation and/or other materials provided with the // distribution. // // THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS // "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT // LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR // A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT // OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, // SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT // LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, // DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY // THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT // (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE // OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. package proto import ( "io" "reflect" ) func makeStdDoubleValueMarshaler(u *marshalInfo) (sizer, marshaler) { return func(ptr pointer, tagsize int) int { t := ptr.asPointerTo(u.typ).Interface().(*float64) v := &float64Value{*t} siz := Size(v) return tagsize + SizeVarint(uint64(siz)) + siz }, func(b []byte, ptr pointer, wiretag uint64, deterministic bool) ([]byte, error) { t := ptr.asPointerTo(u.typ).Interface().(*float64) v := &float64Value{*t} buf, err := Marshal(v) if err != nil { return nil, err } b = appendVarint(b, wiretag) b = appendVarint(b, uint64(len(buf))) b = append(b, buf...) return b, nil } } func makeStdDoubleValuePtrMarshaler(u *marshalInfo) (sizer, marshaler) { return func(ptr pointer, tagsize int) int { if ptr.isNil() { return 0 } t := ptr.asPointerTo(reflect.PtrTo(u.typ)).Elem().Interface().(*float64) v := &float64Value{*t} siz := Size(v) return tagsize + SizeVarint(uint64(siz)) + siz }, func(b []byte, ptr pointer, wiretag uint64, deterministic bool) ([]byte, error) { if ptr.isNil() { return b, nil } t := ptr.asPointerTo(reflect.PtrTo(u.typ)).Elem().Interface().(*float64) v := &float64Value{*t} buf, err := Marshal(v) if err != nil { return nil, err } b = appendVarint(b, wiretag) b = appendVarint(b, uint64(len(buf))) b = append(b, buf...) return b, nil } } func makeStdDoubleValueSliceMarshaler(u *marshalInfo) (sizer, marshaler) { return func(ptr pointer, tagsize int) int { s := ptr.getSlice(u.typ) n := 0 for i := 0; i < s.Len(); i++ { elem := s.Index(i) t := elem.Interface().(float64) v := &float64Value{t} siz := Size(v) n += siz + SizeVarint(uint64(siz)) + tagsize } return n }, func(b []byte, ptr pointer, wiretag uint64, deterministic bool) ([]byte, error) { s := ptr.getSlice(u.typ) for i := 0; i < s.Len(); i++ { elem := s.Index(i) t := elem.Interface().(float64) v := &float64Value{t} siz := Size(v) buf, err := Marshal(v) if err != nil { return nil, err } b = appendVarint(b, wiretag) b = appendVarint(b, uint64(siz)) b = append(b, buf...) } return b, nil } } func makeStdDoubleValuePtrSliceMarshaler(u *marshalInfo) (sizer, marshaler) { return func(ptr pointer, tagsize int) int { s := ptr.getSlice(reflect.PtrTo(u.typ)) n := 0 for i := 0; i < s.Len(); i++ { elem := s.Index(i) t := elem.Interface().(*float64) v := &float64Value{*t} siz := Size(v) n += siz + SizeVarint(uint64(siz)) + tagsize } return n }, func(b []byte, ptr pointer, wiretag uint64, deterministic bool) ([]byte, error) { s := ptr.getSlice(reflect.PtrTo(u.typ)) for i := 0; i < s.Len(); i++ { elem := s.Index(i) t := elem.Interface().(*float64) v := &float64Value{*t} siz := Size(v) buf, err := Marshal(v) if err != nil { return nil, err } b = appendVarint(b, wiretag) b = appendVarint(b, uint64(siz)) b = append(b, buf...) } return b, nil } } func makeStdDoubleValueUnmarshaler(sub *unmarshalInfo, name string) unmarshaler { return func(b []byte, f pointer, w int) ([]byte, error) { if w != WireBytes { return nil, errInternalBadWireType } x, n := decodeVarint(b) if n == 0 { return nil, io.ErrUnexpectedEOF } b = b[n:] if x > uint64(len(b)) { return nil, io.ErrUnexpectedEOF } m := &float64Value{} if err := Unmarshal(b[:x], m); err != nil { return nil, err } s := f.asPointerTo(sub.typ).Elem() s.Set(reflect.ValueOf(m.Value)) return b[x:], nil } } func makeStdDoubleValuePtrUnmarshaler(sub *unmarshalInfo, name string) unmarshaler { return func(b []byte, f pointer, w int) ([]byte, error) { if w != WireBytes { return nil, errInternalBadWireType } x, n := decodeVarint(b) if n == 0 { return nil, io.ErrUnexpectedEOF } b = b[n:] if x > uint64(len(b)) { return nil, io.ErrUnexpectedEOF } m := &float64Value{} if err := Unmarshal(b[:x], m); err != nil { return nil, err } s := f.asPointerTo(reflect.PtrTo(sub.typ)).Elem() s.Set(reflect.ValueOf(&m.Value)) return b[x:], nil } } func makeStdDoubleValuePtrSliceUnmarshaler(sub *unmarshalInfo, name string) unmarshaler { return func(b []byte, f pointer, w int) ([]byte, error) { if w != WireBytes { return nil, errInternalBadWireType } x, n := decodeVarint(b) if n == 0 { return nil, io.ErrUnexpectedEOF } b = b[n:] if x > uint64(len(b)) { return nil, io.ErrUnexpectedEOF } m := &float64Value{} if err := Unmarshal(b[:x], m); err != nil { return nil, err } slice := f.getSlice(reflect.PtrTo(sub.typ)) newSlice := reflect.Append(slice, reflect.ValueOf(&m.Value)) slice.Set(newSlice) return b[x:], nil } } func makeStdDoubleValueSliceUnmarshaler(sub *unmarshalInfo, name string) unmarshaler { return func(b []byte, f pointer, w int) ([]byte, error) { if w != WireBytes { return nil, errInternalBadWireType } x, n := decodeVarint(b) if n == 0 { return nil, io.ErrUnexpectedEOF } b = b[n:] if x > uint64(len(b)) { return nil, io.ErrUnexpectedEOF } m := &float64Value{} if err := Unmarshal(b[:x], m); err != nil { return nil, err } slice := f.getSlice(sub.typ) newSlice := reflect.Append(slice, reflect.ValueOf(m.Value)) slice.Set(newSlice) return b[x:], nil } } func makeStdFloatValueMarshaler(u *marshalInfo) (sizer, marshaler) { return func(ptr pointer, tagsize int) int { t := ptr.asPointerTo(u.typ).Interface().(*float32) v := &float32Value{*t} siz := Size(v) return tagsize + SizeVarint(uint64(siz)) + siz }, func(b []byte, ptr pointer, wiretag uint64, deterministic bool) ([]byte, error) { t := ptr.asPointerTo(u.typ).Interface().(*float32) v := &float32Value{*t} buf, err := Marshal(v) if err != nil { return nil, err } b = appendVarint(b, wiretag) b = appendVarint(b, uint64(len(buf))) b = append(b, buf...) return b, nil } } func makeStdFloatValuePtrMarshaler(u *marshalInfo) (sizer, marshaler) { return func(ptr pointer, tagsize int) int { if ptr.isNil() { return 0 } t := ptr.asPointerTo(reflect.PtrTo(u.typ)).Elem().Interface().(*float32) v := &float32Value{*t} siz := Size(v) return tagsize + SizeVarint(uint64(siz)) + siz }, func(b []byte, ptr pointer, wiretag uint64, deterministic bool) ([]byte, error) { if ptr.isNil() { return b, nil } t := ptr.asPointerTo(reflect.PtrTo(u.typ)).Elem().Interface().(*float32) v := &float32Value{*t} buf, err := Marshal(v) if err != nil { return nil, err } b = appendVarint(b, wiretag) b = appendVarint(b, uint64(len(buf))) b = append(b, buf...) return b, nil } } func makeStdFloatValueSliceMarshaler(u *marshalInfo) (sizer, marshaler) { return func(ptr pointer, tagsize int) int { s := ptr.getSlice(u.typ) n := 0 for i := 0; i < s.Len(); i++ { elem := s.Index(i) t := elem.Interface().(float32) v := &float32Value{t} siz := Size(v) n += siz + SizeVarint(uint64(siz)) + tagsize } return n }, func(b []byte, ptr pointer, wiretag uint64, deterministic bool) ([]byte, error) { s := ptr.getSlice(u.typ) for i := 0; i < s.Len(); i++ { elem := s.Index(i) t := elem.Interface().(float32) v := &float32Value{t} siz := Size(v) buf, err := Marshal(v) if err != nil { return nil, err } b = appendVarint(b, wiretag) b = appendVarint(b, uint64(siz)) b = append(b, buf...) } return b, nil } } func makeStdFloatValuePtrSliceMarshaler(u *marshalInfo) (sizer, marshaler) { return func(ptr pointer, tagsize int) int { s := ptr.getSlice(reflect.PtrTo(u.typ)) n := 0 for i := 0; i < s.Len(); i++ { elem := s.Index(i) t := elem.Interface().(*float32) v := &float32Value{*t} siz := Size(v) n += siz + SizeVarint(uint64(siz)) + tagsize } return n }, func(b []byte, ptr pointer, wiretag uint64, deterministic bool) ([]byte, error) { s := ptr.getSlice(reflect.PtrTo(u.typ)) for i := 0; i < s.Len(); i++ { elem := s.Index(i) t := elem.Interface().(*float32) v := &float32Value{*t} siz := Size(v) buf, err := Marshal(v) if err != nil { return nil, err } b = appendVarint(b, wiretag) b = appendVarint(b, uint64(siz)) b = append(b, buf...) } return b, nil } } func makeStdFloatValueUnmarshaler(sub *unmarshalInfo, name string) unmarshaler { return func(b []byte, f pointer, w int) ([]byte, error) { if w != WireBytes { return nil, errInternalBadWireType } x, n := decodeVarint(b) if n == 0 { return nil, io.ErrUnexpectedEOF } b = b[n:] if x > uint64(len(b)) { return nil, io.ErrUnexpectedEOF } m := &float32Value{} if err := Unmarshal(b[:x], m); err != nil { return nil, err } s := f.asPointerTo(sub.typ).Elem() s.Set(reflect.ValueOf(m.Value)) return b[x:], nil } } func makeStdFloatValuePtrUnmarshaler(sub *unmarshalInfo, name string) unmarshaler { return func(b []byte, f pointer, w int) ([]byte, error) { if w != WireBytes { return nil, errInternalBadWireType } x, n := decodeVarint(b) if n == 0 { return nil, io.ErrUnexpectedEOF } b = b[n:] if x > uint64(len(b)) { return nil, io.ErrUnexpectedEOF } m := &float32Value{} if err := Unmarshal(b[:x], m); err != nil { return nil, err } s := f.asPointerTo(reflect.PtrTo(sub.typ)).Elem() s.Set(reflect.ValueOf(&m.Value)) return b[x:], nil } } func makeStdFloatValuePtrSliceUnmarshaler(sub *unmarshalInfo, name string) unmarshaler { return func(b []byte, f pointer, w int) ([]byte, error) { if w != WireBytes { return nil, errInternalBadWireType } x, n := decodeVarint(b) if n == 0 { return nil, io.ErrUnexpectedEOF } b = b[n:] if x > uint64(len(b)) { return nil, io.ErrUnexpectedEOF } m := &float32Value{} if err := Unmarshal(b[:x], m); err != nil { return nil, err } slice := f.getSlice(reflect.PtrTo(sub.typ)) newSlice := reflect.Append(slice, reflect.ValueOf(&m.Value)) slice.Set(newSlice) return b[x:], nil } } func makeStdFloatValueSliceUnmarshaler(sub *unmarshalInfo, name string) unmarshaler { return func(b []byte, f pointer, w int) ([]byte, error) { if w != WireBytes { return nil, errInternalBadWireType } x, n := decodeVarint(b) if n == 0 { return nil, io.ErrUnexpectedEOF } b = b[n:] if x > uint64(len(b)) { return nil, io.ErrUnexpectedEOF } m := &float32Value{} if err := Unmarshal(b[:x], m); err != nil { return nil, err } slice := f.getSlice(sub.typ) newSlice := reflect.Append(slice, reflect.ValueOf(m.Value)) slice.Set(newSlice) return b[x:], nil } } func makeStdInt64ValueMarshaler(u *marshalInfo) (sizer, marshaler) { return func(ptr pointer, tagsize int) int { t := ptr.asPointerTo(u.typ).Interface().(*int64) v := &int64Value{*t} siz := Size(v) return tagsize + SizeVarint(uint64(siz)) + siz }, func(b []byte, ptr pointer, wiretag uint64, deterministic bool) ([]byte, error) { t := ptr.asPointerTo(u.typ).Interface().(*int64) v := &int64Value{*t} buf, err := Marshal(v) if err != nil { return nil, err } b = appendVarint(b, wiretag) b = appendVarint(b, uint64(len(buf))) b = append(b, buf...) return b, nil } } func makeStdInt64ValuePtrMarshaler(u *marshalInfo) (sizer, marshaler) { return func(ptr pointer, tagsize int) int { if ptr.isNil() { return 0 } t := ptr.asPointerTo(reflect.PtrTo(u.typ)).Elem().Interface().(*int64) v := &int64Value{*t} siz := Size(v) return tagsize + SizeVarint(uint64(siz)) + siz }, func(b []byte, ptr pointer, wiretag uint64, deterministic bool) ([]byte, error) { if ptr.isNil() { return b, nil } t := ptr.asPointerTo(reflect.PtrTo(u.typ)).Elem().Interface().(*int64) v := &int64Value{*t} buf, err := Marshal(v) if err != nil { return nil, err } b = appendVarint(b, wiretag) b = appendVarint(b, uint64(len(buf))) b = append(b, buf...) return b, nil } } func makeStdInt64ValueSliceMarshaler(u *marshalInfo) (sizer, marshaler) { return func(ptr pointer, tagsize int) int { s := ptr.getSlice(u.typ) n := 0 for i := 0; i < s.Len(); i++ { elem := s.Index(i) t := elem.Interface().(int64) v := &int64Value{t} siz := Size(v) n += siz + SizeVarint(uint64(siz)) + tagsize } return n }, func(b []byte, ptr pointer, wiretag uint64, deterministic bool) ([]byte, error) { s := ptr.getSlice(u.typ) for i := 0; i < s.Len(); i++ { elem := s.Index(i) t := elem.Interface().(int64) v := &int64Value{t} siz := Size(v) buf, err := Marshal(v) if err != nil { return nil, err } b = appendVarint(b, wiretag) b = appendVarint(b, uint64(siz)) b = append(b, buf...) } return b, nil } } func makeStdInt64ValuePtrSliceMarshaler(u *marshalInfo) (sizer, marshaler) { return func(ptr pointer, tagsize int) int { s := ptr.getSlice(reflect.PtrTo(u.typ)) n := 0 for i := 0; i < s.Len(); i++ { elem := s.Index(i) t := elem.Interface().(*int64) v := &int64Value{*t} siz := Size(v) n += siz + SizeVarint(uint64(siz)) + tagsize } return n }, func(b []byte, ptr pointer, wiretag uint64, deterministic bool) ([]byte, error) { s := ptr.getSlice(reflect.PtrTo(u.typ)) for i := 0; i < s.Len(); i++ { elem := s.Index(i) t := elem.Interface().(*int64) v := &int64Value{*t} siz := Size(v) buf, err := Marshal(v) if err != nil { return nil, err } b = appendVarint(b, wiretag) b = appendVarint(b, uint64(siz)) b = append(b, buf...) } return b, nil } } func makeStdInt64ValueUnmarshaler(sub *unmarshalInfo, name string) unmarshaler { return func(b []byte, f pointer, w int) ([]byte, error) { if w != WireBytes { return nil, errInternalBadWireType } x, n := decodeVarint(b) if n == 0 { return nil, io.ErrUnexpectedEOF } b = b[n:] if x > uint64(len(b)) { return nil, io.ErrUnexpectedEOF } m := &int64Value{} if err := Unmarshal(b[:x], m); err != nil { return nil, err } s := f.asPointerTo(sub.typ).Elem() s.Set(reflect.ValueOf(m.Value)) return b[x:], nil } } func makeStdInt64ValuePtrUnmarshaler(sub *unmarshalInfo, name string) unmarshaler { return func(b []byte, f pointer, w int) ([]byte, error) { if w != WireBytes { return nil, errInternalBadWireType } x, n := decodeVarint(b) if n == 0 { return nil, io.ErrUnexpectedEOF } b = b[n:] if x > uint64(len(b)) { return nil, io.ErrUnexpectedEOF } m := &int64Value{} if err := Unmarshal(b[:x], m); err != nil { return nil, err } s := f.asPointerTo(reflect.PtrTo(sub.typ)).Elem() s.Set(reflect.ValueOf(&m.Value)) return b[x:], nil } } func makeStdInt64ValuePtrSliceUnmarshaler(sub *unmarshalInfo, name string) unmarshaler { return func(b []byte, f pointer, w int) ([]byte, error) { if w != WireBytes { return nil, errInternalBadWireType } x, n := decodeVarint(b) if n == 0 { return nil, io.ErrUnexpectedEOF } b = b[n:] if x > uint64(len(b)) { return nil, io.ErrUnexpectedEOF } m := &int64Value{} if err := Unmarshal(b[:x], m); err != nil { return nil, err } slice := f.getSlice(reflect.PtrTo(sub.typ)) newSlice := reflect.Append(slice, reflect.ValueOf(&m.Value)) slice.Set(newSlice) return b[x:], nil } } func makeStdInt64ValueSliceUnmarshaler(sub *unmarshalInfo, name string) unmarshaler { return func(b []byte, f pointer, w int) ([]byte, error) { if w != WireBytes { return nil, errInternalBadWireType } x, n := decodeVarint(b) if n == 0 { return nil, io.ErrUnexpectedEOF } b = b[n:] if x > uint64(len(b)) { return nil, io.ErrUnexpectedEOF } m := &int64Value{} if err := Unmarshal(b[:x], m); err != nil { return nil, err } slice := f.getSlice(sub.typ) newSlice := reflect.Append(slice, reflect.ValueOf(m.Value)) slice.Set(newSlice) return b[x:], nil } } func makeStdUInt64ValueMarshaler(u *marshalInfo) (sizer, marshaler) { return func(ptr pointer, tagsize int) int { t := ptr.asPointerTo(u.typ).Interface().(*uint64) v := &uint64Value{*t} siz := Size(v) return tagsize + SizeVarint(uint64(siz)) + siz }, func(b []byte, ptr pointer, wiretag uint64, deterministic bool) ([]byte, error) { t := ptr.asPointerTo(u.typ).Interface().(*uint64) v := &uint64Value{*t} buf, err := Marshal(v) if err != nil { return nil, err } b = appendVarint(b, wiretag) b = appendVarint(b, uint64(len(buf))) b = append(b, buf...) return b, nil } } func makeStdUInt64ValuePtrMarshaler(u *marshalInfo) (sizer, marshaler) { return func(ptr pointer, tagsize int) int { if ptr.isNil() { return 0 } t := ptr.asPointerTo(reflect.PtrTo(u.typ)).Elem().Interface().(*uint64) v := &uint64Value{*t} siz := Size(v) return tagsize + SizeVarint(uint64(siz)) + siz }, func(b []byte, ptr pointer, wiretag uint64, deterministic bool) ([]byte, error) { if ptr.isNil() { return b, nil } t := ptr.asPointerTo(reflect.PtrTo(u.typ)).Elem().Interface().(*uint64) v := &uint64Value{*t} buf, err := Marshal(v) if err != nil { return nil, err } b = appendVarint(b, wiretag) b = appendVarint(b, uint64(len(buf))) b = append(b, buf...) return b, nil } } func makeStdUInt64ValueSliceMarshaler(u *marshalInfo) (sizer, marshaler) { return func(ptr pointer, tagsize int) int { s := ptr.getSlice(u.typ) n := 0 for i := 0; i < s.Len(); i++ { elem := s.Index(i) t := elem.Interface().(uint64) v := &uint64Value{t} siz := Size(v) n += siz + SizeVarint(uint64(siz)) + tagsize } return n }, func(b []byte, ptr pointer, wiretag uint64, deterministic bool) ([]byte, error) { s := ptr.getSlice(u.typ) for i := 0; i < s.Len(); i++ { elem := s.Index(i) t := elem.Interface().(uint64) v := &uint64Value{t} siz := Size(v) buf, err := Marshal(v) if err != nil { return nil, err } b = appendVarint(b, wiretag) b = appendVarint(b, uint64(siz)) b = append(b, buf...) } return b, nil } } func makeStdUInt64ValuePtrSliceMarshaler(u *marshalInfo) (sizer, marshaler) { return func(ptr pointer, tagsize int) int { s := ptr.getSlice(reflect.PtrTo(u.typ)) n := 0 for i := 0; i < s.Len(); i++ { elem := s.Index(i) t := elem.Interface().(*uint64) v := &uint64Value{*t} siz := Size(v) n += siz + SizeVarint(uint64(siz)) + tagsize } return n }, func(b []byte, ptr pointer, wiretag uint64, deterministic bool) ([]byte, error) { s := ptr.getSlice(reflect.PtrTo(u.typ)) for i := 0; i < s.Len(); i++ { elem := s.Index(i) t := elem.Interface().(*uint64) v := &uint64Value{*t} siz := Size(v) buf, err := Marshal(v) if err != nil { return nil, err } b = appendVarint(b, wiretag) b = appendVarint(b, uint64(siz)) b = append(b, buf...) } return b, nil } } func makeStdUInt64ValueUnmarshaler(sub *unmarshalInfo, name string) unmarshaler { return func(b []byte, f pointer, w int) ([]byte, error) { if w != WireBytes { return nil, errInternalBadWireType } x, n := decodeVarint(b) if n == 0 { return nil, io.ErrUnexpectedEOF } b = b[n:] if x > uint64(len(b)) { return nil, io.ErrUnexpectedEOF } m := &uint64Value{} if err := Unmarshal(b[:x], m); err != nil { return nil, err } s := f.asPointerTo(sub.typ).Elem() s.Set(reflect.ValueOf(m.Value)) return b[x:], nil } } func makeStdUInt64ValuePtrUnmarshaler(sub *unmarshalInfo, name string) unmarshaler { return func(b []byte, f pointer, w int) ([]byte, error) { if w != WireBytes { return nil, errInternalBadWireType } x, n := decodeVarint(b) if n == 0 { return nil, io.ErrUnexpectedEOF } b = b[n:] if x > uint64(len(b)) { return nil, io.ErrUnexpectedEOF } m := &uint64Value{} if err := Unmarshal(b[:x], m); err != nil { return nil, err } s := f.asPointerTo(reflect.PtrTo(sub.typ)).Elem() s.Set(reflect.ValueOf(&m.Value)) return b[x:], nil } } func makeStdUInt64ValuePtrSliceUnmarshaler(sub *unmarshalInfo, name string) unmarshaler { return func(b []byte, f pointer, w int) ([]byte, error) { if w != WireBytes { return nil, errInternalBadWireType } x, n := decodeVarint(b) if n == 0 { return nil, io.ErrUnexpectedEOF } b = b[n:] if x > uint64(len(b)) { return nil, io.ErrUnexpectedEOF } m := &uint64Value{} if err := Unmarshal(b[:x], m); err != nil { return nil, err } slice := f.getSlice(reflect.PtrTo(sub.typ)) newSlice := reflect.Append(slice, reflect.ValueOf(&m.Value)) slice.Set(newSlice) return b[x:], nil } } func makeStdUInt64ValueSliceUnmarshaler(sub *unmarshalInfo, name string) unmarshaler { return func(b []byte, f pointer, w int) ([]byte, error) { if w != WireBytes { return nil, errInternalBadWireType } x, n := decodeVarint(b) if n == 0 { return nil, io.ErrUnexpectedEOF } b = b[n:] if x > uint64(len(b)) { return nil, io.ErrUnexpectedEOF } m := &uint64Value{} if err := Unmarshal(b[:x], m); err != nil { return nil, err } slice := f.getSlice(sub.typ) newSlice := reflect.Append(slice, reflect.ValueOf(m.Value)) slice.Set(newSlice) return b[x:], nil } } func makeStdInt32ValueMarshaler(u *marshalInfo) (sizer, marshaler) { return func(ptr pointer, tagsize int) int { t := ptr.asPointerTo(u.typ).Interface().(*int32) v := &int32Value{*t} siz := Size(v) return tagsize + SizeVarint(uint64(siz)) + siz }, func(b []byte, ptr pointer, wiretag uint64, deterministic bool) ([]byte, error) { t := ptr.asPointerTo(u.typ).Interface().(*int32) v := &int32Value{*t} buf, err := Marshal(v) if err != nil { return nil, err } b = appendVarint(b, wiretag) b = appendVarint(b, uint64(len(buf))) b = append(b, buf...) return b, nil } } func makeStdInt32ValuePtrMarshaler(u *marshalInfo) (sizer, marshaler) { return func(ptr pointer, tagsize int) int { if ptr.isNil() { return 0 } t := ptr.asPointerTo(reflect.PtrTo(u.typ)).Elem().Interface().(*int32) v := &int32Value{*t} siz := Size(v) return tagsize + SizeVarint(uint64(siz)) + siz }, func(b []byte, ptr pointer, wiretag uint64, deterministic bool) ([]byte, error) { if ptr.isNil() { return b, nil } t := ptr.asPointerTo(reflect.PtrTo(u.typ)).Elem().Interface().(*int32) v := &int32Value{*t} buf, err := Marshal(v) if err != nil { return nil, err } b = appendVarint(b, wiretag) b = appendVarint(b, uint64(len(buf))) b = append(b, buf...) return b, nil } } func makeStdInt32ValueSliceMarshaler(u *marshalInfo) (sizer, marshaler) { return func(ptr pointer, tagsize int) int { s := ptr.getSlice(u.typ) n := 0 for i := 0; i < s.Len(); i++ { elem := s.Index(i) t := elem.Interface().(int32) v := &int32Value{t} siz := Size(v) n += siz + SizeVarint(uint64(siz)) + tagsize } return n }, func(b []byte, ptr pointer, wiretag uint64, deterministic bool) ([]byte, error) { s := ptr.getSlice(u.typ) for i := 0; i < s.Len(); i++ { elem := s.Index(i) t := elem.Interface().(int32) v := &int32Value{t} siz := Size(v) buf, err := Marshal(v) if err != nil { return nil, err } b = appendVarint(b, wiretag) b = appendVarint(b, uint64(siz)) b = append(b, buf...) } return b, nil } } func makeStdInt32ValuePtrSliceMarshaler(u *marshalInfo) (sizer, marshaler) { return func(ptr pointer, tagsize int) int { s := ptr.getSlice(reflect.PtrTo(u.typ)) n := 0 for i := 0; i < s.Len(); i++ { elem := s.Index(i) t := elem.Interface().(*int32) v := &int32Value{*t} siz := Size(v) n += siz + SizeVarint(uint64(siz)) + tagsize } return n }, func(b []byte, ptr pointer, wiretag uint64, deterministic bool) ([]byte, error) { s := ptr.getSlice(reflect.PtrTo(u.typ)) for i := 0; i < s.Len(); i++ { elem := s.Index(i) t := elem.Interface().(*int32) v := &int32Value{*t} siz := Size(v) buf, err := Marshal(v) if err != nil { return nil, err } b = appendVarint(b, wiretag) b = appendVarint(b, uint64(siz)) b = append(b, buf...) } return b, nil } } func makeStdInt32ValueUnmarshaler(sub *unmarshalInfo, name string) unmarshaler { return func(b []byte, f pointer, w int) ([]byte, error) { if w != WireBytes { return nil, errInternalBadWireType } x, n := decodeVarint(b) if n == 0 { return nil, io.ErrUnexpectedEOF } b = b[n:] if x > uint64(len(b)) { return nil, io.ErrUnexpectedEOF } m := &int32Value{} if err := Unmarshal(b[:x], m); err != nil { return nil, err } s := f.asPointerTo(sub.typ).Elem() s.Set(reflect.ValueOf(m.Value)) return b[x:], nil } } func makeStdInt32ValuePtrUnmarshaler(sub *unmarshalInfo, name string) unmarshaler { return func(b []byte, f pointer, w int) ([]byte, error) { if w != WireBytes { return nil, errInternalBadWireType } x, n := decodeVarint(b) if n == 0 { return nil, io.ErrUnexpectedEOF } b = b[n:] if x > uint64(len(b)) { return nil, io.ErrUnexpectedEOF } m := &int32Value{} if err := Unmarshal(b[:x], m); err != nil { return nil, err } s := f.asPointerTo(reflect.PtrTo(sub.typ)).Elem() s.Set(reflect.ValueOf(&m.Value)) return b[x:], nil } } func makeStdInt32ValuePtrSliceUnmarshaler(sub *unmarshalInfo, name string) unmarshaler { return func(b []byte, f pointer, w int) ([]byte, error) { if w != WireBytes { return nil, errInternalBadWireType } x, n := decodeVarint(b) if n == 0 { return nil, io.ErrUnexpectedEOF } b = b[n:] if x > uint64(len(b)) { return nil, io.ErrUnexpectedEOF } m := &int32Value{} if err := Unmarshal(b[:x], m); err != nil { return nil, err } slice := f.getSlice(reflect.PtrTo(sub.typ)) newSlice := reflect.Append(slice, reflect.ValueOf(&m.Value)) slice.Set(newSlice) return b[x:], nil } } func makeStdInt32ValueSliceUnmarshaler(sub *unmarshalInfo, name string) unmarshaler { return func(b []byte, f pointer, w int) ([]byte, error) { if w != WireBytes { return nil, errInternalBadWireType } x, n := decodeVarint(b) if n == 0 { return nil, io.ErrUnexpectedEOF } b = b[n:] if x > uint64(len(b)) { return nil, io.ErrUnexpectedEOF } m := &int32Value{} if err := Unmarshal(b[:x], m); err != nil { return nil, err } slice := f.getSlice(sub.typ) newSlice := reflect.Append(slice, reflect.ValueOf(m.Value)) slice.Set(newSlice) return b[x:], nil } } func makeStdUInt32ValueMarshaler(u *marshalInfo) (sizer, marshaler) { return func(ptr pointer, tagsize int) int { t := ptr.asPointerTo(u.typ).Interface().(*uint32) v := &uint32Value{*t} siz := Size(v) return tagsize + SizeVarint(uint64(siz)) + siz }, func(b []byte, ptr pointer, wiretag uint64, deterministic bool) ([]byte, error) { t := ptr.asPointerTo(u.typ).Interface().(*uint32) v := &uint32Value{*t} buf, err := Marshal(v) if err != nil { return nil, err } b = appendVarint(b, wiretag) b = appendVarint(b, uint64(len(buf))) b = append(b, buf...) return b, nil } } func makeStdUInt32ValuePtrMarshaler(u *marshalInfo) (sizer, marshaler) { return func(ptr pointer, tagsize int) int { if ptr.isNil() { return 0 } t := ptr.asPointerTo(reflect.PtrTo(u.typ)).Elem().Interface().(*uint32) v := &uint32Value{*t} siz := Size(v) return tagsize + SizeVarint(uint64(siz)) + siz }, func(b []byte, ptr pointer, wiretag uint64, deterministic bool) ([]byte, error) { if ptr.isNil() { return b, nil } t := ptr.asPointerTo(reflect.PtrTo(u.typ)).Elem().Interface().(*uint32) v := &uint32Value{*t} buf, err := Marshal(v) if err != nil { return nil, err } b = appendVarint(b, wiretag) b = appendVarint(b, uint64(len(buf))) b = append(b, buf...) return b, nil } } func makeStdUInt32ValueSliceMarshaler(u *marshalInfo) (sizer, marshaler) { return func(ptr pointer, tagsize int) int { s := ptr.getSlice(u.typ) n := 0 for i := 0; i < s.Len(); i++ { elem := s.Index(i) t := elem.Interface().(uint32) v := &uint32Value{t} siz := Size(v) n += siz + SizeVarint(uint64(siz)) + tagsize } return n }, func(b []byte, ptr pointer, wiretag uint64, deterministic bool) ([]byte, error) { s := ptr.getSlice(u.typ) for i := 0; i < s.Len(); i++ { elem := s.Index(i) t := elem.Interface().(uint32) v := &uint32Value{t} siz := Size(v) buf, err := Marshal(v) if err != nil { return nil, err } b = appendVarint(b, wiretag) b = appendVarint(b, uint64(siz)) b = append(b, buf...) } return b, nil } } func makeStdUInt32ValuePtrSliceMarshaler(u *marshalInfo) (sizer, marshaler) { return func(ptr pointer, tagsize int) int { s := ptr.getSlice(reflect.PtrTo(u.typ)) n := 0 for i := 0; i < s.Len(); i++ { elem := s.Index(i) t := elem.Interface().(*uint32) v := &uint32Value{*t} siz := Size(v) n += siz + SizeVarint(uint64(siz)) + tagsize } return n }, func(b []byte, ptr pointer, wiretag uint64, deterministic bool) ([]byte, error) { s := ptr.getSlice(reflect.PtrTo(u.typ)) for i := 0; i < s.Len(); i++ { elem := s.Index(i) t := elem.Interface().(*uint32) v := &uint32Value{*t} siz := Size(v) buf, err := Marshal(v) if err != nil { return nil, err } b = appendVarint(b, wiretag) b = appendVarint(b, uint64(siz)) b = append(b, buf...) } return b, nil } } func makeStdUInt32ValueUnmarshaler(sub *unmarshalInfo, name string) unmarshaler { return func(b []byte, f pointer, w int) ([]byte, error) { if w != WireBytes { return nil, errInternalBadWireType } x, n := decodeVarint(b) if n == 0 { return nil, io.ErrUnexpectedEOF } b = b[n:] if x > uint64(len(b)) { return nil, io.ErrUnexpectedEOF } m := &uint32Value{} if err := Unmarshal(b[:x], m); err != nil { return nil, err } s := f.asPointerTo(sub.typ).Elem() s.Set(reflect.ValueOf(m.Value)) return b[x:], nil } } func makeStdUInt32ValuePtrUnmarshaler(sub *unmarshalInfo, name string) unmarshaler { return func(b []byte, f pointer, w int) ([]byte, error) { if w != WireBytes { return nil, errInternalBadWireType } x, n := decodeVarint(b) if n == 0 { return nil, io.ErrUnexpectedEOF } b = b[n:] if x > uint64(len(b)) { return nil, io.ErrUnexpectedEOF } m := &uint32Value{} if err := Unmarshal(b[:x], m); err != nil { return nil, err } s := f.asPointerTo(reflect.PtrTo(sub.typ)).Elem() s.Set(reflect.ValueOf(&m.Value)) return b[x:], nil } } func makeStdUInt32ValuePtrSliceUnmarshaler(sub *unmarshalInfo, name string) unmarshaler { return func(b []byte, f pointer, w int) ([]byte, error) { if w != WireBytes { return nil, errInternalBadWireType } x, n := decodeVarint(b) if n == 0 { return nil, io.ErrUnexpectedEOF } b = b[n:] if x > uint64(len(b)) {
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
true
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/gogo/protobuf/proto/pointer_unsafe_gogo.go
vendor/github.com/gogo/protobuf/proto/pointer_unsafe_gogo.go
// Protocol Buffers for Go with Gadgets // // Copyright (c) 2018, The GoGo Authors. All rights reserved. // http://github.com/gogo/protobuf // // Redistribution and use in source and binary forms, with or without // modification, are permitted provided that the following conditions are // met: // // * Redistributions of source code must retain the above copyright // notice, this list of conditions and the following disclaimer. // * Redistributions in binary form must reproduce the above // copyright notice, this list of conditions and the following disclaimer // in the documentation and/or other materials provided with the // distribution. // // THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS // "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT // LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR // A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT // OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, // SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT // LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, // DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY // THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT // (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE // OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. // +build !purego,!appengine,!js // This file contains the implementation of the proto field accesses using package unsafe. package proto import ( "reflect" "unsafe" ) func (p pointer) getRef() pointer { return pointer{p: (unsafe.Pointer)(&p.p)} } func (p pointer) appendRef(v pointer, typ reflect.Type) { slice := p.getSlice(typ) elem := v.asPointerTo(typ).Elem() newSlice := reflect.Append(slice, elem) slice.Set(newSlice) } func (p pointer) getSlice(typ reflect.Type) reflect.Value { sliceTyp := reflect.SliceOf(typ) slice := p.asPointerTo(sliceTyp) slice = slice.Elem() return slice }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/gogo/protobuf/sortkeys/sortkeys.go
vendor/github.com/gogo/protobuf/sortkeys/sortkeys.go
// Protocol Buffers for Go with Gadgets // // Copyright (c) 2013, The GoGo Authors. All rights reserved. // http://github.com/gogo/protobuf // // Redistribution and use in source and binary forms, with or without // modification, are permitted provided that the following conditions are // met: // // * Redistributions of source code must retain the above copyright // notice, this list of conditions and the following disclaimer. // * Redistributions in binary form must reproduce the above // copyright notice, this list of conditions and the following disclaimer // in the documentation and/or other materials provided with the // distribution. // // THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS // "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT // LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR // A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT // OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, // SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT // LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, // DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY // THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT // (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE // OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. package sortkeys import ( "sort" ) func Strings(l []string) { sort.Strings(l) } func Float64s(l []float64) { sort.Float64s(l) } func Float32s(l []float32) { sort.Sort(Float32Slice(l)) } func Int64s(l []int64) { sort.Sort(Int64Slice(l)) } func Int32s(l []int32) { sort.Sort(Int32Slice(l)) } func Uint64s(l []uint64) { sort.Sort(Uint64Slice(l)) } func Uint32s(l []uint32) { sort.Sort(Uint32Slice(l)) } func Bools(l []bool) { sort.Sort(BoolSlice(l)) } type BoolSlice []bool func (p BoolSlice) Len() int { return len(p) } func (p BoolSlice) Less(i, j int) bool { return p[j] } func (p BoolSlice) Swap(i, j int) { p[i], p[j] = p[j], p[i] } type Int64Slice []int64 func (p Int64Slice) Len() int { return len(p) } func (p Int64Slice) Less(i, j int) bool { return p[i] < p[j] } func (p Int64Slice) Swap(i, j int) { p[i], p[j] = p[j], p[i] } type Int32Slice []int32 func (p Int32Slice) Len() int { return len(p) } func (p Int32Slice) Less(i, j int) bool { return p[i] < p[j] } func (p Int32Slice) Swap(i, j int) { p[i], p[j] = p[j], p[i] } type Uint64Slice []uint64 func (p Uint64Slice) Len() int { return len(p) } func (p Uint64Slice) Less(i, j int) bool { return p[i] < p[j] } func (p Uint64Slice) Swap(i, j int) { p[i], p[j] = p[j], p[i] } type Uint32Slice []uint32 func (p Uint32Slice) Len() int { return len(p) } func (p Uint32Slice) Less(i, j int) bool { return p[i] < p[j] } func (p Uint32Slice) Swap(i, j int) { p[i], p[j] = p[j], p[i] } type Float32Slice []float32 func (p Float32Slice) Len() int { return len(p) } func (p Float32Slice) Less(i, j int) bool { return p[i] < p[j] } func (p Float32Slice) Swap(i, j int) { p[i], p[j] = p[j], p[i] }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/dustin/go-humanize/ftoa.go
vendor/github.com/dustin/go-humanize/ftoa.go
package humanize import ( "strconv" "strings" ) func stripTrailingZeros(s string) string { if !strings.ContainsRune(s, '.') { return s } offset := len(s) - 1 for offset > 0 { if s[offset] == '.' { offset-- break } if s[offset] != '0' { break } offset-- } return s[:offset+1] } func stripTrailingDigits(s string, digits int) string { if i := strings.Index(s, "."); i >= 0 { if digits <= 0 { return s[:i] } i++ if i+digits >= len(s) { return s } return s[:i+digits] } return s } // Ftoa converts a float to a string with no trailing zeros. func Ftoa(num float64) string { return stripTrailingZeros(strconv.FormatFloat(num, 'f', 6, 64)) } // FtoaWithDigits converts a float to a string but limits the resulting string // to the given number of decimal places, and no trailing zeros. func FtoaWithDigits(num float64, digits int) string { return stripTrailingZeros(stripTrailingDigits(strconv.FormatFloat(num, 'f', 6, 64), digits)) }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/dustin/go-humanize/commaf.go
vendor/github.com/dustin/go-humanize/commaf.go
//go:build go1.6 // +build go1.6 package humanize import ( "bytes" "math/big" "strings" ) // BigCommaf produces a string form of the given big.Float in base 10 // with commas after every three orders of magnitude. func BigCommaf(v *big.Float) string { buf := &bytes.Buffer{} if v.Sign() < 0 { buf.Write([]byte{'-'}) v.Abs(v) } comma := []byte{','} parts := strings.Split(v.Text('f', -1), ".") pos := 0 if len(parts[0])%3 != 0 { pos += len(parts[0]) % 3 buf.WriteString(parts[0][:pos]) buf.Write(comma) } for ; pos < len(parts[0]); pos += 3 { buf.WriteString(parts[0][pos : pos+3]) buf.Write(comma) } buf.Truncate(buf.Len() - 1) if len(parts) > 1 { buf.Write([]byte{'.'}) buf.WriteString(parts[1]) } return buf.String() }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/dustin/go-humanize/comma.go
vendor/github.com/dustin/go-humanize/comma.go
package humanize import ( "bytes" "math" "math/big" "strconv" "strings" ) // Comma produces a string form of the given number in base 10 with // commas after every three orders of magnitude. // // e.g. Comma(834142) -> 834,142 func Comma(v int64) string { sign := "" // Min int64 can't be negated to a usable value, so it has to be special cased. if v == math.MinInt64 { return "-9,223,372,036,854,775,808" } if v < 0 { sign = "-" v = 0 - v } parts := []string{"", "", "", "", "", "", ""} j := len(parts) - 1 for v > 999 { parts[j] = strconv.FormatInt(v%1000, 10) switch len(parts[j]) { case 2: parts[j] = "0" + parts[j] case 1: parts[j] = "00" + parts[j] } v = v / 1000 j-- } parts[j] = strconv.Itoa(int(v)) return sign + strings.Join(parts[j:], ",") } // Commaf produces a string form of the given number in base 10 with // commas after every three orders of magnitude. // // e.g. Commaf(834142.32) -> 834,142.32 func Commaf(v float64) string { buf := &bytes.Buffer{} if v < 0 { buf.Write([]byte{'-'}) v = 0 - v } comma := []byte{','} parts := strings.Split(strconv.FormatFloat(v, 'f', -1, 64), ".") pos := 0 if len(parts[0])%3 != 0 { pos += len(parts[0]) % 3 buf.WriteString(parts[0][:pos]) buf.Write(comma) } for ; pos < len(parts[0]); pos += 3 { buf.WriteString(parts[0][pos : pos+3]) buf.Write(comma) } buf.Truncate(buf.Len() - 1) if len(parts) > 1 { buf.Write([]byte{'.'}) buf.WriteString(parts[1]) } return buf.String() } // CommafWithDigits works like the Commaf but limits the resulting // string to the given number of decimal places. // // e.g. CommafWithDigits(834142.32, 1) -> 834,142.3 func CommafWithDigits(f float64, decimals int) string { return stripTrailingDigits(Commaf(f), decimals) } // BigComma produces a string form of the given big.Int in base 10 // with commas after every three orders of magnitude. func BigComma(b *big.Int) string { sign := "" if b.Sign() < 0 { sign = "-" b.Abs(b) } athousand := big.NewInt(1000) c := (&big.Int{}).Set(b) _, m := oom(c, athousand) parts := make([]string, m+1) j := len(parts) - 1 mod := &big.Int{} for b.Cmp(athousand) >= 0 { b.DivMod(b, athousand, mod) parts[j] = strconv.FormatInt(mod.Int64(), 10) switch len(parts[j]) { case 2: parts[j] = "0" + parts[j] case 1: parts[j] = "00" + parts[j] } j-- } parts[j] = strconv.Itoa(int(b.Int64())) return sign + strings.Join(parts[j:], ",") }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/dustin/go-humanize/humanize.go
vendor/github.com/dustin/go-humanize/humanize.go
/* Package humanize converts boring ugly numbers to human-friendly strings and back. Durations can be turned into strings such as "3 days ago", numbers representing sizes like 82854982 into useful strings like, "83 MB" or "79 MiB" (whichever you prefer). */ package humanize
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/dustin/go-humanize/si.go
vendor/github.com/dustin/go-humanize/si.go
package humanize import ( "errors" "math" "regexp" "strconv" ) var siPrefixTable = map[float64]string{ -30: "q", // quecto -27: "r", // ronto -24: "y", // yocto -21: "z", // zepto -18: "a", // atto -15: "f", // femto -12: "p", // pico -9: "n", // nano -6: "µ", // micro -3: "m", // milli 0: "", 3: "k", // kilo 6: "M", // mega 9: "G", // giga 12: "T", // tera 15: "P", // peta 18: "E", // exa 21: "Z", // zetta 24: "Y", // yotta 27: "R", // ronna 30: "Q", // quetta } var revSIPrefixTable = revfmap(siPrefixTable) // revfmap reverses the map and precomputes the power multiplier func revfmap(in map[float64]string) map[string]float64 { rv := map[string]float64{} for k, v := range in { rv[v] = math.Pow(10, k) } return rv } var riParseRegex *regexp.Regexp func init() { ri := `^([\-0-9.]+)\s?([` for _, v := range siPrefixTable { ri += v } ri += `]?)(.*)` riParseRegex = regexp.MustCompile(ri) } // ComputeSI finds the most appropriate SI prefix for the given number // and returns the prefix along with the value adjusted to be within // that prefix. // // See also: SI, ParseSI. // // e.g. ComputeSI(2.2345e-12) -> (2.2345, "p") func ComputeSI(input float64) (float64, string) { if input == 0 { return 0, "" } mag := math.Abs(input) exponent := math.Floor(logn(mag, 10)) exponent = math.Floor(exponent/3) * 3 value := mag / math.Pow(10, exponent) // Handle special case where value is exactly 1000.0 // Should return 1 M instead of 1000 k if value == 1000.0 { exponent += 3 value = mag / math.Pow(10, exponent) } value = math.Copysign(value, input) prefix := siPrefixTable[exponent] return value, prefix } // SI returns a string with default formatting. // // SI uses Ftoa to format float value, removing trailing zeros. // // See also: ComputeSI, ParseSI. // // e.g. SI(1000000, "B") -> 1 MB // e.g. SI(2.2345e-12, "F") -> 2.2345 pF func SI(input float64, unit string) string { value, prefix := ComputeSI(input) return Ftoa(value) + " " + prefix + unit } // SIWithDigits works like SI but limits the resulting string to the // given number of decimal places. // // e.g. SIWithDigits(1000000, 0, "B") -> 1 MB // e.g. SIWithDigits(2.2345e-12, 2, "F") -> 2.23 pF func SIWithDigits(input float64, decimals int, unit string) string { value, prefix := ComputeSI(input) return FtoaWithDigits(value, decimals) + " " + prefix + unit } var errInvalid = errors.New("invalid input") // ParseSI parses an SI string back into the number and unit. // // See also: SI, ComputeSI. // // e.g. ParseSI("2.2345 pF") -> (2.2345e-12, "F", nil) func ParseSI(input string) (float64, string, error) { found := riParseRegex.FindStringSubmatch(input) if len(found) != 4 { return 0, "", errInvalid } mag := revSIPrefixTable[found[2]] unit := found[3] base, err := strconv.ParseFloat(found[1], 64) return base * mag, unit, err }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/dustin/go-humanize/ordinals.go
vendor/github.com/dustin/go-humanize/ordinals.go
package humanize import "strconv" // Ordinal gives you the input number in a rank/ordinal format. // // Ordinal(3) -> 3rd func Ordinal(x int) string { suffix := "th" switch x % 10 { case 1: if x%100 != 11 { suffix = "st" } case 2: if x%100 != 12 { suffix = "nd" } case 3: if x%100 != 13 { suffix = "rd" } } return strconv.Itoa(x) + suffix }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/dustin/go-humanize/big.go
vendor/github.com/dustin/go-humanize/big.go
package humanize import ( "math/big" ) // order of magnitude (to a max order) func oomm(n, b *big.Int, maxmag int) (float64, int) { mag := 0 m := &big.Int{} for n.Cmp(b) >= 0 { n.DivMod(n, b, m) mag++ if mag == maxmag && maxmag >= 0 { break } } return float64(n.Int64()) + (float64(m.Int64()) / float64(b.Int64())), mag } // total order of magnitude // (same as above, but with no upper limit) func oom(n, b *big.Int) (float64, int) { mag := 0 m := &big.Int{} for n.Cmp(b) >= 0 { n.DivMod(n, b, m) mag++ } return float64(n.Int64()) + (float64(m.Int64()) / float64(b.Int64())), mag }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/dustin/go-humanize/times.go
vendor/github.com/dustin/go-humanize/times.go
package humanize import ( "fmt" "math" "sort" "time" ) // Seconds-based time units const ( Day = 24 * time.Hour Week = 7 * Day Month = 30 * Day Year = 12 * Month LongTime = 37 * Year ) // Time formats a time into a relative string. // // Time(someT) -> "3 weeks ago" func Time(then time.Time) string { return RelTime(then, time.Now(), "ago", "from now") } // A RelTimeMagnitude struct contains a relative time point at which // the relative format of time will switch to a new format string. A // slice of these in ascending order by their "D" field is passed to // CustomRelTime to format durations. // // The Format field is a string that may contain a "%s" which will be // replaced with the appropriate signed label (e.g. "ago" or "from // now") and a "%d" that will be replaced by the quantity. // // The DivBy field is the amount of time the time difference must be // divided by in order to display correctly. // // e.g. if D is 2*time.Minute and you want to display "%d minutes %s" // DivBy should be time.Minute so whatever the duration is will be // expressed in minutes. type RelTimeMagnitude struct { D time.Duration Format string DivBy time.Duration } var defaultMagnitudes = []RelTimeMagnitude{ {time.Second, "now", time.Second}, {2 * time.Second, "1 second %s", 1}, {time.Minute, "%d seconds %s", time.Second}, {2 * time.Minute, "1 minute %s", 1}, {time.Hour, "%d minutes %s", time.Minute}, {2 * time.Hour, "1 hour %s", 1}, {Day, "%d hours %s", time.Hour}, {2 * Day, "1 day %s", 1}, {Week, "%d days %s", Day}, {2 * Week, "1 week %s", 1}, {Month, "%d weeks %s", Week}, {2 * Month, "1 month %s", 1}, {Year, "%d months %s", Month}, {18 * Month, "1 year %s", 1}, {2 * Year, "2 years %s", 1}, {LongTime, "%d years %s", Year}, {math.MaxInt64, "a long while %s", 1}, } // RelTime formats a time into a relative string. // // It takes two times and two labels. In addition to the generic time // delta string (e.g. 5 minutes), the labels are used applied so that // the label corresponding to the smaller time is applied. // // RelTime(timeInPast, timeInFuture, "earlier", "later") -> "3 weeks earlier" func RelTime(a, b time.Time, albl, blbl string) string { return CustomRelTime(a, b, albl, blbl, defaultMagnitudes) } // CustomRelTime formats a time into a relative string. // // It takes two times two labels and a table of relative time formats. // In addition to the generic time delta string (e.g. 5 minutes), the // labels are used applied so that the label corresponding to the // smaller time is applied. func CustomRelTime(a, b time.Time, albl, blbl string, magnitudes []RelTimeMagnitude) string { lbl := albl diff := b.Sub(a) if a.After(b) { lbl = blbl diff = a.Sub(b) } n := sort.Search(len(magnitudes), func(i int) bool { return magnitudes[i].D > diff }) if n >= len(magnitudes) { n = len(magnitudes) - 1 } mag := magnitudes[n] args := []interface{}{} escaped := false for _, ch := range mag.Format { if escaped { switch ch { case 's': args = append(args, lbl) case 'd': args = append(args, diff/mag.DivBy) } escaped = false } else { escaped = ch == '%' } } return fmt.Sprintf(mag.Format, args...) }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/dustin/go-humanize/bytes.go
vendor/github.com/dustin/go-humanize/bytes.go
package humanize import ( "fmt" "math" "strconv" "strings" "unicode" ) // IEC Sizes. // kibis of bits const ( Byte = 1 << (iota * 10) KiByte MiByte GiByte TiByte PiByte EiByte ) // SI Sizes. const ( IByte = 1 KByte = IByte * 1000 MByte = KByte * 1000 GByte = MByte * 1000 TByte = GByte * 1000 PByte = TByte * 1000 EByte = PByte * 1000 ) var bytesSizeTable = map[string]uint64{ "b": Byte, "kib": KiByte, "kb": KByte, "mib": MiByte, "mb": MByte, "gib": GiByte, "gb": GByte, "tib": TiByte, "tb": TByte, "pib": PiByte, "pb": PByte, "eib": EiByte, "eb": EByte, // Without suffix "": Byte, "ki": KiByte, "k": KByte, "mi": MiByte, "m": MByte, "gi": GiByte, "g": GByte, "ti": TiByte, "t": TByte, "pi": PiByte, "p": PByte, "ei": EiByte, "e": EByte, } func logn(n, b float64) float64 { return math.Log(n) / math.Log(b) } func humanateBytes(s uint64, base float64, sizes []string) string { if s < 10 { return fmt.Sprintf("%d B", s) } e := math.Floor(logn(float64(s), base)) suffix := sizes[int(e)] val := math.Floor(float64(s)/math.Pow(base, e)*10+0.5) / 10 f := "%.0f %s" if val < 10 { f = "%.1f %s" } return fmt.Sprintf(f, val, suffix) } // Bytes produces a human readable representation of an SI size. // // See also: ParseBytes. // // Bytes(82854982) -> 83 MB func Bytes(s uint64) string { sizes := []string{"B", "kB", "MB", "GB", "TB", "PB", "EB"} return humanateBytes(s, 1000, sizes) } // IBytes produces a human readable representation of an IEC size. // // See also: ParseBytes. // // IBytes(82854982) -> 79 MiB func IBytes(s uint64) string { sizes := []string{"B", "KiB", "MiB", "GiB", "TiB", "PiB", "EiB"} return humanateBytes(s, 1024, sizes) } // ParseBytes parses a string representation of bytes into the number // of bytes it represents. // // See Also: Bytes, IBytes. // // ParseBytes("42 MB") -> 42000000, nil // ParseBytes("42 mib") -> 44040192, nil func ParseBytes(s string) (uint64, error) { lastDigit := 0 hasComma := false for _, r := range s { if !(unicode.IsDigit(r) || r == '.' || r == ',') { break } if r == ',' { hasComma = true } lastDigit++ } num := s[:lastDigit] if hasComma { num = strings.Replace(num, ",", "", -1) } f, err := strconv.ParseFloat(num, 64) if err != nil { return 0, err } extra := strings.ToLower(strings.TrimSpace(s[lastDigit:])) if m, ok := bytesSizeTable[extra]; ok { f *= float64(m) if f >= math.MaxUint64 { return 0, fmt.Errorf("too large: %v", s) } return uint64(f), nil } return 0, fmt.Errorf("unhandled size name: %v", extra) }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/dustin/go-humanize/number.go
vendor/github.com/dustin/go-humanize/number.go
package humanize /* Slightly adapted from the source to fit go-humanize. Author: https://github.com/gorhill Source: https://gist.github.com/gorhill/5285193 */ import ( "math" "strconv" ) var ( renderFloatPrecisionMultipliers = [...]float64{ 1, 10, 100, 1000, 10000, 100000, 1000000, 10000000, 100000000, 1000000000, } renderFloatPrecisionRounders = [...]float64{ 0.5, 0.05, 0.005, 0.0005, 0.00005, 0.000005, 0.0000005, 0.00000005, 0.000000005, 0.0000000005, } ) // FormatFloat produces a formatted number as string based on the following user-specified criteria: // * thousands separator // * decimal separator // * decimal precision // // Usage: s := RenderFloat(format, n) // The format parameter tells how to render the number n. // // See examples: http://play.golang.org/p/LXc1Ddm1lJ // // Examples of format strings, given n = 12345.6789: // "#,###.##" => "12,345.67" // "#,###." => "12,345" // "#,###" => "12345,678" // "#\u202F###,##" => "12 345,68" // "#.###,###### => 12.345,678900 // "" (aka default format) => 12,345.67 // // The highest precision allowed is 9 digits after the decimal symbol. // There is also a version for integer number, FormatInteger(), // which is convenient for calls within template. func FormatFloat(format string, n float64) string { // Special cases: // NaN = "NaN" // +Inf = "+Infinity" // -Inf = "-Infinity" if math.IsNaN(n) { return "NaN" } if n > math.MaxFloat64 { return "Infinity" } if n < (0.0 - math.MaxFloat64) { return "-Infinity" } // default format precision := 2 decimalStr := "." thousandStr := "," positiveStr := "" negativeStr := "-" if len(format) > 0 { format := []rune(format) // If there is an explicit format directive, // then default values are these: precision = 9 thousandStr = "" // collect indices of meaningful formatting directives formatIndx := []int{} for i, char := range format { if char != '#' && char != '0' { formatIndx = append(formatIndx, i) } } if len(formatIndx) > 0 { // Directive at index 0: // Must be a '+' // Raise an error if not the case // index: 0123456789 // +0.000,000 // +000,000.0 // +0000.00 // +0000 if formatIndx[0] == 0 { if format[formatIndx[0]] != '+' { panic("RenderFloat(): invalid positive sign directive") } positiveStr = "+" formatIndx = formatIndx[1:] } // Two directives: // First is thousands separator // Raise an error if not followed by 3-digit // 0123456789 // 0.000,000 // 000,000.00 if len(formatIndx) == 2 { if (formatIndx[1] - formatIndx[0]) != 4 { panic("RenderFloat(): thousands separator directive must be followed by 3 digit-specifiers") } thousandStr = string(format[formatIndx[0]]) formatIndx = formatIndx[1:] } // One directive: // Directive is decimal separator // The number of digit-specifier following the separator indicates wanted precision // 0123456789 // 0.00 // 000,0000 if len(formatIndx) == 1 { decimalStr = string(format[formatIndx[0]]) precision = len(format) - formatIndx[0] - 1 } } } // generate sign part var signStr string if n >= 0.000000001 { signStr = positiveStr } else if n <= -0.000000001 { signStr = negativeStr n = -n } else { signStr = "" n = 0.0 } // split number into integer and fractional parts intf, fracf := math.Modf(n + renderFloatPrecisionRounders[precision]) // generate integer part string intStr := strconv.FormatInt(int64(intf), 10) // add thousand separator if required if len(thousandStr) > 0 { for i := len(intStr); i > 3; { i -= 3 intStr = intStr[:i] + thousandStr + intStr[i:] } } // no fractional part, we can leave now if precision == 0 { return signStr + intStr } // generate fractional part fracStr := strconv.Itoa(int(fracf * renderFloatPrecisionMultipliers[precision])) // may need padding if len(fracStr) < precision { fracStr = "000000000000000"[:precision-len(fracStr)] + fracStr } return signStr + intStr + decimalStr + fracStr } // FormatInteger produces a formatted number as string. // See FormatFloat. func FormatInteger(format string, n int) string { return FormatFloat(format, float64(n)) }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/dustin/go-humanize/bigbytes.go
vendor/github.com/dustin/go-humanize/bigbytes.go
package humanize import ( "fmt" "math/big" "strings" "unicode" ) var ( bigIECExp = big.NewInt(1024) // BigByte is one byte in bit.Ints BigByte = big.NewInt(1) // BigKiByte is 1,024 bytes in bit.Ints BigKiByte = (&big.Int{}).Mul(BigByte, bigIECExp) // BigMiByte is 1,024 k bytes in bit.Ints BigMiByte = (&big.Int{}).Mul(BigKiByte, bigIECExp) // BigGiByte is 1,024 m bytes in bit.Ints BigGiByte = (&big.Int{}).Mul(BigMiByte, bigIECExp) // BigTiByte is 1,024 g bytes in bit.Ints BigTiByte = (&big.Int{}).Mul(BigGiByte, bigIECExp) // BigPiByte is 1,024 t bytes in bit.Ints BigPiByte = (&big.Int{}).Mul(BigTiByte, bigIECExp) // BigEiByte is 1,024 p bytes in bit.Ints BigEiByte = (&big.Int{}).Mul(BigPiByte, bigIECExp) // BigZiByte is 1,024 e bytes in bit.Ints BigZiByte = (&big.Int{}).Mul(BigEiByte, bigIECExp) // BigYiByte is 1,024 z bytes in bit.Ints BigYiByte = (&big.Int{}).Mul(BigZiByte, bigIECExp) // BigRiByte is 1,024 y bytes in bit.Ints BigRiByte = (&big.Int{}).Mul(BigYiByte, bigIECExp) // BigQiByte is 1,024 r bytes in bit.Ints BigQiByte = (&big.Int{}).Mul(BigRiByte, bigIECExp) ) var ( bigSIExp = big.NewInt(1000) // BigSIByte is one SI byte in big.Ints BigSIByte = big.NewInt(1) // BigKByte is 1,000 SI bytes in big.Ints BigKByte = (&big.Int{}).Mul(BigSIByte, bigSIExp) // BigMByte is 1,000 SI k bytes in big.Ints BigMByte = (&big.Int{}).Mul(BigKByte, bigSIExp) // BigGByte is 1,000 SI m bytes in big.Ints BigGByte = (&big.Int{}).Mul(BigMByte, bigSIExp) // BigTByte is 1,000 SI g bytes in big.Ints BigTByte = (&big.Int{}).Mul(BigGByte, bigSIExp) // BigPByte is 1,000 SI t bytes in big.Ints BigPByte = (&big.Int{}).Mul(BigTByte, bigSIExp) // BigEByte is 1,000 SI p bytes in big.Ints BigEByte = (&big.Int{}).Mul(BigPByte, bigSIExp) // BigZByte is 1,000 SI e bytes in big.Ints BigZByte = (&big.Int{}).Mul(BigEByte, bigSIExp) // BigYByte is 1,000 SI z bytes in big.Ints BigYByte = (&big.Int{}).Mul(BigZByte, bigSIExp) // BigRByte is 1,000 SI y bytes in big.Ints BigRByte = (&big.Int{}).Mul(BigYByte, bigSIExp) // BigQByte is 1,000 SI r bytes in big.Ints BigQByte = (&big.Int{}).Mul(BigRByte, bigSIExp) ) var bigBytesSizeTable = map[string]*big.Int{ "b": BigByte, "kib": BigKiByte, "kb": BigKByte, "mib": BigMiByte, "mb": BigMByte, "gib": BigGiByte, "gb": BigGByte, "tib": BigTiByte, "tb": BigTByte, "pib": BigPiByte, "pb": BigPByte, "eib": BigEiByte, "eb": BigEByte, "zib": BigZiByte, "zb": BigZByte, "yib": BigYiByte, "yb": BigYByte, "rib": BigRiByte, "rb": BigRByte, "qib": BigQiByte, "qb": BigQByte, // Without suffix "": BigByte, "ki": BigKiByte, "k": BigKByte, "mi": BigMiByte, "m": BigMByte, "gi": BigGiByte, "g": BigGByte, "ti": BigTiByte, "t": BigTByte, "pi": BigPiByte, "p": BigPByte, "ei": BigEiByte, "e": BigEByte, "z": BigZByte, "zi": BigZiByte, "y": BigYByte, "yi": BigYiByte, "r": BigRByte, "ri": BigRiByte, "q": BigQByte, "qi": BigQiByte, } var ten = big.NewInt(10) func humanateBigBytes(s, base *big.Int, sizes []string) string { if s.Cmp(ten) < 0 { return fmt.Sprintf("%d B", s) } c := (&big.Int{}).Set(s) val, mag := oomm(c, base, len(sizes)-1) suffix := sizes[mag] f := "%.0f %s" if val < 10 { f = "%.1f %s" } return fmt.Sprintf(f, val, suffix) } // BigBytes produces a human readable representation of an SI size. // // See also: ParseBigBytes. // // BigBytes(82854982) -> 83 MB func BigBytes(s *big.Int) string { sizes := []string{"B", "kB", "MB", "GB", "TB", "PB", "EB", "ZB", "YB", "RB", "QB"} return humanateBigBytes(s, bigSIExp, sizes) } // BigIBytes produces a human readable representation of an IEC size. // // See also: ParseBigBytes. // // BigIBytes(82854982) -> 79 MiB func BigIBytes(s *big.Int) string { sizes := []string{"B", "KiB", "MiB", "GiB", "TiB", "PiB", "EiB", "ZiB", "YiB", "RiB", "QiB"} return humanateBigBytes(s, bigIECExp, sizes) } // ParseBigBytes parses a string representation of bytes into the number // of bytes it represents. // // See also: BigBytes, BigIBytes. // // ParseBigBytes("42 MB") -> 42000000, nil // ParseBigBytes("42 mib") -> 44040192, nil func ParseBigBytes(s string) (*big.Int, error) { lastDigit := 0 hasComma := false for _, r := range s { if !(unicode.IsDigit(r) || r == '.' || r == ',') { break } if r == ',' { hasComma = true } lastDigit++ } num := s[:lastDigit] if hasComma { num = strings.Replace(num, ",", "", -1) } val := &big.Rat{} _, err := fmt.Sscanf(num, "%f", val) if err != nil { return nil, err } extra := strings.ToLower(strings.TrimSpace(s[lastDigit:])) if m, ok := bigBytesSizeTable[extra]; ok { mv := (&big.Rat{}).SetInt(m) val.Mul(val, mv) rv := &big.Int{} rv.Div(val.Num(), val.Denom()) return rv, nil } return nil, fmt.Errorf("unhandled size name: %v", extra) }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/proc_environ.go
vendor/github.com/prometheus/procfs/proc_environ.go
// Copyright 2019 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package procfs import ( "strings" "github.com/prometheus/procfs/internal/util" ) // Environ reads process environments from `/proc/<pid>/environ`. func (p Proc) Environ() ([]string, error) { environments := make([]string, 0) data, err := util.ReadFileNoStat(p.path("environ")) if err != nil { return environments, err } environments = strings.Split(string(data), "\000") if len(environments) > 0 { environments = environments[:len(environments)-1] } return environments, nil }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/net_softnet.go
vendor/github.com/prometheus/procfs/net_softnet.go
// Copyright 2019 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package procfs import ( "bufio" "bytes" "fmt" "io" "strconv" "strings" "github.com/prometheus/procfs/internal/util" ) // For the proc file format details, // See: // * Linux 2.6.23 https://elixir.bootlin.com/linux/v2.6.23/source/net/core/dev.c#L2343 // * Linux 2.6.39 https://elixir.bootlin.com/linux/v2.6.39/source/net/core/dev.c#L4086 // * Linux 4.18 https://elixir.bootlin.com/linux/v4.18/source/net/core/net-procfs.c#L162 // * Linux 5.14 https://elixir.bootlin.com/linux/v5.14/source/net/core/net-procfs.c#L169 // SoftnetStat contains a single row of data from /proc/net/softnet_stat. type SoftnetStat struct { // Number of processed packets. Processed uint32 // Number of dropped packets. Dropped uint32 // Number of times processing packets ran out of quota. TimeSqueezed uint32 // Number of collision occur while obtaining device lock while transmitting. CPUCollision uint32 // Number of times cpu woken up received_rps. ReceivedRps uint32 // number of times flow limit has been reached. FlowLimitCount uint32 // Softnet backlog status. SoftnetBacklogLen uint32 // CPU id owning this softnet_data. Index uint32 // softnet_data's Width. Width int } var softNetProcFile = "net/softnet_stat" // NetSoftnetStat reads data from /proc/net/softnet_stat. func (fs FS) NetSoftnetStat() ([]SoftnetStat, error) { b, err := util.ReadFileNoStat(fs.proc.Path(softNetProcFile)) if err != nil { return nil, err } entries, err := parseSoftnet(bytes.NewReader(b)) if err != nil { return nil, fmt.Errorf("%w: /proc/net/softnet_stat: %w", ErrFileParse, err) } return entries, nil } func parseSoftnet(r io.Reader) ([]SoftnetStat, error) { const minColumns = 9 s := bufio.NewScanner(r) var stats []SoftnetStat cpuIndex := 0 for s.Scan() { columns := strings.Fields(s.Text()) width := len(columns) softnetStat := SoftnetStat{} if width < minColumns { return nil, fmt.Errorf("%w: detected %d columns, but expected at least %d", ErrFileParse, width, minColumns) } // Linux 2.6.23 https://elixir.bootlin.com/linux/v2.6.23/source/net/core/dev.c#L2347 if width >= minColumns { us, err := parseHexUint32s(columns[0:9]) if err != nil { return nil, err } softnetStat.Processed = us[0] softnetStat.Dropped = us[1] softnetStat.TimeSqueezed = us[2] softnetStat.CPUCollision = us[8] } // Linux 2.6.39 https://elixir.bootlin.com/linux/v2.6.39/source/net/core/dev.c#L4086 if width >= 10 { us, err := parseHexUint32s(columns[9:10]) if err != nil { return nil, err } softnetStat.ReceivedRps = us[0] } // Linux 4.18 https://elixir.bootlin.com/linux/v4.18/source/net/core/net-procfs.c#L162 if width >= 11 { us, err := parseHexUint32s(columns[10:11]) if err != nil { return nil, err } softnetStat.FlowLimitCount = us[0] } // Linux 5.14 https://elixir.bootlin.com/linux/v5.14/source/net/core/net-procfs.c#L169 if width >= 13 { us, err := parseHexUint32s(columns[11:13]) if err != nil { return nil, err } softnetStat.SoftnetBacklogLen = us[0] softnetStat.Index = us[1] } else { // For older kernels, create the Index based on the scan line number. softnetStat.Index = uint32(cpuIndex) } softnetStat.Width = width stats = append(stats, softnetStat) cpuIndex++ } return stats, nil } func parseHexUint32s(ss []string) ([]uint32, error) { us := make([]uint32, 0, len(ss)) for _, s := range ss { u, err := strconv.ParseUint(s, 16, 32) if err != nil { return nil, err } us = append(us, uint32(u)) } return us, nil }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/vm.go
vendor/github.com/prometheus/procfs/vm.go
// Copyright 2019 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. //go:build !windows // +build !windows package procfs import ( "fmt" "os" "path/filepath" "strings" "github.com/prometheus/procfs/internal/util" ) // The VM interface is described at // // https://www.kernel.org/doc/Documentation/sysctl/vm.txt // // Each setting is exposed as a single file. // Each file contains one line with a single numerical value, except lowmem_reserve_ratio which holds an array // and numa_zonelist_order (deprecated) which is a string. type VM struct { AdminReserveKbytes *int64 // /proc/sys/vm/admin_reserve_kbytes BlockDump *int64 // /proc/sys/vm/block_dump CompactUnevictableAllowed *int64 // /proc/sys/vm/compact_unevictable_allowed DirtyBackgroundBytes *int64 // /proc/sys/vm/dirty_background_bytes DirtyBackgroundRatio *int64 // /proc/sys/vm/dirty_background_ratio DirtyBytes *int64 // /proc/sys/vm/dirty_bytes DirtyExpireCentisecs *int64 // /proc/sys/vm/dirty_expire_centisecs DirtyRatio *int64 // /proc/sys/vm/dirty_ratio DirtytimeExpireSeconds *int64 // /proc/sys/vm/dirtytime_expire_seconds DirtyWritebackCentisecs *int64 // /proc/sys/vm/dirty_writeback_centisecs DropCaches *int64 // /proc/sys/vm/drop_caches ExtfragThreshold *int64 // /proc/sys/vm/extfrag_threshold HugetlbShmGroup *int64 // /proc/sys/vm/hugetlb_shm_group LaptopMode *int64 // /proc/sys/vm/laptop_mode LegacyVaLayout *int64 // /proc/sys/vm/legacy_va_layout LowmemReserveRatio []*int64 // /proc/sys/vm/lowmem_reserve_ratio MaxMapCount *int64 // /proc/sys/vm/max_map_count MemoryFailureEarlyKill *int64 // /proc/sys/vm/memory_failure_early_kill MemoryFailureRecovery *int64 // /proc/sys/vm/memory_failure_recovery MinFreeKbytes *int64 // /proc/sys/vm/min_free_kbytes MinSlabRatio *int64 // /proc/sys/vm/min_slab_ratio MinUnmappedRatio *int64 // /proc/sys/vm/min_unmapped_ratio MmapMinAddr *int64 // /proc/sys/vm/mmap_min_addr NrHugepages *int64 // /proc/sys/vm/nr_hugepages NrHugepagesMempolicy *int64 // /proc/sys/vm/nr_hugepages_mempolicy NrOvercommitHugepages *int64 // /proc/sys/vm/nr_overcommit_hugepages NumaStat *int64 // /proc/sys/vm/numa_stat NumaZonelistOrder string // /proc/sys/vm/numa_zonelist_order OomDumpTasks *int64 // /proc/sys/vm/oom_dump_tasks OomKillAllocatingTask *int64 // /proc/sys/vm/oom_kill_allocating_task OvercommitKbytes *int64 // /proc/sys/vm/overcommit_kbytes OvercommitMemory *int64 // /proc/sys/vm/overcommit_memory OvercommitRatio *int64 // /proc/sys/vm/overcommit_ratio PageCluster *int64 // /proc/sys/vm/page-cluster PanicOnOom *int64 // /proc/sys/vm/panic_on_oom PercpuPagelistFraction *int64 // /proc/sys/vm/percpu_pagelist_fraction StatInterval *int64 // /proc/sys/vm/stat_interval Swappiness *int64 // /proc/sys/vm/swappiness UserReserveKbytes *int64 // /proc/sys/vm/user_reserve_kbytes VfsCachePressure *int64 // /proc/sys/vm/vfs_cache_pressure WatermarkBoostFactor *int64 // /proc/sys/vm/watermark_boost_factor WatermarkScaleFactor *int64 // /proc/sys/vm/watermark_scale_factor ZoneReclaimMode *int64 // /proc/sys/vm/zone_reclaim_mode } // VM reads the VM statistics from the specified `proc` filesystem. func (fs FS) VM() (*VM, error) { path := fs.proc.Path("sys/vm") file, err := os.Stat(path) if err != nil { return nil, err } if !file.Mode().IsDir() { return nil, fmt.Errorf("%w: %s is not a directory", ErrFileRead, path) } files, err := os.ReadDir(path) if err != nil { return nil, err } var vm VM for _, f := range files { if f.IsDir() { continue } name := filepath.Join(path, f.Name()) // ignore errors on read, as there are some write only // in /proc/sys/vm value, err := util.SysReadFile(name) if err != nil { continue } vp := util.NewValueParser(value) switch f.Name() { case "admin_reserve_kbytes": vm.AdminReserveKbytes = vp.PInt64() case "block_dump": vm.BlockDump = vp.PInt64() case "compact_unevictable_allowed": vm.CompactUnevictableAllowed = vp.PInt64() case "dirty_background_bytes": vm.DirtyBackgroundBytes = vp.PInt64() case "dirty_background_ratio": vm.DirtyBackgroundRatio = vp.PInt64() case "dirty_bytes": vm.DirtyBytes = vp.PInt64() case "dirty_expire_centisecs": vm.DirtyExpireCentisecs = vp.PInt64() case "dirty_ratio": vm.DirtyRatio = vp.PInt64() case "dirtytime_expire_seconds": vm.DirtytimeExpireSeconds = vp.PInt64() case "dirty_writeback_centisecs": vm.DirtyWritebackCentisecs = vp.PInt64() case "drop_caches": vm.DropCaches = vp.PInt64() case "extfrag_threshold": vm.ExtfragThreshold = vp.PInt64() case "hugetlb_shm_group": vm.HugetlbShmGroup = vp.PInt64() case "laptop_mode": vm.LaptopMode = vp.PInt64() case "legacy_va_layout": vm.LegacyVaLayout = vp.PInt64() case "lowmem_reserve_ratio": stringSlice := strings.Fields(value) pint64Slice := make([]*int64, 0, len(stringSlice)) for _, value := range stringSlice { vp := util.NewValueParser(value) pint64Slice = append(pint64Slice, vp.PInt64()) } vm.LowmemReserveRatio = pint64Slice case "max_map_count": vm.MaxMapCount = vp.PInt64() case "memory_failure_early_kill": vm.MemoryFailureEarlyKill = vp.PInt64() case "memory_failure_recovery": vm.MemoryFailureRecovery = vp.PInt64() case "min_free_kbytes": vm.MinFreeKbytes = vp.PInt64() case "min_slab_ratio": vm.MinSlabRatio = vp.PInt64() case "min_unmapped_ratio": vm.MinUnmappedRatio = vp.PInt64() case "mmap_min_addr": vm.MmapMinAddr = vp.PInt64() case "nr_hugepages": vm.NrHugepages = vp.PInt64() case "nr_hugepages_mempolicy": vm.NrHugepagesMempolicy = vp.PInt64() case "nr_overcommit_hugepages": vm.NrOvercommitHugepages = vp.PInt64() case "numa_stat": vm.NumaStat = vp.PInt64() case "numa_zonelist_order": vm.NumaZonelistOrder = value case "oom_dump_tasks": vm.OomDumpTasks = vp.PInt64() case "oom_kill_allocating_task": vm.OomKillAllocatingTask = vp.PInt64() case "overcommit_kbytes": vm.OvercommitKbytes = vp.PInt64() case "overcommit_memory": vm.OvercommitMemory = vp.PInt64() case "overcommit_ratio": vm.OvercommitRatio = vp.PInt64() case "page-cluster": vm.PageCluster = vp.PInt64() case "panic_on_oom": vm.PanicOnOom = vp.PInt64() case "percpu_pagelist_fraction": vm.PercpuPagelistFraction = vp.PInt64() case "stat_interval": vm.StatInterval = vp.PInt64() case "swappiness": vm.Swappiness = vp.PInt64() case "user_reserve_kbytes": vm.UserReserveKbytes = vp.PInt64() case "vfs_cache_pressure": vm.VfsCachePressure = vp.PInt64() case "watermark_boost_factor": vm.WatermarkBoostFactor = vp.PInt64() case "watermark_scale_factor": vm.WatermarkScaleFactor = vp.PInt64() case "zone_reclaim_mode": vm.ZoneReclaimMode = vp.PInt64() } if err := vp.Err(); err != nil { return nil, err } } return &vm, nil }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/cpuinfo_loong64.go
vendor/github.com/prometheus/procfs/cpuinfo_loong64.go
// Copyright 2022 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. //go:build linux // +build linux package procfs var parseCPUInfo = parseCPUInfoLoong
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/proc.go
vendor/github.com/prometheus/procfs/proc.go
// Copyright 2018 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package procfs import ( "bytes" "errors" "fmt" "io" "os" "strconv" "strings" "github.com/prometheus/procfs/internal/util" ) // Proc provides information about a running process. type Proc struct { // The process ID. PID int fs FS } // Procs represents a list of Proc structs. type Procs []Proc var ( ErrFileParse = errors.New("Error Parsing File") ErrFileRead = errors.New("Error Reading File") ErrMountPoint = errors.New("Error Accessing Mount point") ) func (p Procs) Len() int { return len(p) } func (p Procs) Swap(i, j int) { p[i], p[j] = p[j], p[i] } func (p Procs) Less(i, j int) bool { return p[i].PID < p[j].PID } // Self returns a process for the current process read via /proc/self. func Self() (Proc, error) { fs, err := NewFS(DefaultMountPoint) if err != nil || errors.Unwrap(err) == ErrMountPoint { return Proc{}, err } return fs.Self() } // NewProc returns a process for the given pid under /proc. func NewProc(pid int) (Proc, error) { fs, err := NewFS(DefaultMountPoint) if err != nil { return Proc{}, err } return fs.Proc(pid) } // AllProcs returns a list of all currently available processes under /proc. func AllProcs() (Procs, error) { fs, err := NewFS(DefaultMountPoint) if err != nil { return Procs{}, err } return fs.AllProcs() } // Self returns a process for the current process. func (fs FS) Self() (Proc, error) { p, err := os.Readlink(fs.proc.Path("self")) if err != nil { return Proc{}, err } pid, err := strconv.Atoi(strings.Replace(p, string(fs.proc), "", -1)) if err != nil { return Proc{}, err } return fs.Proc(pid) } // NewProc returns a process for the given pid. // // Deprecated: Use fs.Proc() instead. func (fs FS) NewProc(pid int) (Proc, error) { return fs.Proc(pid) } // Proc returns a process for the given pid. func (fs FS) Proc(pid int) (Proc, error) { if _, err := os.Stat(fs.proc.Path(strconv.Itoa(pid))); err != nil { return Proc{}, err } return Proc{PID: pid, fs: fs}, nil } // AllProcs returns a list of all currently available processes. func (fs FS) AllProcs() (Procs, error) { d, err := os.Open(fs.proc.Path()) if err != nil { return Procs{}, err } defer d.Close() names, err := d.Readdirnames(-1) if err != nil { return Procs{}, fmt.Errorf("%w: Cannot read file: %v: %w", ErrFileRead, names, err) } p := Procs{} for _, n := range names { pid, err := strconv.ParseInt(n, 10, 64) if err != nil { continue } p = append(p, Proc{PID: int(pid), fs: fs}) } return p, nil } // CmdLine returns the command line of a process. func (p Proc) CmdLine() ([]string, error) { data, err := util.ReadFileNoStat(p.path("cmdline")) if err != nil { return nil, err } if len(data) < 1 { return []string{}, nil } return strings.Split(string(bytes.TrimRight(data, "\x00")), "\x00"), nil } // Wchan returns the wchan (wait channel) of a process. func (p Proc) Wchan() (string, error) { f, err := os.Open(p.path("wchan")) if err != nil { return "", err } defer f.Close() data, err := io.ReadAll(f) if err != nil { return "", err } wchan := string(data) if wchan == "" || wchan == "0" { return "", nil } return wchan, nil } // Comm returns the command name of a process. func (p Proc) Comm() (string, error) { data, err := util.ReadFileNoStat(p.path("comm")) if err != nil { return "", err } return strings.TrimSpace(string(data)), nil } // Executable returns the absolute path of the executable command of a process. func (p Proc) Executable() (string, error) { exe, err := os.Readlink(p.path("exe")) if os.IsNotExist(err) { return "", nil } return exe, err } // Cwd returns the absolute path to the current working directory of the process. func (p Proc) Cwd() (string, error) { wd, err := os.Readlink(p.path("cwd")) if os.IsNotExist(err) { return "", nil } return wd, err } // RootDir returns the absolute path to the process's root directory (as set by chroot). func (p Proc) RootDir() (string, error) { rdir, err := os.Readlink(p.path("root")) if os.IsNotExist(err) { return "", nil } return rdir, err } // FileDescriptors returns the currently open file descriptors of a process. func (p Proc) FileDescriptors() ([]uintptr, error) { names, err := p.fileDescriptors() if err != nil { return nil, err } fds := make([]uintptr, len(names)) for i, n := range names { fd, err := strconv.ParseInt(n, 10, 32) if err != nil { return nil, fmt.Errorf("%w: Cannot parse line: %v: %w", ErrFileParse, i, err) } fds[i] = uintptr(fd) } return fds, nil } // FileDescriptorTargets returns the targets of all file descriptors of a process. // If a file descriptor is not a symlink to a file (like a socket), that value will be the empty string. func (p Proc) FileDescriptorTargets() ([]string, error) { names, err := p.fileDescriptors() if err != nil { return nil, err } targets := make([]string, len(names)) for i, name := range names { target, err := os.Readlink(p.path("fd", name)) if err == nil { targets[i] = target } } return targets, nil } // FileDescriptorsLen returns the number of currently open file descriptors of // a process. func (p Proc) FileDescriptorsLen() (int, error) { // Use fast path if available (Linux v6.2): https://github.com/torvalds/linux/commit/f1f1f2569901 if p.fs.isReal { stat, err := os.Stat(p.path("fd")) if err != nil { return 0, err } size := stat.Size() if size > 0 { return int(size), nil } } fds, err := p.fileDescriptors() if err != nil { return 0, err } return len(fds), nil } // MountStats retrieves statistics and configuration for mount points in a // process's namespace. func (p Proc) MountStats() ([]*Mount, error) { f, err := os.Open(p.path("mountstats")) if err != nil { return nil, err } defer f.Close() return parseMountStats(f) } // MountInfo retrieves mount information for mount points in a // process's namespace. // It supplies information missing in `/proc/self/mounts` and // fixes various other problems with that file too. func (p Proc) MountInfo() ([]*MountInfo, error) { data, err := util.ReadFileNoStat(p.path("mountinfo")) if err != nil { return nil, err } return parseMountInfo(data) } func (p Proc) fileDescriptors() ([]string, error) { d, err := os.Open(p.path("fd")) if err != nil { return nil, err } defer d.Close() names, err := d.Readdirnames(-1) if err != nil { return nil, fmt.Errorf("%w: Cannot read file: %v: %w", ErrFileRead, names, err) } return names, nil } func (p Proc) path(pa ...string) string { return p.fs.proc.Path(append([]string{strconv.Itoa(p.PID)}, pa...)...) } // FileDescriptorsInfo retrieves information about all file descriptors of // the process. func (p Proc) FileDescriptorsInfo() (ProcFDInfos, error) { names, err := p.fileDescriptors() if err != nil { return nil, err } var fdinfos ProcFDInfos for _, n := range names { fdinfo, err := p.FDInfo(n) if err != nil { continue } fdinfos = append(fdinfos, *fdinfo) } return fdinfos, nil } // Schedstat returns task scheduling information for the process. func (p Proc) Schedstat() (ProcSchedstat, error) { contents, err := os.ReadFile(p.path("schedstat")) if err != nil { return ProcSchedstat{}, err } return parseProcSchedstat(string(contents)) }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/net_protocols.go
vendor/github.com/prometheus/procfs/net_protocols.go
// Copyright 2020 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package procfs import ( "bufio" "bytes" "fmt" "strconv" "strings" "github.com/prometheus/procfs/internal/util" ) // NetProtocolStats stores the contents from /proc/net/protocols. type NetProtocolStats map[string]NetProtocolStatLine // NetProtocolStatLine contains a single line parsed from /proc/net/protocols. We // only care about the first six columns as the rest are not likely to change // and only serve to provide a set of capabilities for each protocol. type NetProtocolStatLine struct { Name string // 0 The name of the protocol Size uint64 // 1 The size, in bytes, of a given protocol structure. e.g. sizeof(struct tcp_sock) or sizeof(struct unix_sock) Sockets int64 // 2 Number of sockets in use by this protocol Memory int64 // 3 Number of 4KB pages allocated by all sockets of this protocol Pressure int // 4 This is either yes, no, or NI (not implemented). For the sake of simplicity we treat NI as not experiencing memory pressure. MaxHeader uint64 // 5 Protocol specific max header size Slab bool // 6 Indicates whether or not memory is allocated from the SLAB ModuleName string // 7 The name of the module that implemented this protocol or "kernel" if not from a module Capabilities NetProtocolCapabilities } // NetProtocolCapabilities contains a list of capabilities for each protocol. type NetProtocolCapabilities struct { Close bool // 8 Connect bool // 9 Disconnect bool // 10 Accept bool // 11 IoCtl bool // 12 Init bool // 13 Destroy bool // 14 Shutdown bool // 15 SetSockOpt bool // 16 GetSockOpt bool // 17 SendMsg bool // 18 RecvMsg bool // 19 SendPage bool // 20 Bind bool // 21 BacklogRcv bool // 22 Hash bool // 23 UnHash bool // 24 GetPort bool // 25 EnterMemoryPressure bool // 26 } // NetProtocols reads stats from /proc/net/protocols and returns a map of // PortocolStatLine entries. As of this writing no official Linux Documentation // exists, however the source is fairly self-explanatory and the format seems // stable since its introduction in 2.6.12-rc2 // Linux 2.6.12-rc2 - https://elixir.bootlin.com/linux/v2.6.12-rc2/source/net/core/sock.c#L1452 // Linux 5.10 - https://elixir.bootlin.com/linux/v5.10.4/source/net/core/sock.c#L3586 func (fs FS) NetProtocols() (NetProtocolStats, error) { data, err := util.ReadFileNoStat(fs.proc.Path("net/protocols")) if err != nil { return NetProtocolStats{}, err } return parseNetProtocols(bufio.NewScanner(bytes.NewReader(data))) } func parseNetProtocols(s *bufio.Scanner) (NetProtocolStats, error) { nps := NetProtocolStats{} // Skip the header line s.Scan() for s.Scan() { line, err := nps.parseLine(s.Text()) if err != nil { return NetProtocolStats{}, err } nps[line.Name] = *line } return nps, nil } func (ps NetProtocolStats) parseLine(rawLine string) (*NetProtocolStatLine, error) { line := &NetProtocolStatLine{Capabilities: NetProtocolCapabilities{}} var err error const enabled = "yes" const disabled = "no" fields := strings.Fields(rawLine) line.Name = fields[0] line.Size, err = strconv.ParseUint(fields[1], 10, 64) if err != nil { return nil, err } line.Sockets, err = strconv.ParseInt(fields[2], 10, 64) if err != nil { return nil, err } line.Memory, err = strconv.ParseInt(fields[3], 10, 64) if err != nil { return nil, err } if fields[4] == enabled { line.Pressure = 1 } else if fields[4] == disabled { line.Pressure = 0 } else { line.Pressure = -1 } line.MaxHeader, err = strconv.ParseUint(fields[5], 10, 64) if err != nil { return nil, err } if fields[6] == enabled { line.Slab = true } else if fields[6] == disabled { line.Slab = false } else { return nil, fmt.Errorf("%w: capability for protocol: %s", ErrFileParse, line.Name) } line.ModuleName = fields[7] err = line.Capabilities.parseCapabilities(fields[8:]) if err != nil { return nil, err } return line, nil } func (pc *NetProtocolCapabilities) parseCapabilities(capabilities []string) error { // The capabilities are all bools so we can loop over to map them capabilityFields := [...]*bool{ &pc.Close, &pc.Connect, &pc.Disconnect, &pc.Accept, &pc.IoCtl, &pc.Init, &pc.Destroy, &pc.Shutdown, &pc.SetSockOpt, &pc.GetSockOpt, &pc.SendMsg, &pc.RecvMsg, &pc.SendPage, &pc.Bind, &pc.BacklogRcv, &pc.Hash, &pc.UnHash, &pc.GetPort, &pc.EnterMemoryPressure, } for i := 0; i < len(capabilities); i++ { if capabilities[i] == "y" { *capabilityFields[i] = true } else if capabilities[i] == "n" { *capabilityFields[i] = false } else { return fmt.Errorf("%w: capability block for protocol: position %d", ErrFileParse, i) } } return nil }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/proc_stat.go
vendor/github.com/prometheus/procfs/proc_stat.go
// Copyright 2018 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package procfs import ( "bytes" "fmt" "os" "github.com/prometheus/procfs/internal/util" ) // Originally, this USER_HZ value was dynamically retrieved via a sysconf call // which required cgo. However, that caused a lot of problems regarding // cross-compilation. Alternatives such as running a binary to determine the // value, or trying to derive it in some other way were all problematic. After // much research it was determined that USER_HZ is actually hardcoded to 100 on // all Go-supported platforms as of the time of this writing. This is why we // decided to hardcode it here as well. It is not impossible that there could // be systems with exceptions, but they should be very exotic edge cases, and // in that case, the worst outcome will be two misreported metrics. // // See also the following discussions: // // - https://github.com/prometheus/node_exporter/issues/52 // - https://github.com/prometheus/procfs/pull/2 // - http://stackoverflow.com/questions/17410841/how-does-user-hz-solve-the-jiffy-scaling-issue const userHZ = 100 // ProcStat provides status information about the process, // read from /proc/[pid]/stat. type ProcStat struct { // The process ID. PID int // The filename of the executable. Comm string // The process state. State string // The PID of the parent of this process. PPID int // The process group ID of the process. PGRP int // The session ID of the process. Session int // The controlling terminal of the process. TTY int // The ID of the foreground process group of the controlling terminal of // the process. TPGID int // The kernel flags word of the process. Flags uint // The number of minor faults the process has made which have not required // loading a memory page from disk. MinFlt uint // The number of minor faults that the process's waited-for children have // made. CMinFlt uint // The number of major faults the process has made which have required // loading a memory page from disk. MajFlt uint // The number of major faults that the process's waited-for children have // made. CMajFlt uint // Amount of time that this process has been scheduled in user mode, // measured in clock ticks. UTime uint // Amount of time that this process has been scheduled in kernel mode, // measured in clock ticks. STime uint // Amount of time that this process's waited-for children have been // scheduled in user mode, measured in clock ticks. CUTime int // Amount of time that this process's waited-for children have been // scheduled in kernel mode, measured in clock ticks. CSTime int // For processes running a real-time scheduling policy, this is the negated // scheduling priority, minus one. Priority int // The nice value, a value in the range 19 (low priority) to -20 (high // priority). Nice int // Number of threads in this process. NumThreads int // The time the process started after system boot, the value is expressed // in clock ticks. Starttime uint64 // Virtual memory size in bytes. VSize uint // Resident set size in pages. RSS int // Soft limit in bytes on the rss of the process. RSSLimit uint64 // CPU number last executed on. Processor uint // Real-time scheduling priority, a number in the range 1 to 99 for processes // scheduled under a real-time policy, or 0, for non-real-time processes. RTPriority uint // Scheduling policy. Policy uint // Aggregated block I/O delays, measured in clock ticks (centiseconds). DelayAcctBlkIOTicks uint64 // Guest time of the process (time spent running a virtual CPU for a guest // operating system), measured in clock ticks. GuestTime int // Guest time of the process's children, measured in clock ticks. CGuestTime int proc FS } // NewStat returns the current status information of the process. // // Deprecated: Use p.Stat() instead. func (p Proc) NewStat() (ProcStat, error) { return p.Stat() } // Stat returns the current status information of the process. func (p Proc) Stat() (ProcStat, error) { data, err := util.ReadFileNoStat(p.path("stat")) if err != nil { return ProcStat{}, err } var ( ignoreInt64 int64 ignoreUint64 uint64 s = ProcStat{PID: p.PID, proc: p.fs} l = bytes.Index(data, []byte("(")) r = bytes.LastIndex(data, []byte(")")) ) if l < 0 || r < 0 { return ProcStat{}, fmt.Errorf("%w: unexpected format, couldn't extract comm %q", ErrFileParse, data) } s.Comm = string(data[l+1 : r]) // Check the following resources for the details about the particular stat // fields and their data types: // * https://man7.org/linux/man-pages/man5/proc.5.html // * https://man7.org/linux/man-pages/man3/scanf.3.html _, err = fmt.Fscan( bytes.NewBuffer(data[r+2:]), &s.State, &s.PPID, &s.PGRP, &s.Session, &s.TTY, &s.TPGID, &s.Flags, &s.MinFlt, &s.CMinFlt, &s.MajFlt, &s.CMajFlt, &s.UTime, &s.STime, &s.CUTime, &s.CSTime, &s.Priority, &s.Nice, &s.NumThreads, &ignoreInt64, &s.Starttime, &s.VSize, &s.RSS, &s.RSSLimit, &ignoreUint64, &ignoreUint64, &ignoreUint64, &ignoreUint64, &ignoreUint64, &ignoreUint64, &ignoreUint64, &ignoreUint64, &ignoreUint64, &ignoreUint64, &ignoreUint64, &ignoreUint64, &ignoreInt64, &s.Processor, &s.RTPriority, &s.Policy, &s.DelayAcctBlkIOTicks, &s.GuestTime, &s.CGuestTime, ) if err != nil { return ProcStat{}, err } return s, nil } // VirtualMemory returns the virtual memory size in bytes. func (s ProcStat) VirtualMemory() uint { return s.VSize } // ResidentMemory returns the resident memory size in bytes. func (s ProcStat) ResidentMemory() int { return s.RSS * os.Getpagesize() } // StartTime returns the unix timestamp of the process in seconds. func (s ProcStat) StartTime() (float64, error) { stat, err := s.proc.Stat() if err != nil { return 0, err } return float64(stat.BootTime) + (float64(s.Starttime) / userHZ), nil } // CPUTime returns the total CPU user and system time in seconds. func (s ProcStat) CPUTime() float64 { return float64(s.UTime+s.STime) / userHZ }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/cpuinfo.go
vendor/github.com/prometheus/procfs/cpuinfo.go
// Copyright 2019 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. //go:build linux // +build linux package procfs import ( "bufio" "bytes" "errors" "fmt" "regexp" "strconv" "strings" "github.com/prometheus/procfs/internal/util" ) // CPUInfo contains general information about a system CPU found in /proc/cpuinfo. type CPUInfo struct { Processor uint VendorID string CPUFamily string Model string ModelName string Stepping string Microcode string CPUMHz float64 CacheSize string PhysicalID string Siblings uint CoreID string CPUCores uint APICID string InitialAPICID string FPU string FPUException string CPUIDLevel uint WP string Flags []string Bugs []string BogoMips float64 CLFlushSize uint CacheAlignment uint AddressSizes string PowerManagement string } var ( cpuinfoClockRegexp = regexp.MustCompile(`([\d.]+)`) cpuinfoS390XProcessorRegexp = regexp.MustCompile(`^processor\s+(\d+):.*`) ) // CPUInfo returns information about current system CPUs. // See https://www.kernel.org/doc/Documentation/filesystems/proc.txt func (fs FS) CPUInfo() ([]CPUInfo, error) { data, err := util.ReadFileNoStat(fs.proc.Path("cpuinfo")) if err != nil { return nil, err } return parseCPUInfo(data) } func parseCPUInfoX86(info []byte) ([]CPUInfo, error) { scanner := bufio.NewScanner(bytes.NewReader(info)) // find the first "processor" line firstLine := firstNonEmptyLine(scanner) if !strings.HasPrefix(firstLine, "processor") || !strings.Contains(firstLine, ":") { return nil, fmt.Errorf("%w: Cannot parse line: %q", ErrFileParse, firstLine) } field := strings.SplitN(firstLine, ": ", 2) v, err := strconv.ParseUint(field[1], 0, 32) if err != nil { return nil, err } firstcpu := CPUInfo{Processor: uint(v)} cpuinfo := []CPUInfo{firstcpu} i := 0 for scanner.Scan() { line := scanner.Text() if !strings.Contains(line, ":") { continue } field := strings.SplitN(line, ": ", 2) switch strings.TrimSpace(field[0]) { case "processor": cpuinfo = append(cpuinfo, CPUInfo{}) // start of the next processor i++ v, err := strconv.ParseUint(field[1], 0, 32) if err != nil { return nil, err } cpuinfo[i].Processor = uint(v) case "vendor", "vendor_id": cpuinfo[i].VendorID = field[1] case "cpu family": cpuinfo[i].CPUFamily = field[1] case "model": cpuinfo[i].Model = field[1] case "model name": cpuinfo[i].ModelName = field[1] case "stepping": cpuinfo[i].Stepping = field[1] case "microcode": cpuinfo[i].Microcode = field[1] case "cpu MHz": v, err := strconv.ParseFloat(field[1], 64) if err != nil { return nil, err } cpuinfo[i].CPUMHz = v case "cache size": cpuinfo[i].CacheSize = field[1] case "physical id": cpuinfo[i].PhysicalID = field[1] case "siblings": v, err := strconv.ParseUint(field[1], 0, 32) if err != nil { return nil, err } cpuinfo[i].Siblings = uint(v) case "core id": cpuinfo[i].CoreID = field[1] case "cpu cores": v, err := strconv.ParseUint(field[1], 0, 32) if err != nil { return nil, err } cpuinfo[i].CPUCores = uint(v) case "apicid": cpuinfo[i].APICID = field[1] case "initial apicid": cpuinfo[i].InitialAPICID = field[1] case "fpu": cpuinfo[i].FPU = field[1] case "fpu_exception": cpuinfo[i].FPUException = field[1] case "cpuid level": v, err := strconv.ParseUint(field[1], 0, 32) if err != nil { return nil, err } cpuinfo[i].CPUIDLevel = uint(v) case "wp": cpuinfo[i].WP = field[1] case "flags": cpuinfo[i].Flags = strings.Fields(field[1]) case "bugs": cpuinfo[i].Bugs = strings.Fields(field[1]) case "bogomips": v, err := strconv.ParseFloat(field[1], 64) if err != nil { return nil, err } cpuinfo[i].BogoMips = v case "clflush size": v, err := strconv.ParseUint(field[1], 0, 32) if err != nil { return nil, err } cpuinfo[i].CLFlushSize = uint(v) case "cache_alignment": v, err := strconv.ParseUint(field[1], 0, 32) if err != nil { return nil, err } cpuinfo[i].CacheAlignment = uint(v) case "address sizes": cpuinfo[i].AddressSizes = field[1] case "power management": cpuinfo[i].PowerManagement = field[1] } } return cpuinfo, nil } func parseCPUInfoARM(info []byte) ([]CPUInfo, error) { scanner := bufio.NewScanner(bytes.NewReader(info)) firstLine := firstNonEmptyLine(scanner) match, err := regexp.MatchString("^[Pp]rocessor", firstLine) if !match || !strings.Contains(firstLine, ":") { return nil, fmt.Errorf("%w: Cannot parse line: %q: %w", ErrFileParse, firstLine, err) } field := strings.SplitN(firstLine, ": ", 2) cpuinfo := []CPUInfo{} featuresLine := "" commonCPUInfo := CPUInfo{} i := 0 if strings.TrimSpace(field[0]) == "Processor" { commonCPUInfo = CPUInfo{ModelName: field[1]} i = -1 } else { v, err := strconv.ParseUint(field[1], 0, 32) if err != nil { return nil, err } firstcpu := CPUInfo{Processor: uint(v)} cpuinfo = []CPUInfo{firstcpu} } for scanner.Scan() { line := scanner.Text() if !strings.Contains(line, ":") { continue } field := strings.SplitN(line, ": ", 2) switch strings.TrimSpace(field[0]) { case "processor": cpuinfo = append(cpuinfo, commonCPUInfo) // start of the next processor i++ v, err := strconv.ParseUint(field[1], 0, 32) if err != nil { return nil, err } cpuinfo[i].Processor = uint(v) case "BogoMIPS": if i == -1 { cpuinfo = append(cpuinfo, commonCPUInfo) // There is only one processor i++ cpuinfo[i].Processor = 0 } v, err := strconv.ParseFloat(field[1], 64) if err != nil { return nil, err } cpuinfo[i].BogoMips = v case "Features": featuresLine = line case "model name": cpuinfo[i].ModelName = field[1] } } fields := strings.SplitN(featuresLine, ": ", 2) for i := range cpuinfo { cpuinfo[i].Flags = strings.Fields(fields[1]) } return cpuinfo, nil } func parseCPUInfoS390X(info []byte) ([]CPUInfo, error) { scanner := bufio.NewScanner(bytes.NewReader(info)) firstLine := firstNonEmptyLine(scanner) if !strings.HasPrefix(firstLine, "vendor_id") || !strings.Contains(firstLine, ":") { return nil, fmt.Errorf("%w: Cannot parse line: %q", ErrFileParse, firstLine) } field := strings.SplitN(firstLine, ": ", 2) cpuinfo := []CPUInfo{} commonCPUInfo := CPUInfo{VendorID: field[1]} for scanner.Scan() { line := scanner.Text() if !strings.Contains(line, ":") { continue } field := strings.SplitN(line, ": ", 2) switch strings.TrimSpace(field[0]) { case "bogomips per cpu": v, err := strconv.ParseFloat(field[1], 64) if err != nil { return nil, err } commonCPUInfo.BogoMips = v case "features": commonCPUInfo.Flags = strings.Fields(field[1]) } if strings.HasPrefix(line, "processor") { match := cpuinfoS390XProcessorRegexp.FindStringSubmatch(line) if len(match) < 2 { return nil, fmt.Errorf("%w: %q", ErrFileParse, firstLine) } cpu := commonCPUInfo v, err := strconv.ParseUint(match[1], 0, 32) if err != nil { return nil, err } cpu.Processor = uint(v) cpuinfo = append(cpuinfo, cpu) } if strings.HasPrefix(line, "cpu number") { break } } i := 0 for scanner.Scan() { line := scanner.Text() if !strings.Contains(line, ":") { continue } field := strings.SplitN(line, ": ", 2) switch strings.TrimSpace(field[0]) { case "cpu number": i++ case "cpu MHz dynamic": clock := cpuinfoClockRegexp.FindString(strings.TrimSpace(field[1])) v, err := strconv.ParseFloat(clock, 64) if err != nil { return nil, err } cpuinfo[i].CPUMHz = v case "physical id": cpuinfo[i].PhysicalID = field[1] case "core id": cpuinfo[i].CoreID = field[1] case "cpu cores": v, err := strconv.ParseUint(field[1], 0, 32) if err != nil { return nil, err } cpuinfo[i].CPUCores = uint(v) case "siblings": v, err := strconv.ParseUint(field[1], 0, 32) if err != nil { return nil, err } cpuinfo[i].Siblings = uint(v) } } return cpuinfo, nil } func parseCPUInfoMips(info []byte) ([]CPUInfo, error) { scanner := bufio.NewScanner(bytes.NewReader(info)) // find the first "processor" line firstLine := firstNonEmptyLine(scanner) if !strings.HasPrefix(firstLine, "system type") || !strings.Contains(firstLine, ":") { return nil, fmt.Errorf("%w: %q", ErrFileParse, firstLine) } field := strings.SplitN(firstLine, ": ", 2) cpuinfo := []CPUInfo{} systemType := field[1] i := 0 for scanner.Scan() { line := scanner.Text() if !strings.Contains(line, ":") { continue } field := strings.SplitN(line, ": ", 2) switch strings.TrimSpace(field[0]) { case "processor": v, err := strconv.ParseUint(field[1], 0, 32) if err != nil { return nil, err } i = int(v) cpuinfo = append(cpuinfo, CPUInfo{}) // start of the next processor cpuinfo[i].Processor = uint(v) cpuinfo[i].VendorID = systemType case "cpu model": cpuinfo[i].ModelName = field[1] case "BogoMIPS": v, err := strconv.ParseFloat(field[1], 64) if err != nil { return nil, err } cpuinfo[i].BogoMips = v } } return cpuinfo, nil } func parseCPUInfoLoong(info []byte) ([]CPUInfo, error) { scanner := bufio.NewScanner(bytes.NewReader(info)) // find the first "processor" line firstLine := firstNonEmptyLine(scanner) if !strings.HasPrefix(firstLine, "system type") || !strings.Contains(firstLine, ":") { return nil, fmt.Errorf("%w: %q", ErrFileParse, firstLine) } field := strings.SplitN(firstLine, ": ", 2) cpuinfo := []CPUInfo{} systemType := field[1] i := 0 for scanner.Scan() { line := scanner.Text() if !strings.Contains(line, ":") { continue } field := strings.SplitN(line, ": ", 2) switch strings.TrimSpace(field[0]) { case "processor": v, err := strconv.ParseUint(field[1], 0, 32) if err != nil { return nil, err } i = int(v) cpuinfo = append(cpuinfo, CPUInfo{}) // start of the next processor cpuinfo[i].Processor = uint(v) cpuinfo[i].VendorID = systemType case "CPU Family": cpuinfo[i].CPUFamily = field[1] case "Model Name": cpuinfo[i].ModelName = field[1] } } return cpuinfo, nil } func parseCPUInfoPPC(info []byte) ([]CPUInfo, error) { scanner := bufio.NewScanner(bytes.NewReader(info)) firstLine := firstNonEmptyLine(scanner) if !strings.HasPrefix(firstLine, "processor") || !strings.Contains(firstLine, ":") { return nil, fmt.Errorf("%w: %q", ErrFileParse, firstLine) } field := strings.SplitN(firstLine, ": ", 2) v, err := strconv.ParseUint(field[1], 0, 32) if err != nil { return nil, err } firstcpu := CPUInfo{Processor: uint(v)} cpuinfo := []CPUInfo{firstcpu} i := 0 for scanner.Scan() { line := scanner.Text() if !strings.Contains(line, ":") { continue } field := strings.SplitN(line, ": ", 2) switch strings.TrimSpace(field[0]) { case "processor": cpuinfo = append(cpuinfo, CPUInfo{}) // start of the next processor i++ v, err := strconv.ParseUint(field[1], 0, 32) if err != nil { return nil, err } cpuinfo[i].Processor = uint(v) case "cpu": cpuinfo[i].VendorID = field[1] case "clock": clock := cpuinfoClockRegexp.FindString(strings.TrimSpace(field[1])) v, err := strconv.ParseFloat(clock, 64) if err != nil { return nil, err } cpuinfo[i].CPUMHz = v } } return cpuinfo, nil } func parseCPUInfoRISCV(info []byte) ([]CPUInfo, error) { scanner := bufio.NewScanner(bytes.NewReader(info)) firstLine := firstNonEmptyLine(scanner) if !strings.HasPrefix(firstLine, "processor") || !strings.Contains(firstLine, ":") { return nil, fmt.Errorf("%w: %q", ErrFileParse, firstLine) } field := strings.SplitN(firstLine, ": ", 2) v, err := strconv.ParseUint(field[1], 0, 32) if err != nil { return nil, err } firstcpu := CPUInfo{Processor: uint(v)} cpuinfo := []CPUInfo{firstcpu} i := 0 for scanner.Scan() { line := scanner.Text() if !strings.Contains(line, ":") { continue } field := strings.SplitN(line, ": ", 2) switch strings.TrimSpace(field[0]) { case "processor": v, err := strconv.ParseUint(field[1], 0, 32) if err != nil { return nil, err } i = int(v) cpuinfo = append(cpuinfo, CPUInfo{}) // start of the next processor cpuinfo[i].Processor = uint(v) case "hart": cpuinfo[i].CoreID = field[1] case "isa": cpuinfo[i].ModelName = field[1] } } return cpuinfo, nil } func parseCPUInfoDummy(_ []byte) ([]CPUInfo, error) { // nolint:unused,deadcode return nil, errors.New("not implemented") } // firstNonEmptyLine advances the scanner to the first non-empty line // and returns the contents of that line. func firstNonEmptyLine(scanner *bufio.Scanner) string { for scanner.Scan() { line := scanner.Text() if strings.TrimSpace(line) != "" { return line } } return "" }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/proc_status.go
vendor/github.com/prometheus/procfs/proc_status.go
// Copyright 2018 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package procfs import ( "bytes" "math/bits" "sort" "strconv" "strings" "github.com/prometheus/procfs/internal/util" ) // ProcStatus provides status information about the process, // read from /proc/[pid]/status. type ProcStatus struct { // The process ID. PID int // The process name. Name string // Thread group ID. TGID int // List of Pid namespace. NSpids []uint64 // Peak virtual memory size. VmPeak uint64 // nolint:revive // Virtual memory size. VmSize uint64 // nolint:revive // Locked memory size. VmLck uint64 // nolint:revive // Pinned memory size. VmPin uint64 // nolint:revive // Peak resident set size. VmHWM uint64 // nolint:revive // Resident set size (sum of RssAnnon RssFile and RssShmem). VmRSS uint64 // nolint:revive // Size of resident anonymous memory. RssAnon uint64 // nolint:revive // Size of resident file mappings. RssFile uint64 // nolint:revive // Size of resident shared memory. RssShmem uint64 // nolint:revive // Size of data segments. VmData uint64 // nolint:revive // Size of stack segments. VmStk uint64 // nolint:revive // Size of text segments. VmExe uint64 // nolint:revive // Shared library code size. VmLib uint64 // nolint:revive // Page table entries size. VmPTE uint64 // nolint:revive // Size of second-level page tables. VmPMD uint64 // nolint:revive // Swapped-out virtual memory size by anonymous private. VmSwap uint64 // nolint:revive // Size of hugetlb memory portions HugetlbPages uint64 // Number of voluntary context switches. VoluntaryCtxtSwitches uint64 // Number of involuntary context switches. NonVoluntaryCtxtSwitches uint64 // UIDs of the process (Real, effective, saved set, and filesystem UIDs) UIDs [4]uint64 // GIDs of the process (Real, effective, saved set, and filesystem GIDs) GIDs [4]uint64 // CpusAllowedList: List of cpu cores processes are allowed to run on. CpusAllowedList []uint64 } // NewStatus returns the current status information of the process. func (p Proc) NewStatus() (ProcStatus, error) { data, err := util.ReadFileNoStat(p.path("status")) if err != nil { return ProcStatus{}, err } s := ProcStatus{PID: p.PID} lines := strings.Split(string(data), "\n") for _, line := range lines { if !bytes.Contains([]byte(line), []byte(":")) { continue } kv := strings.SplitN(line, ":", 2) // removes spaces k := strings.TrimSpace(kv[0]) v := strings.TrimSpace(kv[1]) // removes "kB" v = strings.TrimSuffix(v, " kB") // value to int when possible // we can skip error check here, 'cause vKBytes is not used when value is a string vKBytes, _ := strconv.ParseUint(v, 10, 64) // convert kB to B vBytes := vKBytes * 1024 err = s.fillStatus(k, v, vKBytes, vBytes) if err != nil { return ProcStatus{}, err } } return s, nil } func (s *ProcStatus) fillStatus(k string, vString string, vUint uint64, vUintBytes uint64) error { switch k { case "Tgid": s.TGID = int(vUint) case "Name": s.Name = vString case "Uid": var err error for i, v := range strings.Split(vString, "\t") { s.UIDs[i], err = strconv.ParseUint(v, 10, bits.UintSize) if err != nil { return err } } case "Gid": var err error for i, v := range strings.Split(vString, "\t") { s.GIDs[i], err = strconv.ParseUint(v, 10, bits.UintSize) if err != nil { return err } } case "NSpid": s.NSpids = calcNSPidsList(vString) case "VmPeak": s.VmPeak = vUintBytes case "VmSize": s.VmSize = vUintBytes case "VmLck": s.VmLck = vUintBytes case "VmPin": s.VmPin = vUintBytes case "VmHWM": s.VmHWM = vUintBytes case "VmRSS": s.VmRSS = vUintBytes case "RssAnon": s.RssAnon = vUintBytes case "RssFile": s.RssFile = vUintBytes case "RssShmem": s.RssShmem = vUintBytes case "VmData": s.VmData = vUintBytes case "VmStk": s.VmStk = vUintBytes case "VmExe": s.VmExe = vUintBytes case "VmLib": s.VmLib = vUintBytes case "VmPTE": s.VmPTE = vUintBytes case "VmPMD": s.VmPMD = vUintBytes case "VmSwap": s.VmSwap = vUintBytes case "HugetlbPages": s.HugetlbPages = vUintBytes case "voluntary_ctxt_switches": s.VoluntaryCtxtSwitches = vUint case "nonvoluntary_ctxt_switches": s.NonVoluntaryCtxtSwitches = vUint case "Cpus_allowed_list": s.CpusAllowedList = calcCpusAllowedList(vString) } return nil } // TotalCtxtSwitches returns the total context switch. func (s ProcStatus) TotalCtxtSwitches() uint64 { return s.VoluntaryCtxtSwitches + s.NonVoluntaryCtxtSwitches } func calcCpusAllowedList(cpuString string) []uint64 { s := strings.Split(cpuString, ",") var g []uint64 for _, cpu := range s { // parse cpu ranges, example: 1-3=[1,2,3] if l := strings.Split(strings.TrimSpace(cpu), "-"); len(l) > 1 { startCPU, _ := strconv.ParseUint(l[0], 10, 64) endCPU, _ := strconv.ParseUint(l[1], 10, 64) for i := startCPU; i <= endCPU; i++ { g = append(g, i) } } else if len(l) == 1 { cpu, _ := strconv.ParseUint(l[0], 10, 64) g = append(g, cpu) } } sort.Slice(g, func(i, j int) bool { return g[i] < g[j] }) return g } func calcNSPidsList(nspidsString string) []uint64 { s := strings.Split(nspidsString, " ") var nspids []uint64 for _, nspid := range s { nspid, _ := strconv.ParseUint(nspid, 10, 64) if nspid == 0 { continue } nspids = append(nspids, nspid) } return nspids }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/proc_smaps.go
vendor/github.com/prometheus/procfs/proc_smaps.go
// Copyright 2020 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. //go:build !windows // +build !windows package procfs import ( "bufio" "errors" "fmt" "os" "regexp" "strconv" "strings" "github.com/prometheus/procfs/internal/util" ) var ( // match the header line before each mapped zone in `/proc/pid/smaps`. procSMapsHeaderLine = regexp.MustCompile(`^[a-f0-9].*$`) ) type ProcSMapsRollup struct { // Amount of the mapping that is currently resident in RAM. Rss uint64 // Process's proportional share of this mapping. Pss uint64 // Size in bytes of clean shared pages. SharedClean uint64 // Size in bytes of dirty shared pages. SharedDirty uint64 // Size in bytes of clean private pages. PrivateClean uint64 // Size in bytes of dirty private pages. PrivateDirty uint64 // Amount of memory currently marked as referenced or accessed. Referenced uint64 // Amount of memory that does not belong to any file. Anonymous uint64 // Amount would-be-anonymous memory currently on swap. Swap uint64 // Process's proportional memory on swap. SwapPss uint64 } // ProcSMapsRollup reads from /proc/[pid]/smaps_rollup to get summed memory information of the // process. // // If smaps_rollup does not exists (require kernel >= 4.15), the content of /proc/pid/smaps will // we read and summed. func (p Proc) ProcSMapsRollup() (ProcSMapsRollup, error) { data, err := util.ReadFileNoStat(p.path("smaps_rollup")) if err != nil && os.IsNotExist(err) { return p.procSMapsRollupManual() } if err != nil { return ProcSMapsRollup{}, err } lines := strings.Split(string(data), "\n") smaps := ProcSMapsRollup{} // skip first line which don't contains information we need lines = lines[1:] for _, line := range lines { if line == "" { continue } if err := smaps.parseLine(line); err != nil { return ProcSMapsRollup{}, err } } return smaps, nil } // Read /proc/pid/smaps and do the roll-up in Go code. func (p Proc) procSMapsRollupManual() (ProcSMapsRollup, error) { file, err := os.Open(p.path("smaps")) if err != nil { return ProcSMapsRollup{}, err } defer file.Close() smaps := ProcSMapsRollup{} scan := bufio.NewScanner(file) for scan.Scan() { line := scan.Text() if procSMapsHeaderLine.MatchString(line) { continue } if err := smaps.parseLine(line); err != nil { return ProcSMapsRollup{}, err } } return smaps, nil } func (s *ProcSMapsRollup) parseLine(line string) error { kv := strings.SplitN(line, ":", 2) if len(kv) != 2 { fmt.Println(line) return errors.New("invalid net/dev line, missing colon") } k := kv[0] if k == "VmFlags" { return nil } v := strings.TrimSpace(kv[1]) v = strings.TrimSuffix(v, " kB") vKBytes, err := strconv.ParseUint(v, 10, 64) if err != nil { return err } vBytes := vKBytes * 1024 s.addValue(k, vBytes) return nil } func (s *ProcSMapsRollup) addValue(k string, vUintBytes uint64) { switch k { case "Rss": s.Rss += vUintBytes case "Pss": s.Pss += vUintBytes case "Shared_Clean": s.SharedClean += vUintBytes case "Shared_Dirty": s.SharedDirty += vUintBytes case "Private_Clean": s.PrivateClean += vUintBytes case "Private_Dirty": s.PrivateDirty += vUintBytes case "Referenced": s.Referenced += vUintBytes case "Anonymous": s.Anonymous += vUintBytes case "Swap": s.Swap += vUintBytes case "SwapPss": s.SwapPss += vUintBytes } }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/proc_limits.go
vendor/github.com/prometheus/procfs/proc_limits.go
// Copyright 2018 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package procfs import ( "bufio" "fmt" "os" "regexp" "strconv" ) // ProcLimits represents the soft limits for each of the process's resource // limits. For more information see getrlimit(2): // http://man7.org/linux/man-pages/man2/getrlimit.2.html. type ProcLimits struct { // CPU time limit in seconds. CPUTime uint64 // Maximum size of files that the process may create. FileSize uint64 // Maximum size of the process's data segment (initialized data, // uninitialized data, and heap). DataSize uint64 // Maximum size of the process stack in bytes. StackSize uint64 // Maximum size of a core file. CoreFileSize uint64 // Limit of the process's resident set in pages. ResidentSet uint64 // Maximum number of processes that can be created for the real user ID of // the calling process. Processes uint64 // Value one greater than the maximum file descriptor number that can be // opened by this process. OpenFiles uint64 // Maximum number of bytes of memory that may be locked into RAM. LockedMemory uint64 // Maximum size of the process's virtual memory address space in bytes. AddressSpace uint64 // Limit on the combined number of flock(2) locks and fcntl(2) leases that // this process may establish. FileLocks uint64 // Limit of signals that may be queued for the real user ID of the calling // process. PendingSignals uint64 // Limit on the number of bytes that can be allocated for POSIX message // queues for the real user ID of the calling process. MsqqueueSize uint64 // Limit of the nice priority set using setpriority(2) or nice(2). NicePriority uint64 // Limit of the real-time priority set using sched_setscheduler(2) or // sched_setparam(2). RealtimePriority uint64 // Limit (in microseconds) on the amount of CPU time that a process // scheduled under a real-time scheduling policy may consume without making // a blocking system call. RealtimeTimeout uint64 } const ( limitsFields = 4 limitsUnlimited = "unlimited" ) var ( limitsMatch = regexp.MustCompile(`(Max \w+\s{0,1}?\w*\s{0,1}\w*)\s{2,}(\w+)\s+(\w+)`) ) // NewLimits returns the current soft limits of the process. // // Deprecated: Use p.Limits() instead. func (p Proc) NewLimits() (ProcLimits, error) { return p.Limits() } // Limits returns the current soft limits of the process. func (p Proc) Limits() (ProcLimits, error) { f, err := os.Open(p.path("limits")) if err != nil { return ProcLimits{}, err } defer f.Close() var ( l = ProcLimits{} s = bufio.NewScanner(f) ) s.Scan() // Skip limits header for s.Scan() { //fields := limitsMatch.Split(s.Text(), limitsFields) fields := limitsMatch.FindStringSubmatch(s.Text()) if len(fields) != limitsFields { return ProcLimits{}, fmt.Errorf("%w: couldn't parse %q line %q", ErrFileParse, f.Name(), s.Text()) } switch fields[1] { case "Max cpu time": l.CPUTime, err = parseUint(fields[2]) case "Max file size": l.FileSize, err = parseUint(fields[2]) case "Max data size": l.DataSize, err = parseUint(fields[2]) case "Max stack size": l.StackSize, err = parseUint(fields[2]) case "Max core file size": l.CoreFileSize, err = parseUint(fields[2]) case "Max resident set": l.ResidentSet, err = parseUint(fields[2]) case "Max processes": l.Processes, err = parseUint(fields[2]) case "Max open files": l.OpenFiles, err = parseUint(fields[2]) case "Max locked memory": l.LockedMemory, err = parseUint(fields[2]) case "Max address space": l.AddressSpace, err = parseUint(fields[2]) case "Max file locks": l.FileLocks, err = parseUint(fields[2]) case "Max pending signals": l.PendingSignals, err = parseUint(fields[2]) case "Max msgqueue size": l.MsqqueueSize, err = parseUint(fields[2]) case "Max nice priority": l.NicePriority, err = parseUint(fields[2]) case "Max realtime priority": l.RealtimePriority, err = parseUint(fields[2]) case "Max realtime timeout": l.RealtimeTimeout, err = parseUint(fields[2]) } if err != nil { return ProcLimits{}, err } } return l, s.Err() } func parseUint(s string) (uint64, error) { if s == limitsUnlimited { return 18446744073709551615, nil } i, err := strconv.ParseUint(s, 10, 64) if err != nil { return 0, fmt.Errorf("%w: couldn't parse value %q: %w", ErrFileParse, s, err) } return i, nil }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/net_tls_stat.go
vendor/github.com/prometheus/procfs/net_tls_stat.go
// Copyright 2023 Prometheus Team // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package procfs import ( "bufio" "fmt" "os" "strconv" "strings" ) // TLSStat struct represents data in /proc/net/tls_stat. // See https://docs.kernel.org/networking/tls.html#statistics type TLSStat struct { // number of TX sessions currently installed where host handles cryptography TLSCurrTxSw int // number of RX sessions currently installed where host handles cryptography TLSCurrRxSw int // number of TX sessions currently installed where NIC handles cryptography TLSCurrTxDevice int // number of RX sessions currently installed where NIC handles cryptography TLSCurrRxDevice int //number of TX sessions opened with host cryptography TLSTxSw int //number of RX sessions opened with host cryptography TLSRxSw int // number of TX sessions opened with NIC cryptography TLSTxDevice int // number of RX sessions opened with NIC cryptography TLSRxDevice int // record decryption failed (e.g. due to incorrect authentication tag) TLSDecryptError int // number of RX resyncs sent to NICs handling cryptography TLSRxDeviceResync int // number of RX records which had to be re-decrypted due to TLS_RX_EXPECT_NO_PAD mis-prediction. Note that this counter will also increment for non-data records. TLSDecryptRetry int // number of data RX records which had to be re-decrypted due to TLS_RX_EXPECT_NO_PAD mis-prediction. TLSRxNoPadViolation int } // NewTLSStat reads the tls_stat statistics. func NewTLSStat() (TLSStat, error) { fs, err := NewFS(DefaultMountPoint) if err != nil { return TLSStat{}, err } return fs.NewTLSStat() } // NewTLSStat reads the tls_stat statistics. func (fs FS) NewTLSStat() (TLSStat, error) { file, err := os.Open(fs.proc.Path("net/tls_stat")) if err != nil { return TLSStat{}, err } defer file.Close() var ( tlsstat = TLSStat{} s = bufio.NewScanner(file) ) for s.Scan() { fields := strings.Fields(s.Text()) if len(fields) != 2 { return TLSStat{}, fmt.Errorf("%w: %q line %q", ErrFileParse, file.Name(), s.Text()) } name := fields[0] value, err := strconv.Atoi(fields[1]) if err != nil { return TLSStat{}, err } switch name { case "TlsCurrTxSw": tlsstat.TLSCurrTxSw = value case "TlsCurrRxSw": tlsstat.TLSCurrRxSw = value case "TlsCurrTxDevice": tlsstat.TLSCurrTxDevice = value case "TlsCurrRxDevice": tlsstat.TLSCurrRxDevice = value case "TlsTxSw": tlsstat.TLSTxSw = value case "TlsRxSw": tlsstat.TLSRxSw = value case "TlsTxDevice": tlsstat.TLSTxDevice = value case "TlsRxDevice": tlsstat.TLSRxDevice = value case "TlsDecryptError": tlsstat.TLSDecryptError = value case "TlsRxDeviceResync": tlsstat.TLSRxDeviceResync = value case "TlsDecryptRetry": tlsstat.TLSDecryptRetry = value case "TlsRxNoPadViolation": tlsstat.TLSRxNoPadViolation = value } } return tlsstat, s.Err() }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/stat.go
vendor/github.com/prometheus/procfs/stat.go
// Copyright 2018 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package procfs import ( "bufio" "bytes" "fmt" "io" "strconv" "strings" "github.com/prometheus/procfs/internal/fs" "github.com/prometheus/procfs/internal/util" ) // CPUStat shows how much time the cpu spend in various stages. type CPUStat struct { User float64 Nice float64 System float64 Idle float64 Iowait float64 IRQ float64 SoftIRQ float64 Steal float64 Guest float64 GuestNice float64 } // SoftIRQStat represent the softirq statistics as exported in the procfs stat file. // A nice introduction can be found at https://0xax.gitbooks.io/linux-insides/content/interrupts/interrupts-9.html // It is possible to get per-cpu stats by reading `/proc/softirqs`. type SoftIRQStat struct { Hi uint64 Timer uint64 NetTx uint64 NetRx uint64 Block uint64 BlockIoPoll uint64 Tasklet uint64 Sched uint64 Hrtimer uint64 Rcu uint64 } // Stat represents kernel/system statistics. type Stat struct { // Boot time in seconds since the Epoch. BootTime uint64 // Summed up cpu statistics. CPUTotal CPUStat // Per-CPU statistics. CPU map[int64]CPUStat // Number of times interrupts were handled, which contains numbered and unnumbered IRQs. IRQTotal uint64 // Number of times a numbered IRQ was triggered. IRQ []uint64 // Number of times a context switch happened. ContextSwitches uint64 // Number of times a process was created. ProcessCreated uint64 // Number of processes currently running. ProcessesRunning uint64 // Number of processes currently blocked (waiting for IO). ProcessesBlocked uint64 // Number of times a softirq was scheduled. SoftIRQTotal uint64 // Detailed softirq statistics. SoftIRQ SoftIRQStat } // Parse a cpu statistics line and returns the CPUStat struct plus the cpu id (or -1 for the overall sum). func parseCPUStat(line string) (CPUStat, int64, error) { cpuStat := CPUStat{} var cpu string count, err := fmt.Sscanf(line, "%s %f %f %f %f %f %f %f %f %f %f", &cpu, &cpuStat.User, &cpuStat.Nice, &cpuStat.System, &cpuStat.Idle, &cpuStat.Iowait, &cpuStat.IRQ, &cpuStat.SoftIRQ, &cpuStat.Steal, &cpuStat.Guest, &cpuStat.GuestNice) if err != nil && err != io.EOF { return CPUStat{}, -1, fmt.Errorf("%w: couldn't parse %q (cpu): %w", ErrFileParse, line, err) } if count == 0 { return CPUStat{}, -1, fmt.Errorf("%w: couldn't parse %q (cpu): 0 elements parsed", ErrFileParse, line) } cpuStat.User /= userHZ cpuStat.Nice /= userHZ cpuStat.System /= userHZ cpuStat.Idle /= userHZ cpuStat.Iowait /= userHZ cpuStat.IRQ /= userHZ cpuStat.SoftIRQ /= userHZ cpuStat.Steal /= userHZ cpuStat.Guest /= userHZ cpuStat.GuestNice /= userHZ if cpu == "cpu" { return cpuStat, -1, nil } cpuID, err := strconv.ParseInt(cpu[3:], 10, 64) if err != nil { return CPUStat{}, -1, fmt.Errorf("%w: couldn't parse %q (cpu/cpuid): %w", ErrFileParse, line, err) } return cpuStat, cpuID, nil } // Parse a softirq line. func parseSoftIRQStat(line string) (SoftIRQStat, uint64, error) { softIRQStat := SoftIRQStat{} var total uint64 var prefix string _, err := fmt.Sscanf(line, "%s %d %d %d %d %d %d %d %d %d %d %d", &prefix, &total, &softIRQStat.Hi, &softIRQStat.Timer, &softIRQStat.NetTx, &softIRQStat.NetRx, &softIRQStat.Block, &softIRQStat.BlockIoPoll, &softIRQStat.Tasklet, &softIRQStat.Sched, &softIRQStat.Hrtimer, &softIRQStat.Rcu) if err != nil { return SoftIRQStat{}, 0, fmt.Errorf("%w: couldn't parse %q (softirq): %w", ErrFileParse, line, err) } return softIRQStat, total, nil } // NewStat returns information about current cpu/process statistics. // See https://www.kernel.org/doc/Documentation/filesystems/proc.txt // // Deprecated: Use fs.Stat() instead. func NewStat() (Stat, error) { fs, err := NewFS(fs.DefaultProcMountPoint) if err != nil { return Stat{}, err } return fs.Stat() } // NewStat returns information about current cpu/process statistics. // See: https://www.kernel.org/doc/Documentation/filesystems/proc.txt // // Deprecated: Use fs.Stat() instead. func (fs FS) NewStat() (Stat, error) { return fs.Stat() } // Stat returns information about current cpu/process statistics. // See: https://www.kernel.org/doc/Documentation/filesystems/proc.txt func (fs FS) Stat() (Stat, error) { fileName := fs.proc.Path("stat") data, err := util.ReadFileNoStat(fileName) if err != nil { return Stat{}, err } procStat, err := parseStat(bytes.NewReader(data), fileName) if err != nil { return Stat{}, err } return procStat, nil } // parseStat parses the metrics from /proc/[pid]/stat. func parseStat(r io.Reader, fileName string) (Stat, error) { var ( scanner = bufio.NewScanner(r) stat = Stat{ CPU: make(map[int64]CPUStat), } err error ) // Increase default scanner buffer to handle very long `intr` lines. buf := make([]byte, 0, 8*1024) scanner.Buffer(buf, 1024*1024) for scanner.Scan() { line := scanner.Text() parts := strings.Fields(scanner.Text()) // require at least <key> <value> if len(parts) < 2 { continue } switch { case parts[0] == "btime": if stat.BootTime, err = strconv.ParseUint(parts[1], 10, 64); err != nil { return Stat{}, fmt.Errorf("%w: couldn't parse %q (btime): %w", ErrFileParse, parts[1], err) } case parts[0] == "intr": if stat.IRQTotal, err = strconv.ParseUint(parts[1], 10, 64); err != nil { return Stat{}, fmt.Errorf("%w: couldn't parse %q (intr): %w", ErrFileParse, parts[1], err) } numberedIRQs := parts[2:] stat.IRQ = make([]uint64, len(numberedIRQs)) for i, count := range numberedIRQs { if stat.IRQ[i], err = strconv.ParseUint(count, 10, 64); err != nil { return Stat{}, fmt.Errorf("%w: couldn't parse %q (intr%d): %w", ErrFileParse, count, i, err) } } case parts[0] == "ctxt": if stat.ContextSwitches, err = strconv.ParseUint(parts[1], 10, 64); err != nil { return Stat{}, fmt.Errorf("%w: couldn't parse %q (ctxt): %w", ErrFileParse, parts[1], err) } case parts[0] == "processes": if stat.ProcessCreated, err = strconv.ParseUint(parts[1], 10, 64); err != nil { return Stat{}, fmt.Errorf("%w: couldn't parse %q (processes): %w", ErrFileParse, parts[1], err) } case parts[0] == "procs_running": if stat.ProcessesRunning, err = strconv.ParseUint(parts[1], 10, 64); err != nil { return Stat{}, fmt.Errorf("%w: couldn't parse %q (procs_running): %w", ErrFileParse, parts[1], err) } case parts[0] == "procs_blocked": if stat.ProcessesBlocked, err = strconv.ParseUint(parts[1], 10, 64); err != nil { return Stat{}, fmt.Errorf("%w: couldn't parse %q (procs_blocked): %w", ErrFileParse, parts[1], err) } case parts[0] == "softirq": softIRQStats, total, err := parseSoftIRQStat(line) if err != nil { return Stat{}, err } stat.SoftIRQTotal = total stat.SoftIRQ = softIRQStats case strings.HasPrefix(parts[0], "cpu"): cpuStat, cpuID, err := parseCPUStat(line) if err != nil { return Stat{}, err } if cpuID == -1 { stat.CPUTotal = cpuStat } else { stat.CPU[cpuID] = cpuStat } } } if err := scanner.Err(); err != nil { return Stat{}, fmt.Errorf("%w: couldn't parse %q: %w", ErrFileParse, fileName, err) } return stat, nil }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/mountinfo.go
vendor/github.com/prometheus/procfs/mountinfo.go
// Copyright 2019 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package procfs import ( "bufio" "bytes" "fmt" "strconv" "strings" "github.com/prometheus/procfs/internal/util" ) // A MountInfo is a type that describes the details, options // for each mount, parsed from /proc/self/mountinfo. // The fields described in each entry of /proc/self/mountinfo // is described in the following man page. // http://man7.org/linux/man-pages/man5/proc.5.html type MountInfo struct { // Unique ID for the mount MountID int // The ID of the parent mount ParentID int // The value of `st_dev` for the files on this FS MajorMinorVer string // The pathname of the directory in the FS that forms // the root for this mount Root string // The pathname of the mount point relative to the root MountPoint string // Mount options Options map[string]string // Zero or more optional fields OptionalFields map[string]string // The Filesystem type FSType string // FS specific information or "none" Source string // Superblock options SuperOptions map[string]string } // Reads each line of the mountinfo file, and returns a list of formatted MountInfo structs. func parseMountInfo(info []byte) ([]*MountInfo, error) { mounts := []*MountInfo{} scanner := bufio.NewScanner(bytes.NewReader(info)) for scanner.Scan() { mountString := scanner.Text() parsedMounts, err := parseMountInfoString(mountString) if err != nil { return nil, err } mounts = append(mounts, parsedMounts) } err := scanner.Err() return mounts, err } // Parses a mountinfo file line, and converts it to a MountInfo struct. // An important check here is to see if the hyphen separator, as if it does not exist, // it means that the line is malformed. func parseMountInfoString(mountString string) (*MountInfo, error) { var err error mountInfo := strings.Split(mountString, " ") mountInfoLength := len(mountInfo) if mountInfoLength < 10 { return nil, fmt.Errorf("%w: Too few fields in mount string: %s", ErrFileParse, mountString) } if mountInfo[mountInfoLength-4] != "-" { return nil, fmt.Errorf("%w: couldn't find separator in expected field: %s", ErrFileParse, mountInfo[mountInfoLength-4]) } mount := &MountInfo{ MajorMinorVer: mountInfo[2], Root: mountInfo[3], MountPoint: mountInfo[4], Options: mountOptionsParser(mountInfo[5]), OptionalFields: nil, FSType: mountInfo[mountInfoLength-3], Source: mountInfo[mountInfoLength-2], SuperOptions: mountOptionsParser(mountInfo[mountInfoLength-1]), } mount.MountID, err = strconv.Atoi(mountInfo[0]) if err != nil { return nil, fmt.Errorf("%w: mount ID: %q", ErrFileParse, mount.MountID) } mount.ParentID, err = strconv.Atoi(mountInfo[1]) if err != nil { return nil, fmt.Errorf("%w: parent ID: %q", ErrFileParse, mount.ParentID) } // Has optional fields, which is a space separated list of values. // Example: shared:2 master:7 if mountInfo[6] != "" { mount.OptionalFields, err = mountOptionsParseOptionalFields(mountInfo[6 : mountInfoLength-4]) if err != nil { return nil, fmt.Errorf("%w: %w", ErrFileParse, err) } } return mount, nil } // mountOptionsIsValidField checks a string against a valid list of optional fields keys. func mountOptionsIsValidField(s string) bool { switch s { case "shared", "master", "propagate_from", "unbindable": return true } return false } // mountOptionsParseOptionalFields parses a list of optional fields strings into a double map of strings. func mountOptionsParseOptionalFields(o []string) (map[string]string, error) { optionalFields := make(map[string]string) for _, field := range o { optionSplit := strings.SplitN(field, ":", 2) value := "" if len(optionSplit) == 2 { value = optionSplit[1] } if mountOptionsIsValidField(optionSplit[0]) { optionalFields[optionSplit[0]] = value } } return optionalFields, nil } // mountOptionsParser parses the mount options, superblock options. func mountOptionsParser(mountOptions string) map[string]string { opts := make(map[string]string) options := strings.Split(mountOptions, ",") for _, opt := range options { splitOption := strings.Split(opt, "=") if len(splitOption) < 2 { key := splitOption[0] opts[key] = "" } else { key, value := splitOption[0], splitOption[1] opts[key] = value } } return opts } // GetMounts retrieves mountinfo information from `/proc/self/mountinfo`. func GetMounts() ([]*MountInfo, error) { data, err := util.ReadFileNoStat("/proc/self/mountinfo") if err != nil { return nil, err } return parseMountInfo(data) } // GetProcMounts retrieves mountinfo information from a processes' `/proc/<pid>/mountinfo`. func GetProcMounts(pid int) ([]*MountInfo, error) { data, err := util.ReadFileNoStat(fmt.Sprintf("/proc/%d/mountinfo", pid)) if err != nil { return nil, err } return parseMountInfo(data) }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/proc_netstat.go
vendor/github.com/prometheus/procfs/proc_netstat.go
// Copyright 2022 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package procfs import ( "bufio" "bytes" "fmt" "io" "strconv" "strings" "github.com/prometheus/procfs/internal/util" ) // ProcNetstat models the content of /proc/<pid>/net/netstat. type ProcNetstat struct { // The process ID. PID int TcpExt IpExt } type TcpExt struct { // nolint:revive SyncookiesSent *float64 SyncookiesRecv *float64 SyncookiesFailed *float64 EmbryonicRsts *float64 PruneCalled *float64 RcvPruned *float64 OfoPruned *float64 OutOfWindowIcmps *float64 LockDroppedIcmps *float64 ArpFilter *float64 TW *float64 TWRecycled *float64 TWKilled *float64 PAWSActive *float64 PAWSEstab *float64 DelayedACKs *float64 DelayedACKLocked *float64 DelayedACKLost *float64 ListenOverflows *float64 ListenDrops *float64 TCPHPHits *float64 TCPPureAcks *float64 TCPHPAcks *float64 TCPRenoRecovery *float64 TCPSackRecovery *float64 TCPSACKReneging *float64 TCPSACKReorder *float64 TCPRenoReorder *float64 TCPTSReorder *float64 TCPFullUndo *float64 TCPPartialUndo *float64 TCPDSACKUndo *float64 TCPLossUndo *float64 TCPLostRetransmit *float64 TCPRenoFailures *float64 TCPSackFailures *float64 TCPLossFailures *float64 TCPFastRetrans *float64 TCPSlowStartRetrans *float64 TCPTimeouts *float64 TCPLossProbes *float64 TCPLossProbeRecovery *float64 TCPRenoRecoveryFail *float64 TCPSackRecoveryFail *float64 TCPRcvCollapsed *float64 TCPDSACKOldSent *float64 TCPDSACKOfoSent *float64 TCPDSACKRecv *float64 TCPDSACKOfoRecv *float64 TCPAbortOnData *float64 TCPAbortOnClose *float64 TCPAbortOnMemory *float64 TCPAbortOnTimeout *float64 TCPAbortOnLinger *float64 TCPAbortFailed *float64 TCPMemoryPressures *float64 TCPMemoryPressuresChrono *float64 TCPSACKDiscard *float64 TCPDSACKIgnoredOld *float64 TCPDSACKIgnoredNoUndo *float64 TCPSpuriousRTOs *float64 TCPMD5NotFound *float64 TCPMD5Unexpected *float64 TCPMD5Failure *float64 TCPSackShifted *float64 TCPSackMerged *float64 TCPSackShiftFallback *float64 TCPBacklogDrop *float64 PFMemallocDrop *float64 TCPMinTTLDrop *float64 TCPDeferAcceptDrop *float64 IPReversePathFilter *float64 TCPTimeWaitOverflow *float64 TCPReqQFullDoCookies *float64 TCPReqQFullDrop *float64 TCPRetransFail *float64 TCPRcvCoalesce *float64 TCPRcvQDrop *float64 TCPOFOQueue *float64 TCPOFODrop *float64 TCPOFOMerge *float64 TCPChallengeACK *float64 TCPSYNChallenge *float64 TCPFastOpenActive *float64 TCPFastOpenActiveFail *float64 TCPFastOpenPassive *float64 TCPFastOpenPassiveFail *float64 TCPFastOpenListenOverflow *float64 TCPFastOpenCookieReqd *float64 TCPFastOpenBlackhole *float64 TCPSpuriousRtxHostQueues *float64 BusyPollRxPackets *float64 TCPAutoCorking *float64 TCPFromZeroWindowAdv *float64 TCPToZeroWindowAdv *float64 TCPWantZeroWindowAdv *float64 TCPSynRetrans *float64 TCPOrigDataSent *float64 TCPHystartTrainDetect *float64 TCPHystartTrainCwnd *float64 TCPHystartDelayDetect *float64 TCPHystartDelayCwnd *float64 TCPACKSkippedSynRecv *float64 TCPACKSkippedPAWS *float64 TCPACKSkippedSeq *float64 TCPACKSkippedFinWait2 *float64 TCPACKSkippedTimeWait *float64 TCPACKSkippedChallenge *float64 TCPWinProbe *float64 TCPKeepAlive *float64 TCPMTUPFail *float64 TCPMTUPSuccess *float64 TCPWqueueTooBig *float64 } type IpExt struct { // nolint:revive InNoRoutes *float64 InTruncatedPkts *float64 InMcastPkts *float64 OutMcastPkts *float64 InBcastPkts *float64 OutBcastPkts *float64 InOctets *float64 OutOctets *float64 InMcastOctets *float64 OutMcastOctets *float64 InBcastOctets *float64 OutBcastOctets *float64 InCsumErrors *float64 InNoECTPkts *float64 InECT1Pkts *float64 InECT0Pkts *float64 InCEPkts *float64 ReasmOverlaps *float64 } func (p Proc) Netstat() (ProcNetstat, error) { filename := p.path("net/netstat") data, err := util.ReadFileNoStat(filename) if err != nil { return ProcNetstat{PID: p.PID}, err } procNetstat, err := parseProcNetstat(bytes.NewReader(data), filename) procNetstat.PID = p.PID return procNetstat, err } // parseProcNetstat parses the metrics from proc/<pid>/net/netstat file // and returns a ProcNetstat structure. func parseProcNetstat(r io.Reader, fileName string) (ProcNetstat, error) { var ( scanner = bufio.NewScanner(r) procNetstat = ProcNetstat{} ) for scanner.Scan() { nameParts := strings.Split(scanner.Text(), " ") scanner.Scan() valueParts := strings.Split(scanner.Text(), " ") // Remove trailing :. protocol := strings.TrimSuffix(nameParts[0], ":") if len(nameParts) != len(valueParts) { return procNetstat, fmt.Errorf("%w: mismatch field count mismatch in %s: %s", ErrFileParse, fileName, protocol) } for i := 1; i < len(nameParts); i++ { value, err := strconv.ParseFloat(valueParts[i], 64) if err != nil { return procNetstat, err } key := nameParts[i] switch protocol { case "TcpExt": switch key { case "SyncookiesSent": procNetstat.TcpExt.SyncookiesSent = &value case "SyncookiesRecv": procNetstat.TcpExt.SyncookiesRecv = &value case "SyncookiesFailed": procNetstat.TcpExt.SyncookiesFailed = &value case "EmbryonicRsts": procNetstat.TcpExt.EmbryonicRsts = &value case "PruneCalled": procNetstat.TcpExt.PruneCalled = &value case "RcvPruned": procNetstat.TcpExt.RcvPruned = &value case "OfoPruned": procNetstat.TcpExt.OfoPruned = &value case "OutOfWindowIcmps": procNetstat.TcpExt.OutOfWindowIcmps = &value case "LockDroppedIcmps": procNetstat.TcpExt.LockDroppedIcmps = &value case "ArpFilter": procNetstat.TcpExt.ArpFilter = &value case "TW": procNetstat.TcpExt.TW = &value case "TWRecycled": procNetstat.TcpExt.TWRecycled = &value case "TWKilled": procNetstat.TcpExt.TWKilled = &value case "PAWSActive": procNetstat.TcpExt.PAWSActive = &value case "PAWSEstab": procNetstat.TcpExt.PAWSEstab = &value case "DelayedACKs": procNetstat.TcpExt.DelayedACKs = &value case "DelayedACKLocked": procNetstat.TcpExt.DelayedACKLocked = &value case "DelayedACKLost": procNetstat.TcpExt.DelayedACKLost = &value case "ListenOverflows": procNetstat.TcpExt.ListenOverflows = &value case "ListenDrops": procNetstat.TcpExt.ListenDrops = &value case "TCPHPHits": procNetstat.TcpExt.TCPHPHits = &value case "TCPPureAcks": procNetstat.TcpExt.TCPPureAcks = &value case "TCPHPAcks": procNetstat.TcpExt.TCPHPAcks = &value case "TCPRenoRecovery": procNetstat.TcpExt.TCPRenoRecovery = &value case "TCPSackRecovery": procNetstat.TcpExt.TCPSackRecovery = &value case "TCPSACKReneging": procNetstat.TcpExt.TCPSACKReneging = &value case "TCPSACKReorder": procNetstat.TcpExt.TCPSACKReorder = &value case "TCPRenoReorder": procNetstat.TcpExt.TCPRenoReorder = &value case "TCPTSReorder": procNetstat.TcpExt.TCPTSReorder = &value case "TCPFullUndo": procNetstat.TcpExt.TCPFullUndo = &value case "TCPPartialUndo": procNetstat.TcpExt.TCPPartialUndo = &value case "TCPDSACKUndo": procNetstat.TcpExt.TCPDSACKUndo = &value case "TCPLossUndo": procNetstat.TcpExt.TCPLossUndo = &value case "TCPLostRetransmit": procNetstat.TcpExt.TCPLostRetransmit = &value case "TCPRenoFailures": procNetstat.TcpExt.TCPRenoFailures = &value case "TCPSackFailures": procNetstat.TcpExt.TCPSackFailures = &value case "TCPLossFailures": procNetstat.TcpExt.TCPLossFailures = &value case "TCPFastRetrans": procNetstat.TcpExt.TCPFastRetrans = &value case "TCPSlowStartRetrans": procNetstat.TcpExt.TCPSlowStartRetrans = &value case "TCPTimeouts": procNetstat.TcpExt.TCPTimeouts = &value case "TCPLossProbes": procNetstat.TcpExt.TCPLossProbes = &value case "TCPLossProbeRecovery": procNetstat.TcpExt.TCPLossProbeRecovery = &value case "TCPRenoRecoveryFail": procNetstat.TcpExt.TCPRenoRecoveryFail = &value case "TCPSackRecoveryFail": procNetstat.TcpExt.TCPSackRecoveryFail = &value case "TCPRcvCollapsed": procNetstat.TcpExt.TCPRcvCollapsed = &value case "TCPDSACKOldSent": procNetstat.TcpExt.TCPDSACKOldSent = &value case "TCPDSACKOfoSent": procNetstat.TcpExt.TCPDSACKOfoSent = &value case "TCPDSACKRecv": procNetstat.TcpExt.TCPDSACKRecv = &value case "TCPDSACKOfoRecv": procNetstat.TcpExt.TCPDSACKOfoRecv = &value case "TCPAbortOnData": procNetstat.TcpExt.TCPAbortOnData = &value case "TCPAbortOnClose": procNetstat.TcpExt.TCPAbortOnClose = &value case "TCPDeferAcceptDrop": procNetstat.TcpExt.TCPDeferAcceptDrop = &value case "IPReversePathFilter": procNetstat.TcpExt.IPReversePathFilter = &value case "TCPTimeWaitOverflow": procNetstat.TcpExt.TCPTimeWaitOverflow = &value case "TCPReqQFullDoCookies": procNetstat.TcpExt.TCPReqQFullDoCookies = &value case "TCPReqQFullDrop": procNetstat.TcpExt.TCPReqQFullDrop = &value case "TCPRetransFail": procNetstat.TcpExt.TCPRetransFail = &value case "TCPRcvCoalesce": procNetstat.TcpExt.TCPRcvCoalesce = &value case "TCPRcvQDrop": procNetstat.TcpExt.TCPRcvQDrop = &value case "TCPOFOQueue": procNetstat.TcpExt.TCPOFOQueue = &value case "TCPOFODrop": procNetstat.TcpExt.TCPOFODrop = &value case "TCPOFOMerge": procNetstat.TcpExt.TCPOFOMerge = &value case "TCPChallengeACK": procNetstat.TcpExt.TCPChallengeACK = &value case "TCPSYNChallenge": procNetstat.TcpExt.TCPSYNChallenge = &value case "TCPFastOpenActive": procNetstat.TcpExt.TCPFastOpenActive = &value case "TCPFastOpenActiveFail": procNetstat.TcpExt.TCPFastOpenActiveFail = &value case "TCPFastOpenPassive": procNetstat.TcpExt.TCPFastOpenPassive = &value case "TCPFastOpenPassiveFail": procNetstat.TcpExt.TCPFastOpenPassiveFail = &value case "TCPFastOpenListenOverflow": procNetstat.TcpExt.TCPFastOpenListenOverflow = &value case "TCPFastOpenCookieReqd": procNetstat.TcpExt.TCPFastOpenCookieReqd = &value case "TCPFastOpenBlackhole": procNetstat.TcpExt.TCPFastOpenBlackhole = &value case "TCPSpuriousRtxHostQueues": procNetstat.TcpExt.TCPSpuriousRtxHostQueues = &value case "BusyPollRxPackets": procNetstat.TcpExt.BusyPollRxPackets = &value case "TCPAutoCorking": procNetstat.TcpExt.TCPAutoCorking = &value case "TCPFromZeroWindowAdv": procNetstat.TcpExt.TCPFromZeroWindowAdv = &value case "TCPToZeroWindowAdv": procNetstat.TcpExt.TCPToZeroWindowAdv = &value case "TCPWantZeroWindowAdv": procNetstat.TcpExt.TCPWantZeroWindowAdv = &value case "TCPSynRetrans": procNetstat.TcpExt.TCPSynRetrans = &value case "TCPOrigDataSent": procNetstat.TcpExt.TCPOrigDataSent = &value case "TCPHystartTrainDetect": procNetstat.TcpExt.TCPHystartTrainDetect = &value case "TCPHystartTrainCwnd": procNetstat.TcpExt.TCPHystartTrainCwnd = &value case "TCPHystartDelayDetect": procNetstat.TcpExt.TCPHystartDelayDetect = &value case "TCPHystartDelayCwnd": procNetstat.TcpExt.TCPHystartDelayCwnd = &value case "TCPACKSkippedSynRecv": procNetstat.TcpExt.TCPACKSkippedSynRecv = &value case "TCPACKSkippedPAWS": procNetstat.TcpExt.TCPACKSkippedPAWS = &value case "TCPACKSkippedSeq": procNetstat.TcpExt.TCPACKSkippedSeq = &value case "TCPACKSkippedFinWait2": procNetstat.TcpExt.TCPACKSkippedFinWait2 = &value case "TCPACKSkippedTimeWait": procNetstat.TcpExt.TCPACKSkippedTimeWait = &value case "TCPACKSkippedChallenge": procNetstat.TcpExt.TCPACKSkippedChallenge = &value case "TCPWinProbe": procNetstat.TcpExt.TCPWinProbe = &value case "TCPKeepAlive": procNetstat.TcpExt.TCPKeepAlive = &value case "TCPMTUPFail": procNetstat.TcpExt.TCPMTUPFail = &value case "TCPMTUPSuccess": procNetstat.TcpExt.TCPMTUPSuccess = &value case "TCPWqueueTooBig": procNetstat.TcpExt.TCPWqueueTooBig = &value } case "IpExt": switch key { case "InNoRoutes": procNetstat.IpExt.InNoRoutes = &value case "InTruncatedPkts": procNetstat.IpExt.InTruncatedPkts = &value case "InMcastPkts": procNetstat.IpExt.InMcastPkts = &value case "OutMcastPkts": procNetstat.IpExt.OutMcastPkts = &value case "InBcastPkts": procNetstat.IpExt.InBcastPkts = &value case "OutBcastPkts": procNetstat.IpExt.OutBcastPkts = &value case "InOctets": procNetstat.IpExt.InOctets = &value case "OutOctets": procNetstat.IpExt.OutOctets = &value case "InMcastOctets": procNetstat.IpExt.InMcastOctets = &value case "OutMcastOctets": procNetstat.IpExt.OutMcastOctets = &value case "InBcastOctets": procNetstat.IpExt.InBcastOctets = &value case "OutBcastOctets": procNetstat.IpExt.OutBcastOctets = &value case "InCsumErrors": procNetstat.IpExt.InCsumErrors = &value case "InNoECTPkts": procNetstat.IpExt.InNoECTPkts = &value case "InECT1Pkts": procNetstat.IpExt.InECT1Pkts = &value case "InECT0Pkts": procNetstat.IpExt.InECT0Pkts = &value case "InCEPkts": procNetstat.IpExt.InCEPkts = &value case "ReasmOverlaps": procNetstat.IpExt.ReasmOverlaps = &value } } } } return procNetstat, scanner.Err() }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/schedstat.go
vendor/github.com/prometheus/procfs/schedstat.go
// Copyright 2019 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package procfs import ( "bufio" "errors" "os" "regexp" "strconv" ) var ( cpuLineRE = regexp.MustCompile(`cpu(\d+) (\d+) (\d+) (\d+) (\d+) (\d+) (\d+) (\d+) (\d+) (\d+)`) procLineRE = regexp.MustCompile(`(\d+) (\d+) (\d+)`) ) // Schedstat contains scheduler statistics from /proc/schedstat // // See // https://www.kernel.org/doc/Documentation/scheduler/sched-stats.txt // for a detailed description of what these numbers mean. // // Note the current kernel documentation claims some of the time units are in // jiffies when they are actually in nanoseconds since 2.6.23 with the // introduction of CFS. A fix to the documentation is pending. See // https://lore.kernel.org/patchwork/project/lkml/list/?series=403473 type Schedstat struct { CPUs []*SchedstatCPU } // SchedstatCPU contains the values from one "cpu<N>" line. type SchedstatCPU struct { CPUNum string RunningNanoseconds uint64 WaitingNanoseconds uint64 RunTimeslices uint64 } // ProcSchedstat contains the values from `/proc/<pid>/schedstat`. type ProcSchedstat struct { RunningNanoseconds uint64 WaitingNanoseconds uint64 RunTimeslices uint64 } // Schedstat reads data from `/proc/schedstat`. func (fs FS) Schedstat() (*Schedstat, error) { file, err := os.Open(fs.proc.Path("schedstat")) if err != nil { return nil, err } defer file.Close() stats := &Schedstat{} scanner := bufio.NewScanner(file) for scanner.Scan() { match := cpuLineRE.FindStringSubmatch(scanner.Text()) if match != nil { cpu := &SchedstatCPU{} cpu.CPUNum = match[1] cpu.RunningNanoseconds, err = strconv.ParseUint(match[8], 10, 64) if err != nil { continue } cpu.WaitingNanoseconds, err = strconv.ParseUint(match[9], 10, 64) if err != nil { continue } cpu.RunTimeslices, err = strconv.ParseUint(match[10], 10, 64) if err != nil { continue } stats.CPUs = append(stats.CPUs, cpu) } } return stats, nil } func parseProcSchedstat(contents string) (ProcSchedstat, error) { var ( stats ProcSchedstat err error ) match := procLineRE.FindStringSubmatch(contents) if match != nil { stats.RunningNanoseconds, err = strconv.ParseUint(match[1], 10, 64) if err != nil { return stats, err } stats.WaitingNanoseconds, err = strconv.ParseUint(match[2], 10, 64) if err != nil { return stats, err } stats.RunTimeslices, err = strconv.ParseUint(match[3], 10, 64) return stats, err } return stats, errors.New("could not parse schedstat") }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/cpuinfo_riscvx.go
vendor/github.com/prometheus/procfs/cpuinfo_riscvx.go
// Copyright 2020 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. //go:build linux && (riscv || riscv64) // +build linux // +build riscv riscv64 package procfs var parseCPUInfo = parseCPUInfoRISCV
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/mountstats.go
vendor/github.com/prometheus/procfs/mountstats.go
// Copyright 2018 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package procfs // While implementing parsing of /proc/[pid]/mountstats, this blog was used // heavily as a reference: // https://utcc.utoronto.ca/~cks/space/blog/linux/NFSMountstatsIndex // // Special thanks to Chris Siebenmann for all of his posts explaining the // various statistics available for NFS. import ( "bufio" "fmt" "io" "strconv" "strings" "time" ) // Constants shared between multiple functions. const ( deviceEntryLen = 8 fieldBytesLen = 8 fieldEventsLen = 27 statVersion10 = "1.0" statVersion11 = "1.1" fieldTransport10TCPLen = 10 fieldTransport10UDPLen = 7 fieldTransport11TCPLen = 13 fieldTransport11UDPLen = 10 // kernel version >= 4.14 MaxLen // See: https://elixir.bootlin.com/linux/v6.4.8/source/net/sunrpc/xprtrdma/xprt_rdma.h#L393 fieldTransport11RDMAMaxLen = 28 // kernel version <= 4.2 MinLen // See: https://elixir.bootlin.com/linux/v4.2.8/source/net/sunrpc/xprtrdma/xprt_rdma.h#L331 fieldTransport11RDMAMinLen = 20 ) // A Mount is a device mount parsed from /proc/[pid]/mountstats. type Mount struct { // Name of the device. Device string // The mount point of the device. Mount string // The filesystem type used by the device. Type string // If available additional statistics related to this Mount. // Use a type assertion to determine if additional statistics are available. Stats MountStats } // A MountStats is a type which contains detailed statistics for a specific // type of Mount. type MountStats interface { mountStats() } // A MountStatsNFS is a MountStats implementation for NFSv3 and v4 mounts. type MountStatsNFS struct { // The version of statistics provided. StatVersion string // The mount options of the NFS mount. Opts map[string]string // The age of the NFS mount. Age time.Duration // Statistics related to byte counters for various operations. Bytes NFSBytesStats // Statistics related to various NFS event occurrences. Events NFSEventsStats // Statistics broken down by filesystem operation. Operations []NFSOperationStats // Statistics about the NFS RPC transport. Transport []NFSTransportStats } // mountStats implements MountStats. func (m MountStatsNFS) mountStats() {} // A NFSBytesStats contains statistics about the number of bytes read and written // by an NFS client to and from an NFS server. type NFSBytesStats struct { // Number of bytes read using the read() syscall. Read uint64 // Number of bytes written using the write() syscall. Write uint64 // Number of bytes read using the read() syscall in O_DIRECT mode. DirectRead uint64 // Number of bytes written using the write() syscall in O_DIRECT mode. DirectWrite uint64 // Number of bytes read from the NFS server, in total. ReadTotal uint64 // Number of bytes written to the NFS server, in total. WriteTotal uint64 // Number of pages read directly via mmap()'d files. ReadPages uint64 // Number of pages written directly via mmap()'d files. WritePages uint64 } // A NFSEventsStats contains statistics about NFS event occurrences. type NFSEventsStats struct { // Number of times cached inode attributes are re-validated from the server. InodeRevalidate uint64 // Number of times cached dentry nodes are re-validated from the server. DnodeRevalidate uint64 // Number of times an inode cache is cleared. DataInvalidate uint64 // Number of times cached inode attributes are invalidated. AttributeInvalidate uint64 // Number of times files or directories have been open()'d. VFSOpen uint64 // Number of times a directory lookup has occurred. VFSLookup uint64 // Number of times permissions have been checked. VFSAccess uint64 // Number of updates (and potential writes) to pages. VFSUpdatePage uint64 // Number of pages read directly via mmap()'d files. VFSReadPage uint64 // Number of times a group of pages have been read. VFSReadPages uint64 // Number of pages written directly via mmap()'d files. VFSWritePage uint64 // Number of times a group of pages have been written. VFSWritePages uint64 // Number of times directory entries have been read with getdents(). VFSGetdents uint64 // Number of times attributes have been set on inodes. VFSSetattr uint64 // Number of pending writes that have been forcefully flushed to the server. VFSFlush uint64 // Number of times fsync() has been called on directories and files. VFSFsync uint64 // Number of times locking has been attempted on a file. VFSLock uint64 // Number of times files have been closed and released. VFSFileRelease uint64 // Unknown. Possibly unused. CongestionWait uint64 // Number of times files have been truncated. Truncation uint64 // Number of times a file has been grown due to writes beyond its existing end. WriteExtension uint64 // Number of times a file was removed while still open by another process. SillyRename uint64 // Number of times the NFS server gave less data than expected while reading. ShortRead uint64 // Number of times the NFS server wrote less data than expected while writing. ShortWrite uint64 // Number of times the NFS server indicated EJUKEBOX; retrieving data from // offline storage. JukeboxDelay uint64 // Number of NFS v4.1+ pNFS reads. PNFSRead uint64 // Number of NFS v4.1+ pNFS writes. PNFSWrite uint64 } // A NFSOperationStats contains statistics for a single operation. type NFSOperationStats struct { // The name of the operation. Operation string // Number of requests performed for this operation. Requests uint64 // Number of times an actual RPC request has been transmitted for this operation. Transmissions uint64 // Number of times a request has had a major timeout. MajorTimeouts uint64 // Number of bytes sent for this operation, including RPC headers and payload. BytesSent uint64 // Number of bytes received for this operation, including RPC headers and payload. BytesReceived uint64 // Duration all requests spent queued for transmission before they were sent. CumulativeQueueMilliseconds uint64 // Duration it took to get a reply back after the request was transmitted. CumulativeTotalResponseMilliseconds uint64 // Duration from when a request was enqueued to when it was completely handled. CumulativeTotalRequestMilliseconds uint64 // The count of operations that complete with tk_status < 0. These statuses usually indicate error conditions. Errors uint64 } // A NFSTransportStats contains statistics for the NFS mount RPC requests and // responses. type NFSTransportStats struct { // The transport protocol used for the NFS mount. Protocol string // The local port used for the NFS mount. Port uint64 // Number of times the client has had to establish a connection from scratch // to the NFS server. Bind uint64 // Number of times the client has made a TCP connection to the NFS server. Connect uint64 // Duration (in jiffies, a kernel internal unit of time) the NFS mount has // spent waiting for connections to the server to be established. ConnectIdleTime uint64 // Duration since the NFS mount last saw any RPC traffic. IdleTimeSeconds uint64 // Number of RPC requests for this mount sent to the NFS server. Sends uint64 // Number of RPC responses for this mount received from the NFS server. Receives uint64 // Number of times the NFS server sent a response with a transaction ID // unknown to this client. BadTransactionIDs uint64 // A running counter, incremented on each request as the current difference // ebetween sends and receives. CumulativeActiveRequests uint64 // A running counter, incremented on each request by the current backlog // queue size. CumulativeBacklog uint64 // Stats below only available with stat version 1.1. // Maximum number of simultaneously active RPC requests ever used. MaximumRPCSlotsUsed uint64 // A running counter, incremented on each request as the current size of the // sending queue. CumulativeSendingQueue uint64 // A running counter, incremented on each request as the current size of the // pending queue. CumulativePendingQueue uint64 // Stats below only available with stat version 1.1. // Transport over RDMA // accessed when sending a call ReadChunkCount uint64 WriteChunkCount uint64 ReplyChunkCount uint64 TotalRdmaRequest uint64 // rarely accessed error counters PullupCopyCount uint64 HardwayRegisterCount uint64 FailedMarshalCount uint64 BadReplyCount uint64 MrsRecovered uint64 MrsOrphaned uint64 MrsAllocated uint64 EmptySendctxQ uint64 // accessed when receiving a reply TotalRdmaReply uint64 FixupCopyCount uint64 ReplyWaitsForSend uint64 LocalInvNeeded uint64 NomsgCallCount uint64 BcallCount uint64 } // parseMountStats parses a /proc/[pid]/mountstats file and returns a slice // of Mount structures containing detailed information about each mount. // If available, statistics for each mount are parsed as well. func parseMountStats(r io.Reader) ([]*Mount, error) { const ( device = "device" statVersionPrefix = "statvers=" nfs3Type = "nfs" nfs4Type = "nfs4" ) var mounts []*Mount s := bufio.NewScanner(r) for s.Scan() { // Only look for device entries in this function ss := strings.Fields(string(s.Bytes())) if len(ss) == 0 || ss[0] != device { continue } m, err := parseMount(ss) if err != nil { return nil, err } // Does this mount also possess statistics information? if len(ss) > deviceEntryLen { // Only NFSv3 and v4 are supported for parsing statistics if m.Type != nfs3Type && m.Type != nfs4Type { return nil, fmt.Errorf("%w: Cannot parse MountStats for %q", ErrFileParse, m.Type) } statVersion := strings.TrimPrefix(ss[8], statVersionPrefix) stats, err := parseMountStatsNFS(s, statVersion) if err != nil { return nil, err } m.Stats = stats } mounts = append(mounts, m) } return mounts, s.Err() } // parseMount parses an entry in /proc/[pid]/mountstats in the format: // // device [device] mounted on [mount] with fstype [type] func parseMount(ss []string) (*Mount, error) { if len(ss) < deviceEntryLen { return nil, fmt.Errorf("%w: Invalid device %q", ErrFileParse, ss) } // Check for specific words appearing at specific indices to ensure // the format is consistent with what we expect format := []struct { i int s string }{ {i: 0, s: "device"}, {i: 2, s: "mounted"}, {i: 3, s: "on"}, {i: 5, s: "with"}, {i: 6, s: "fstype"}, } for _, f := range format { if ss[f.i] != f.s { return nil, fmt.Errorf("%w: Invalid device %q", ErrFileParse, ss) } } return &Mount{ Device: ss[1], Mount: ss[4], Type: ss[7], }, nil } // parseMountStatsNFS parses a MountStatsNFS by scanning additional information // related to NFS statistics. func parseMountStatsNFS(s *bufio.Scanner, statVersion string) (*MountStatsNFS, error) { // Field indicators for parsing specific types of data const ( fieldOpts = "opts:" fieldAge = "age:" fieldBytes = "bytes:" fieldEvents = "events:" fieldPerOpStats = "per-op" fieldTransport = "xprt:" ) stats := &MountStatsNFS{ StatVersion: statVersion, } for s.Scan() { ss := strings.Fields(string(s.Bytes())) if len(ss) == 0 { break } switch ss[0] { case fieldOpts: if len(ss) < 2 { return nil, fmt.Errorf("%w: Incomplete information for NFS stats: %v", ErrFileParse, ss) } if stats.Opts == nil { stats.Opts = map[string]string{} } for _, opt := range strings.Split(ss[1], ",") { split := strings.Split(opt, "=") if len(split) == 2 { stats.Opts[split[0]] = split[1] } else { stats.Opts[opt] = "" } } case fieldAge: if len(ss) < 2 { return nil, fmt.Errorf("%w: Incomplete information for NFS stats: %v", ErrFileParse, ss) } // Age integer is in seconds d, err := time.ParseDuration(ss[1] + "s") if err != nil { return nil, err } stats.Age = d case fieldBytes: if len(ss) < 2 { return nil, fmt.Errorf("%w: Incomplete information for NFS stats: %v", ErrFileParse, ss) } bstats, err := parseNFSBytesStats(ss[1:]) if err != nil { return nil, err } stats.Bytes = *bstats case fieldEvents: if len(ss) < 2 { return nil, fmt.Errorf("%w: Incomplete information for NFS events: %v", ErrFileParse, ss) } estats, err := parseNFSEventsStats(ss[1:]) if err != nil { return nil, err } stats.Events = *estats case fieldTransport: if len(ss) < 3 { return nil, fmt.Errorf("%w: Incomplete information for NFS transport stats: %v", ErrFileParse, ss) } tstats, err := parseNFSTransportStats(ss[1:], statVersion) if err != nil { return nil, err } stats.Transport = append(stats.Transport, *tstats) } // When encountering "per-operation statistics", we must break this // loop and parse them separately to ensure we can terminate parsing // before reaching another device entry; hence why this 'if' statement // is not just another switch case if ss[0] == fieldPerOpStats { break } } if err := s.Err(); err != nil { return nil, err } // NFS per-operation stats appear last before the next device entry perOpStats, err := parseNFSOperationStats(s) if err != nil { return nil, err } stats.Operations = perOpStats return stats, nil } // parseNFSBytesStats parses a NFSBytesStats line using an input set of // integer fields. func parseNFSBytesStats(ss []string) (*NFSBytesStats, error) { if len(ss) != fieldBytesLen { return nil, fmt.Errorf("%w: Invalid NFS bytes stats: %v", ErrFileParse, ss) } ns := make([]uint64, 0, fieldBytesLen) for _, s := range ss { n, err := strconv.ParseUint(s, 10, 64) if err != nil { return nil, err } ns = append(ns, n) } return &NFSBytesStats{ Read: ns[0], Write: ns[1], DirectRead: ns[2], DirectWrite: ns[3], ReadTotal: ns[4], WriteTotal: ns[5], ReadPages: ns[6], WritePages: ns[7], }, nil } // parseNFSEventsStats parses a NFSEventsStats line using an input set of // integer fields. func parseNFSEventsStats(ss []string) (*NFSEventsStats, error) { if len(ss) != fieldEventsLen { return nil, fmt.Errorf("%w: invalid NFS events stats: %v", ErrFileParse, ss) } ns := make([]uint64, 0, fieldEventsLen) for _, s := range ss { n, err := strconv.ParseUint(s, 10, 64) if err != nil { return nil, err } ns = append(ns, n) } return &NFSEventsStats{ InodeRevalidate: ns[0], DnodeRevalidate: ns[1], DataInvalidate: ns[2], AttributeInvalidate: ns[3], VFSOpen: ns[4], VFSLookup: ns[5], VFSAccess: ns[6], VFSUpdatePage: ns[7], VFSReadPage: ns[8], VFSReadPages: ns[9], VFSWritePage: ns[10], VFSWritePages: ns[11], VFSGetdents: ns[12], VFSSetattr: ns[13], VFSFlush: ns[14], VFSFsync: ns[15], VFSLock: ns[16], VFSFileRelease: ns[17], CongestionWait: ns[18], Truncation: ns[19], WriteExtension: ns[20], SillyRename: ns[21], ShortRead: ns[22], ShortWrite: ns[23], JukeboxDelay: ns[24], PNFSRead: ns[25], PNFSWrite: ns[26], }, nil } // parseNFSOperationStats parses a slice of NFSOperationStats by scanning // additional information about per-operation statistics until an empty // line is reached. func parseNFSOperationStats(s *bufio.Scanner) ([]NFSOperationStats, error) { const ( // Minimum number of expected fields in each per-operation statistics set minFields = 9 ) var ops []NFSOperationStats for s.Scan() { ss := strings.Fields(string(s.Bytes())) if len(ss) == 0 { // Must break when reading a blank line after per-operation stats to // enable top-level function to parse the next device entry break } if len(ss) < minFields { return nil, fmt.Errorf("%w: invalid NFS per-operations stats: %v", ErrFileParse, ss) } // Skip string operation name for integers ns := make([]uint64, 0, minFields-1) for _, st := range ss[1:] { n, err := strconv.ParseUint(st, 10, 64) if err != nil { return nil, err } ns = append(ns, n) } opStats := NFSOperationStats{ Operation: strings.TrimSuffix(ss[0], ":"), Requests: ns[0], Transmissions: ns[1], MajorTimeouts: ns[2], BytesSent: ns[3], BytesReceived: ns[4], CumulativeQueueMilliseconds: ns[5], CumulativeTotalResponseMilliseconds: ns[6], CumulativeTotalRequestMilliseconds: ns[7], } if len(ns) > 8 { opStats.Errors = ns[8] } ops = append(ops, opStats) } return ops, s.Err() } // parseNFSTransportStats parses a NFSTransportStats line using an input set of // integer fields matched to a specific stats version. func parseNFSTransportStats(ss []string, statVersion string) (*NFSTransportStats, error) { // Extract the protocol field. It is the only string value in the line protocol := ss[0] ss = ss[1:] switch statVersion { case statVersion10: var expectedLength int if protocol == "tcp" { expectedLength = fieldTransport10TCPLen } else if protocol == "udp" { expectedLength = fieldTransport10UDPLen } else { return nil, fmt.Errorf("%w: Invalid NFS protocol \"%s\" in stats 1.0 statement: %v", ErrFileParse, protocol, ss) } if len(ss) != expectedLength { return nil, fmt.Errorf("%w: Invalid NFS transport stats 1.0 statement: %v", ErrFileParse, ss) } case statVersion11: var expectedLength int if protocol == "tcp" { expectedLength = fieldTransport11TCPLen } else if protocol == "udp" { expectedLength = fieldTransport11UDPLen } else if protocol == "rdma" { expectedLength = fieldTransport11RDMAMinLen } else { return nil, fmt.Errorf("%w: invalid NFS protocol \"%s\" in stats 1.1 statement: %v", ErrFileParse, protocol, ss) } if (len(ss) != expectedLength && (protocol == "tcp" || protocol == "udp")) || (protocol == "rdma" && len(ss) < expectedLength) { return nil, fmt.Errorf("%w: invalid NFS transport stats 1.1 statement: %v, protocol: %v", ErrFileParse, ss, protocol) } default: return nil, fmt.Errorf("%w: Unrecognized NFS transport stats version: %q, protocol: %v", ErrFileParse, statVersion, protocol) } // Allocate enough for v1.1 stats since zero value for v1.1 stats will be okay // in a v1.0 response. Since the stat length is bigger for TCP stats, we use // the TCP length here. // // Note: slice length must be set to length of v1.1 stats to avoid a panic when // only v1.0 stats are present. // See: https://github.com/prometheus/node_exporter/issues/571. // // Note: NFS Over RDMA slice length is fieldTransport11RDMAMaxLen ns := make([]uint64, fieldTransport11RDMAMaxLen+3) for i, s := range ss { n, err := strconv.ParseUint(s, 10, 64) if err != nil { return nil, err } ns[i] = n } // The fields differ depending on the transport protocol (TCP or UDP) // From https://utcc.utoronto.ca/%7Ecks/space/blog/linux/NFSMountstatsXprt // // For the udp RPC transport there is no connection count, connect idle time, // or idle time (fields #3, #4, and #5); all other fields are the same. So // we set them to 0 here. if protocol == "udp" { ns = append(ns[:2], append(make([]uint64, 3), ns[2:]...)...) } else if protocol == "tcp" { ns = append(ns[:fieldTransport11TCPLen], make([]uint64, fieldTransport11RDMAMaxLen-fieldTransport11TCPLen+3)...) } else if protocol == "rdma" { ns = append(ns[:fieldTransport10TCPLen], append(make([]uint64, 3), ns[fieldTransport10TCPLen:]...)...) } return &NFSTransportStats{ // NFS xprt over tcp or udp Protocol: protocol, Port: ns[0], Bind: ns[1], Connect: ns[2], ConnectIdleTime: ns[3], IdleTimeSeconds: ns[4], Sends: ns[5], Receives: ns[6], BadTransactionIDs: ns[7], CumulativeActiveRequests: ns[8], CumulativeBacklog: ns[9], // NFS xprt over tcp or udp // And statVersion 1.1 MaximumRPCSlotsUsed: ns[10], CumulativeSendingQueue: ns[11], CumulativePendingQueue: ns[12], // NFS xprt over rdma // And stat Version 1.1 ReadChunkCount: ns[13], WriteChunkCount: ns[14], ReplyChunkCount: ns[15], TotalRdmaRequest: ns[16], PullupCopyCount: ns[17], HardwayRegisterCount: ns[18], FailedMarshalCount: ns[19], BadReplyCount: ns[20], MrsRecovered: ns[21], MrsOrphaned: ns[22], MrsAllocated: ns[23], EmptySendctxQ: ns[24], TotalRdmaReply: ns[25], FixupCopyCount: ns[26], ReplyWaitsForSend: ns[27], LocalInvNeeded: ns[28], NomsgCallCount: ns[29], BcallCount: ns[30], }, nil }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/loadavg.go
vendor/github.com/prometheus/procfs/loadavg.go
// Copyright 2019 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package procfs import ( "fmt" "strconv" "strings" "github.com/prometheus/procfs/internal/util" ) // LoadAvg represents an entry in /proc/loadavg. type LoadAvg struct { Load1 float64 Load5 float64 Load15 float64 } // LoadAvg returns loadavg from /proc. func (fs FS) LoadAvg() (*LoadAvg, error) { path := fs.proc.Path("loadavg") data, err := util.ReadFileNoStat(path) if err != nil { return nil, err } return parseLoad(data) } // Parse /proc loadavg and return 1m, 5m and 15m. func parseLoad(loadavgBytes []byte) (*LoadAvg, error) { loads := make([]float64, 3) parts := strings.Fields(string(loadavgBytes)) if len(parts) < 3 { return nil, fmt.Errorf("%w: Malformed line %q", ErrFileParse, string(loadavgBytes)) } var err error for i, load := range parts[0:3] { loads[i], err = strconv.ParseFloat(load, 64) if err != nil { return nil, fmt.Errorf("%w: Cannot parse load: %f: %w", ErrFileParse, loads[i], err) } } return &LoadAvg{ Load1: loads[0], Load5: loads[1], Load15: loads[2], }, nil }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/meminfo.go
vendor/github.com/prometheus/procfs/meminfo.go
// Copyright 2019 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package procfs import ( "bufio" "bytes" "fmt" "io" "strconv" "strings" "github.com/prometheus/procfs/internal/util" ) // Meminfo represents memory statistics. type Meminfo struct { // Total usable ram (i.e. physical ram minus a few reserved // bits and the kernel binary code) MemTotal *uint64 // The sum of LowFree+HighFree MemFree *uint64 // An estimate of how much memory is available for starting // new applications, without swapping. Calculated from // MemFree, SReclaimable, the size of the file LRU lists, and // the low watermarks in each zone. The estimate takes into // account that the system needs some page cache to function // well, and that not all reclaimable slab will be // reclaimable, due to items being in use. The impact of those // factors will vary from system to system. MemAvailable *uint64 // Relatively temporary storage for raw disk blocks shouldn't // get tremendously large (20MB or so) Buffers *uint64 Cached *uint64 // Memory that once was swapped out, is swapped back in but // still also is in the swapfile (if memory is needed it // doesn't need to be swapped out AGAIN because it is already // in the swapfile. This saves I/O) SwapCached *uint64 // Memory that has been used more recently and usually not // reclaimed unless absolutely necessary. Active *uint64 // Memory which has been less recently used. It is more // eligible to be reclaimed for other purposes Inactive *uint64 ActiveAnon *uint64 InactiveAnon *uint64 ActiveFile *uint64 InactiveFile *uint64 Unevictable *uint64 Mlocked *uint64 // total amount of swap space available SwapTotal *uint64 // Memory which has been evicted from RAM, and is temporarily // on the disk SwapFree *uint64 // Memory which is waiting to get written back to the disk Dirty *uint64 // Memory which is actively being written back to the disk Writeback *uint64 // Non-file backed pages mapped into userspace page tables AnonPages *uint64 // files which have been mapped, such as libraries Mapped *uint64 Shmem *uint64 // in-kernel data structures cache Slab *uint64 // Part of Slab, that might be reclaimed, such as caches SReclaimable *uint64 // Part of Slab, that cannot be reclaimed on memory pressure SUnreclaim *uint64 KernelStack *uint64 // amount of memory dedicated to the lowest level of page // tables. PageTables *uint64 // NFS pages sent to the server, but not yet committed to // stable storage NFSUnstable *uint64 // Memory used for block device "bounce buffers" Bounce *uint64 // Memory used by FUSE for temporary writeback buffers WritebackTmp *uint64 // Based on the overcommit ratio ('vm.overcommit_ratio'), // this is the total amount of memory currently available to // be allocated on the system. This limit is only adhered to // if strict overcommit accounting is enabled (mode 2 in // 'vm.overcommit_memory'). // The CommitLimit is calculated with the following formula: // CommitLimit = ([total RAM pages] - [total huge TLB pages]) * // overcommit_ratio / 100 + [total swap pages] // For example, on a system with 1G of physical RAM and 7G // of swap with a `vm.overcommit_ratio` of 30 it would // yield a CommitLimit of 7.3G. // For more details, see the memory overcommit documentation // in vm/overcommit-accounting. CommitLimit *uint64 // The amount of memory presently allocated on the system. // The committed memory is a sum of all of the memory which // has been allocated by processes, even if it has not been // "used" by them as of yet. A process which malloc()'s 1G // of memory, but only touches 300M of it will show up as // using 1G. This 1G is memory which has been "committed" to // by the VM and can be used at any time by the allocating // application. With strict overcommit enabled on the system // (mode 2 in 'vm.overcommit_memory'),allocations which would // exceed the CommitLimit (detailed above) will not be permitted. // This is useful if one needs to guarantee that processes will // not fail due to lack of memory once that memory has been // successfully allocated. CommittedAS *uint64 // total size of vmalloc memory area VmallocTotal *uint64 // amount of vmalloc area which is used VmallocUsed *uint64 // largest contiguous block of vmalloc area which is free VmallocChunk *uint64 Percpu *uint64 HardwareCorrupted *uint64 AnonHugePages *uint64 ShmemHugePages *uint64 ShmemPmdMapped *uint64 CmaTotal *uint64 CmaFree *uint64 HugePagesTotal *uint64 HugePagesFree *uint64 HugePagesRsvd *uint64 HugePagesSurp *uint64 Hugepagesize *uint64 DirectMap4k *uint64 DirectMap2M *uint64 DirectMap1G *uint64 // The struct fields below are the byte-normalized counterparts to the // existing struct fields. Values are normalized using the optional // unit field in the meminfo line. MemTotalBytes *uint64 MemFreeBytes *uint64 MemAvailableBytes *uint64 BuffersBytes *uint64 CachedBytes *uint64 SwapCachedBytes *uint64 ActiveBytes *uint64 InactiveBytes *uint64 ActiveAnonBytes *uint64 InactiveAnonBytes *uint64 ActiveFileBytes *uint64 InactiveFileBytes *uint64 UnevictableBytes *uint64 MlockedBytes *uint64 SwapTotalBytes *uint64 SwapFreeBytes *uint64 DirtyBytes *uint64 WritebackBytes *uint64 AnonPagesBytes *uint64 MappedBytes *uint64 ShmemBytes *uint64 SlabBytes *uint64 SReclaimableBytes *uint64 SUnreclaimBytes *uint64 KernelStackBytes *uint64 PageTablesBytes *uint64 NFSUnstableBytes *uint64 BounceBytes *uint64 WritebackTmpBytes *uint64 CommitLimitBytes *uint64 CommittedASBytes *uint64 VmallocTotalBytes *uint64 VmallocUsedBytes *uint64 VmallocChunkBytes *uint64 PercpuBytes *uint64 HardwareCorruptedBytes *uint64 AnonHugePagesBytes *uint64 ShmemHugePagesBytes *uint64 ShmemPmdMappedBytes *uint64 CmaTotalBytes *uint64 CmaFreeBytes *uint64 HugepagesizeBytes *uint64 DirectMap4kBytes *uint64 DirectMap2MBytes *uint64 DirectMap1GBytes *uint64 } // Meminfo returns an information about current kernel/system memory statistics. // See https://www.kernel.org/doc/Documentation/filesystems/proc.txt func (fs FS) Meminfo() (Meminfo, error) { b, err := util.ReadFileNoStat(fs.proc.Path("meminfo")) if err != nil { return Meminfo{}, err } m, err := parseMemInfo(bytes.NewReader(b)) if err != nil { return Meminfo{}, fmt.Errorf("%w: %w", ErrFileParse, err) } return *m, nil } func parseMemInfo(r io.Reader) (*Meminfo, error) { var m Meminfo s := bufio.NewScanner(r) for s.Scan() { fields := strings.Fields(s.Text()) var val, valBytes uint64 val, err := strconv.ParseUint(fields[1], 0, 64) if err != nil { return nil, err } switch len(fields) { case 2: // No unit present, use the parsed the value as bytes directly. valBytes = val case 3: // Unit present in optional 3rd field, convert it to // bytes. The only unit supported within the Linux // kernel is `kB`. if fields[2] != "kB" { return nil, fmt.Errorf("%w: Unsupported unit in optional 3rd field %q", ErrFileParse, fields[2]) } valBytes = 1024 * val default: return nil, fmt.Errorf("%w: Malformed line %q", ErrFileParse, s.Text()) } switch fields[0] { case "MemTotal:": m.MemTotal = &val m.MemTotalBytes = &valBytes case "MemFree:": m.MemFree = &val m.MemFreeBytes = &valBytes case "MemAvailable:": m.MemAvailable = &val m.MemAvailableBytes = &valBytes case "Buffers:": m.Buffers = &val m.BuffersBytes = &valBytes case "Cached:": m.Cached = &val m.CachedBytes = &valBytes case "SwapCached:": m.SwapCached = &val m.SwapCachedBytes = &valBytes case "Active:": m.Active = &val m.ActiveBytes = &valBytes case "Inactive:": m.Inactive = &val m.InactiveBytes = &valBytes case "Active(anon):": m.ActiveAnon = &val m.ActiveAnonBytes = &valBytes case "Inactive(anon):": m.InactiveAnon = &val m.InactiveAnonBytes = &valBytes case "Active(file):": m.ActiveFile = &val m.ActiveFileBytes = &valBytes case "Inactive(file):": m.InactiveFile = &val m.InactiveFileBytes = &valBytes case "Unevictable:": m.Unevictable = &val m.UnevictableBytes = &valBytes case "Mlocked:": m.Mlocked = &val m.MlockedBytes = &valBytes case "SwapTotal:": m.SwapTotal = &val m.SwapTotalBytes = &valBytes case "SwapFree:": m.SwapFree = &val m.SwapFreeBytes = &valBytes case "Dirty:": m.Dirty = &val m.DirtyBytes = &valBytes case "Writeback:": m.Writeback = &val m.WritebackBytes = &valBytes case "AnonPages:": m.AnonPages = &val m.AnonPagesBytes = &valBytes case "Mapped:": m.Mapped = &val m.MappedBytes = &valBytes case "Shmem:": m.Shmem = &val m.ShmemBytes = &valBytes case "Slab:": m.Slab = &val m.SlabBytes = &valBytes case "SReclaimable:": m.SReclaimable = &val m.SReclaimableBytes = &valBytes case "SUnreclaim:": m.SUnreclaim = &val m.SUnreclaimBytes = &valBytes case "KernelStack:": m.KernelStack = &val m.KernelStackBytes = &valBytes case "PageTables:": m.PageTables = &val m.PageTablesBytes = &valBytes case "NFS_Unstable:": m.NFSUnstable = &val m.NFSUnstableBytes = &valBytes case "Bounce:": m.Bounce = &val m.BounceBytes = &valBytes case "WritebackTmp:": m.WritebackTmp = &val m.WritebackTmpBytes = &valBytes case "CommitLimit:": m.CommitLimit = &val m.CommitLimitBytes = &valBytes case "Committed_AS:": m.CommittedAS = &val m.CommittedASBytes = &valBytes case "VmallocTotal:": m.VmallocTotal = &val m.VmallocTotalBytes = &valBytes case "VmallocUsed:": m.VmallocUsed = &val m.VmallocUsedBytes = &valBytes case "VmallocChunk:": m.VmallocChunk = &val m.VmallocChunkBytes = &valBytes case "Percpu:": m.Percpu = &val m.PercpuBytes = &valBytes case "HardwareCorrupted:": m.HardwareCorrupted = &val m.HardwareCorruptedBytes = &valBytes case "AnonHugePages:": m.AnonHugePages = &val m.AnonHugePagesBytes = &valBytes case "ShmemHugePages:": m.ShmemHugePages = &val m.ShmemHugePagesBytes = &valBytes case "ShmemPmdMapped:": m.ShmemPmdMapped = &val m.ShmemPmdMappedBytes = &valBytes case "CmaTotal:": m.CmaTotal = &val m.CmaTotalBytes = &valBytes case "CmaFree:": m.CmaFree = &val m.CmaFreeBytes = &valBytes case "HugePages_Total:": m.HugePagesTotal = &val case "HugePages_Free:": m.HugePagesFree = &val case "HugePages_Rsvd:": m.HugePagesRsvd = &val case "HugePages_Surp:": m.HugePagesSurp = &val case "Hugepagesize:": m.Hugepagesize = &val m.HugepagesizeBytes = &valBytes case "DirectMap4k:": m.DirectMap4k = &val m.DirectMap4kBytes = &valBytes case "DirectMap2M:": m.DirectMap2M = &val m.DirectMap2MBytes = &valBytes case "DirectMap1G:": m.DirectMap1G = &val m.DirectMap1GBytes = &valBytes } } return &m, nil }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/cmdline.go
vendor/github.com/prometheus/procfs/cmdline.go
// Copyright 2021 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package procfs import ( "strings" "github.com/prometheus/procfs/internal/util" ) // CmdLine returns the command line of the kernel. func (fs FS) CmdLine() ([]string, error) { data, err := util.ReadFileNoStat(fs.proc.Path("cmdline")) if err != nil { return nil, err } return strings.Fields(string(data)), nil }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/proc_cgroup.go
vendor/github.com/prometheus/procfs/proc_cgroup.go
// Copyright 2020 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package procfs import ( "bufio" "bytes" "fmt" "strconv" "strings" "github.com/prometheus/procfs/internal/util" ) // Cgroup models one line from /proc/[pid]/cgroup. Each Cgroup struct describes the placement of a PID inside a // specific control hierarchy. The kernel has two cgroup APIs, v1 and v2. v1 has one hierarchy per available resource // controller, while v2 has one unified hierarchy shared by all controllers. Regardless of v1 or v2, all hierarchies // contain all running processes, so the question answerable with a Cgroup struct is 'where is this process in // this hierarchy' (where==what path on the specific cgroupfs). By prefixing this path with the mount point of // *this specific* hierarchy, you can locate the relevant pseudo-files needed to read/set the data for this PID // in this hierarchy // // Also see http://man7.org/linux/man-pages/man7/cgroups.7.html type Cgroup struct { // HierarchyID that can be matched to a named hierarchy using /proc/cgroups. Cgroups V2 only has one // hierarchy, so HierarchyID is always 0. For cgroups v1 this is a unique ID number HierarchyID int // Controllers using this hierarchy of processes. Controllers are also known as subsystems. For // Cgroups V2 this may be empty, as all active controllers use the same hierarchy Controllers []string // Path of this control group, relative to the mount point of the cgroupfs representing this specific // hierarchy Path string } // parseCgroupString parses each line of the /proc/[pid]/cgroup file // Line format is hierarchyID:[controller1,controller2]:path. func parseCgroupString(cgroupStr string) (*Cgroup, error) { var err error fields := strings.SplitN(cgroupStr, ":", 3) if len(fields) < 3 { return nil, fmt.Errorf("%w: 3+ fields required, found %d fields in cgroup string: %s", ErrFileParse, len(fields), cgroupStr) } cgroup := &Cgroup{ Path: fields[2], Controllers: nil, } cgroup.HierarchyID, err = strconv.Atoi(fields[0]) if err != nil { return nil, fmt.Errorf("%w: hierarchy ID: %q", ErrFileParse, cgroup.HierarchyID) } if fields[1] != "" { ssNames := strings.Split(fields[1], ",") cgroup.Controllers = append(cgroup.Controllers, ssNames...) } return cgroup, nil } // parseCgroups reads each line of the /proc/[pid]/cgroup file. func parseCgroups(data []byte) ([]Cgroup, error) { var cgroups []Cgroup scanner := bufio.NewScanner(bytes.NewReader(data)) for scanner.Scan() { mountString := scanner.Text() parsedMounts, err := parseCgroupString(mountString) if err != nil { return nil, err } cgroups = append(cgroups, *parsedMounts) } err := scanner.Err() return cgroups, err } // Cgroups reads from /proc/<pid>/cgroups and returns a []*Cgroup struct locating this PID in each process // control hierarchy running on this system. On every system (v1 and v2), all hierarchies contain all processes, // so the len of the returned struct is equal to the number of active hierarchies on this system. func (p Proc) Cgroups() ([]Cgroup, error) { data, err := util.ReadFileNoStat(p.path("cgroup")) if err != nil { return nil, err } return parseCgroups(data) }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/softirqs.go
vendor/github.com/prometheus/procfs/softirqs.go
// Copyright 2022 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package procfs import ( "bufio" "bytes" "fmt" "io" "strconv" "strings" "github.com/prometheus/procfs/internal/util" ) // Softirqs represents the softirq statistics. type Softirqs struct { Hi []uint64 Timer []uint64 NetTx []uint64 NetRx []uint64 Block []uint64 IRQPoll []uint64 Tasklet []uint64 Sched []uint64 HRTimer []uint64 RCU []uint64 } func (fs FS) Softirqs() (Softirqs, error) { fileName := fs.proc.Path("softirqs") data, err := util.ReadFileNoStat(fileName) if err != nil { return Softirqs{}, err } reader := bytes.NewReader(data) return parseSoftirqs(reader) } func parseSoftirqs(r io.Reader) (Softirqs, error) { var ( softirqs = Softirqs{} scanner = bufio.NewScanner(r) ) if !scanner.Scan() { return Softirqs{}, fmt.Errorf("%w: softirqs empty", ErrFileRead) } for scanner.Scan() { parts := strings.Fields(scanner.Text()) var err error // require at least one cpu if len(parts) < 2 { continue } switch { case parts[0] == "HI:": perCPU := parts[1:] softirqs.Hi = make([]uint64, len(perCPU)) for i, count := range perCPU { if softirqs.Hi[i], err = strconv.ParseUint(count, 10, 64); err != nil { return Softirqs{}, fmt.Errorf("%w: couldn't parse %q (HI%d): %w", ErrFileParse, count, i, err) } } case parts[0] == "TIMER:": perCPU := parts[1:] softirqs.Timer = make([]uint64, len(perCPU)) for i, count := range perCPU { if softirqs.Timer[i], err = strconv.ParseUint(count, 10, 64); err != nil { return Softirqs{}, fmt.Errorf("%w: couldn't parse %q (TIMER%d): %w", ErrFileParse, count, i, err) } } case parts[0] == "NET_TX:": perCPU := parts[1:] softirqs.NetTx = make([]uint64, len(perCPU)) for i, count := range perCPU { if softirqs.NetTx[i], err = strconv.ParseUint(count, 10, 64); err != nil { return Softirqs{}, fmt.Errorf("%w: couldn't parse %q (NET_TX%d): %w", ErrFileParse, count, i, err) } } case parts[0] == "NET_RX:": perCPU := parts[1:] softirqs.NetRx = make([]uint64, len(perCPU)) for i, count := range perCPU { if softirqs.NetRx[i], err = strconv.ParseUint(count, 10, 64); err != nil { return Softirqs{}, fmt.Errorf("%w: couldn't parse %q (NET_RX%d): %w", ErrFileParse, count, i, err) } } case parts[0] == "BLOCK:": perCPU := parts[1:] softirqs.Block = make([]uint64, len(perCPU)) for i, count := range perCPU { if softirqs.Block[i], err = strconv.ParseUint(count, 10, 64); err != nil { return Softirqs{}, fmt.Errorf("%w: couldn't parse %q (BLOCK%d): %w", ErrFileParse, count, i, err) } } case parts[0] == "IRQ_POLL:": perCPU := parts[1:] softirqs.IRQPoll = make([]uint64, len(perCPU)) for i, count := range perCPU { if softirqs.IRQPoll[i], err = strconv.ParseUint(count, 10, 64); err != nil { return Softirqs{}, fmt.Errorf("%w: couldn't parse %q (IRQ_POLL%d): %w", ErrFileParse, count, i, err) } } case parts[0] == "TASKLET:": perCPU := parts[1:] softirqs.Tasklet = make([]uint64, len(perCPU)) for i, count := range perCPU { if softirqs.Tasklet[i], err = strconv.ParseUint(count, 10, 64); err != nil { return Softirqs{}, fmt.Errorf("%w: couldn't parse %q (TASKLET%d): %w", ErrFileParse, count, i, err) } } case parts[0] == "SCHED:": perCPU := parts[1:] softirqs.Sched = make([]uint64, len(perCPU)) for i, count := range perCPU { if softirqs.Sched[i], err = strconv.ParseUint(count, 10, 64); err != nil { return Softirqs{}, fmt.Errorf("%w: couldn't parse %q (SCHED%d): %w", ErrFileParse, count, i, err) } } case parts[0] == "HRTIMER:": perCPU := parts[1:] softirqs.HRTimer = make([]uint64, len(perCPU)) for i, count := range perCPU { if softirqs.HRTimer[i], err = strconv.ParseUint(count, 10, 64); err != nil { return Softirqs{}, fmt.Errorf("%w: couldn't parse %q (HRTIMER%d): %w", ErrFileParse, count, i, err) } } case parts[0] == "RCU:": perCPU := parts[1:] softirqs.RCU = make([]uint64, len(perCPU)) for i, count := range perCPU { if softirqs.RCU[i], err = strconv.ParseUint(count, 10, 64); err != nil { return Softirqs{}, fmt.Errorf("%w: couldn't parse %q (RCU%d): %w", ErrFileParse, count, i, err) } } } } if err := scanner.Err(); err != nil { return Softirqs{}, fmt.Errorf("%w: couldn't parse softirqs: %w", ErrFileParse, err) } return softirqs, scanner.Err() }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/zoneinfo.go
vendor/github.com/prometheus/procfs/zoneinfo.go
// Copyright 2019 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. //go:build !windows // +build !windows package procfs import ( "bytes" "fmt" "os" "regexp" "strings" "github.com/prometheus/procfs/internal/util" ) // Zoneinfo holds info parsed from /proc/zoneinfo. type Zoneinfo struct { Node string Zone string NrFreePages *int64 Min *int64 Low *int64 High *int64 Scanned *int64 Spanned *int64 Present *int64 Managed *int64 NrActiveAnon *int64 NrInactiveAnon *int64 NrIsolatedAnon *int64 NrAnonPages *int64 NrAnonTransparentHugepages *int64 NrActiveFile *int64 NrInactiveFile *int64 NrIsolatedFile *int64 NrFilePages *int64 NrSlabReclaimable *int64 NrSlabUnreclaimable *int64 NrMlockStack *int64 NrKernelStack *int64 NrMapped *int64 NrDirty *int64 NrWriteback *int64 NrUnevictable *int64 NrShmem *int64 NrDirtied *int64 NrWritten *int64 NumaHit *int64 NumaMiss *int64 NumaForeign *int64 NumaInterleave *int64 NumaLocal *int64 NumaOther *int64 Protection []*int64 } var nodeZoneRE = regexp.MustCompile(`(\d+), zone\s+(\w+)`) // Zoneinfo parses an zoneinfo-file (/proc/zoneinfo) and returns a slice of // structs containing the relevant info. More information available here: // https://www.kernel.org/doc/Documentation/sysctl/vm.txt func (fs FS) Zoneinfo() ([]Zoneinfo, error) { data, err := os.ReadFile(fs.proc.Path("zoneinfo")) if err != nil { return nil, fmt.Errorf("%w: error reading zoneinfo %q: %w", ErrFileRead, fs.proc.Path("zoneinfo"), err) } zoneinfo, err := parseZoneinfo(data) if err != nil { return nil, fmt.Errorf("%w: error parsing zoneinfo %q: %w", ErrFileParse, fs.proc.Path("zoneinfo"), err) } return zoneinfo, nil } func parseZoneinfo(zoneinfoData []byte) ([]Zoneinfo, error) { zoneinfo := []Zoneinfo{} zoneinfoBlocks := bytes.Split(zoneinfoData, []byte("\nNode")) for _, block := range zoneinfoBlocks { var zoneinfoElement Zoneinfo lines := strings.Split(string(block), "\n") for _, line := range lines { if nodeZone := nodeZoneRE.FindStringSubmatch(line); nodeZone != nil { zoneinfoElement.Node = nodeZone[1] zoneinfoElement.Zone = nodeZone[2] continue } if strings.HasPrefix(strings.TrimSpace(line), "per-node stats") { continue } parts := strings.Fields(strings.TrimSpace(line)) if len(parts) < 2 { continue } vp := util.NewValueParser(parts[1]) switch parts[0] { case "nr_free_pages": zoneinfoElement.NrFreePages = vp.PInt64() case "min": zoneinfoElement.Min = vp.PInt64() case "low": zoneinfoElement.Low = vp.PInt64() case "high": zoneinfoElement.High = vp.PInt64() case "scanned": zoneinfoElement.Scanned = vp.PInt64() case "spanned": zoneinfoElement.Spanned = vp.PInt64() case "present": zoneinfoElement.Present = vp.PInt64() case "managed": zoneinfoElement.Managed = vp.PInt64() case "nr_active_anon": zoneinfoElement.NrActiveAnon = vp.PInt64() case "nr_inactive_anon": zoneinfoElement.NrInactiveAnon = vp.PInt64() case "nr_isolated_anon": zoneinfoElement.NrIsolatedAnon = vp.PInt64() case "nr_anon_pages": zoneinfoElement.NrAnonPages = vp.PInt64() case "nr_anon_transparent_hugepages": zoneinfoElement.NrAnonTransparentHugepages = vp.PInt64() case "nr_active_file": zoneinfoElement.NrActiveFile = vp.PInt64() case "nr_inactive_file": zoneinfoElement.NrInactiveFile = vp.PInt64() case "nr_isolated_file": zoneinfoElement.NrIsolatedFile = vp.PInt64() case "nr_file_pages": zoneinfoElement.NrFilePages = vp.PInt64() case "nr_slab_reclaimable": zoneinfoElement.NrSlabReclaimable = vp.PInt64() case "nr_slab_unreclaimable": zoneinfoElement.NrSlabUnreclaimable = vp.PInt64() case "nr_mlock_stack": zoneinfoElement.NrMlockStack = vp.PInt64() case "nr_kernel_stack": zoneinfoElement.NrKernelStack = vp.PInt64() case "nr_mapped": zoneinfoElement.NrMapped = vp.PInt64() case "nr_dirty": zoneinfoElement.NrDirty = vp.PInt64() case "nr_writeback": zoneinfoElement.NrWriteback = vp.PInt64() case "nr_unevictable": zoneinfoElement.NrUnevictable = vp.PInt64() case "nr_shmem": zoneinfoElement.NrShmem = vp.PInt64() case "nr_dirtied": zoneinfoElement.NrDirtied = vp.PInt64() case "nr_written": zoneinfoElement.NrWritten = vp.PInt64() case "numa_hit": zoneinfoElement.NumaHit = vp.PInt64() case "numa_miss": zoneinfoElement.NumaMiss = vp.PInt64() case "numa_foreign": zoneinfoElement.NumaForeign = vp.PInt64() case "numa_interleave": zoneinfoElement.NumaInterleave = vp.PInt64() case "numa_local": zoneinfoElement.NumaLocal = vp.PInt64() case "numa_other": zoneinfoElement.NumaOther = vp.PInt64() case "protection:": protectionParts := strings.Split(line, ":") protectionValues := strings.Replace(protectionParts[1], "(", "", 1) protectionValues = strings.Replace(protectionValues, ")", "", 1) protectionValues = strings.TrimSpace(protectionValues) protectionStringMap := strings.Split(protectionValues, ", ") val, err := util.ParsePInt64s(protectionStringMap) if err == nil { zoneinfoElement.Protection = val } } } zoneinfo = append(zoneinfo, zoneinfoElement) } return zoneinfo, nil }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/proc_fdinfo.go
vendor/github.com/prometheus/procfs/proc_fdinfo.go
// Copyright 2019 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package procfs import ( "bufio" "bytes" "fmt" "regexp" "github.com/prometheus/procfs/internal/util" ) var ( rPos = regexp.MustCompile(`^pos:\s+(\d+)$`) rFlags = regexp.MustCompile(`^flags:\s+(\d+)$`) rMntID = regexp.MustCompile(`^mnt_id:\s+(\d+)$`) rIno = regexp.MustCompile(`^ino:\s+(\d+)$`) rInotify = regexp.MustCompile(`^inotify`) rInotifyParts = regexp.MustCompile(`^inotify\s+wd:([0-9a-f]+)\s+ino:([0-9a-f]+)\s+sdev:([0-9a-f]+)(?:\s+mask:([0-9a-f]+))?`) ) // ProcFDInfo contains represents file descriptor information. type ProcFDInfo struct { // File descriptor FD string // File offset Pos string // File access mode and status flags Flags string // Mount point ID MntID string // Inode number Ino string // List of inotify lines (structured) in the fdinfo file (kernel 3.8+ only) InotifyInfos []InotifyInfo } // FDInfo constructor. On kernels older than 3.8, InotifyInfos will always be empty. func (p Proc) FDInfo(fd string) (*ProcFDInfo, error) { data, err := util.ReadFileNoStat(p.path("fdinfo", fd)) if err != nil { return nil, err } var text, pos, flags, mntid, ino string var inotify []InotifyInfo scanner := bufio.NewScanner(bytes.NewReader(data)) for scanner.Scan() { text = scanner.Text() if rPos.MatchString(text) { pos = rPos.FindStringSubmatch(text)[1] } else if rFlags.MatchString(text) { flags = rFlags.FindStringSubmatch(text)[1] } else if rMntID.MatchString(text) { mntid = rMntID.FindStringSubmatch(text)[1] } else if rIno.MatchString(text) { ino = rIno.FindStringSubmatch(text)[1] } else if rInotify.MatchString(text) { newInotify, err := parseInotifyInfo(text) if err != nil { return nil, err } inotify = append(inotify, *newInotify) } } i := &ProcFDInfo{ FD: fd, Pos: pos, Flags: flags, MntID: mntid, Ino: ino, InotifyInfos: inotify, } return i, nil } // InotifyInfo represents a single inotify line in the fdinfo file. type InotifyInfo struct { // Watch descriptor number WD string // Inode number Ino string // Device ID Sdev string // Mask of events being monitored Mask string } // InotifyInfo constructor. Only available on kernel 3.8+. func parseInotifyInfo(line string) (*InotifyInfo, error) { m := rInotifyParts.FindStringSubmatch(line) if len(m) >= 4 { var mask string if len(m) == 5 { mask = m[4] } i := &InotifyInfo{ WD: m[1], Ino: m[2], Sdev: m[3], Mask: mask, } return i, nil } return nil, fmt.Errorf("%w: invalid inode entry: %q", ErrFileParse, line) } // ProcFDInfos represents a list of ProcFDInfo structs. type ProcFDInfos []ProcFDInfo func (p ProcFDInfos) Len() int { return len(p) } func (p ProcFDInfos) Swap(i, j int) { p[i], p[j] = p[j], p[i] } func (p ProcFDInfos) Less(i, j int) bool { return p[i].FD < p[j].FD } // InotifyWatchLen returns the total number of inotify watches. func (p ProcFDInfos) InotifyWatchLen() (int, error) { length := 0 for _, f := range p { length += len(f.InotifyInfos) } return length, nil }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/proc_psi.go
vendor/github.com/prometheus/procfs/proc_psi.go
// Copyright 2019 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package procfs // The PSI / pressure interface is described at // https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git/tree/Documentation/accounting/psi.txt // Each resource (cpu, io, memory, ...) is exposed as a single file. // Each file may contain up to two lines, one for "some" pressure and one for "full" pressure. // Each line contains several averages (over n seconds) and a total in µs. // // Example io pressure file: // > some avg10=0.06 avg60=0.21 avg300=0.99 total=8537362 // > full avg10=0.00 avg60=0.13 avg300=0.96 total=8183134 import ( "bufio" "bytes" "fmt" "io" "strings" "github.com/prometheus/procfs/internal/util" ) const lineFormat = "avg10=%f avg60=%f avg300=%f total=%d" // PSILine is a single line of values as returned by `/proc/pressure/*`. // // The Avg entries are averages over n seconds, as a percentage. // The Total line is in microseconds. type PSILine struct { Avg10 float64 Avg60 float64 Avg300 float64 Total uint64 } // PSIStats represent pressure stall information from /proc/pressure/* // // "Some" indicates the share of time in which at least some tasks are stalled. // "Full" indicates the share of time in which all non-idle tasks are stalled simultaneously. type PSIStats struct { Some *PSILine Full *PSILine } // PSIStatsForResource reads pressure stall information for the specified // resource from /proc/pressure/<resource>. At time of writing this can be // either "cpu", "memory" or "io". func (fs FS) PSIStatsForResource(resource string) (PSIStats, error) { data, err := util.ReadFileNoStat(fs.proc.Path(fmt.Sprintf("%s/%s", "pressure", resource))) if err != nil { return PSIStats{}, fmt.Errorf("%w: psi_stats: unavailable for %q: %w", ErrFileRead, resource, err) } return parsePSIStats(bytes.NewReader(data)) } // parsePSIStats parses the specified file for pressure stall information. func parsePSIStats(r io.Reader) (PSIStats, error) { psiStats := PSIStats{} scanner := bufio.NewScanner(r) for scanner.Scan() { l := scanner.Text() prefix := strings.Split(l, " ")[0] switch prefix { case "some": psi := PSILine{} _, err := fmt.Sscanf(l, fmt.Sprintf("some %s", lineFormat), &psi.Avg10, &psi.Avg60, &psi.Avg300, &psi.Total) if err != nil { return PSIStats{}, err } psiStats.Some = &psi case "full": psi := PSILine{} _, err := fmt.Sscanf(l, fmt.Sprintf("full %s", lineFormat), &psi.Avg10, &psi.Avg60, &psi.Avg300, &psi.Total) if err != nil { return PSIStats{}, err } psiStats.Full = &psi default: // If we encounter a line with an unknown prefix, ignore it and move on // Should new measurement types be added in the future we'll simply ignore them instead // of erroring on retrieval continue } } return psiStats, nil }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/buddyinfo.go
vendor/github.com/prometheus/procfs/buddyinfo.go
// Copyright 2017 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package procfs import ( "bufio" "fmt" "io" "os" "strconv" "strings" ) // A BuddyInfo is the details parsed from /proc/buddyinfo. // The data is comprised of an array of free fragments of each size. // The sizes are 2^n*PAGE_SIZE, where n is the array index. type BuddyInfo struct { Node string Zone string Sizes []float64 } // BuddyInfo reads the buddyinfo statistics from the specified `proc` filesystem. func (fs FS) BuddyInfo() ([]BuddyInfo, error) { file, err := os.Open(fs.proc.Path("buddyinfo")) if err != nil { return nil, err } defer file.Close() return parseBuddyInfo(file) } func parseBuddyInfo(r io.Reader) ([]BuddyInfo, error) { var ( buddyInfo = []BuddyInfo{} scanner = bufio.NewScanner(r) bucketCount = -1 ) for scanner.Scan() { var err error line := scanner.Text() parts := strings.Fields(line) if len(parts) < 4 { return nil, fmt.Errorf("%w: Invalid number of fields, found: %v", ErrFileParse, parts) } node := strings.TrimSuffix(parts[1], ",") zone := strings.TrimSuffix(parts[3], ",") arraySize := len(parts[4:]) if bucketCount == -1 { bucketCount = arraySize } else { if bucketCount != arraySize { return nil, fmt.Errorf("%w: mismatch in number of buddyinfo buckets, previous count %d, new count %d", ErrFileParse, bucketCount, arraySize) } } sizes := make([]float64, arraySize) for i := 0; i < arraySize; i++ { sizes[i], err = strconv.ParseFloat(parts[i+4], 64) if err != nil { return nil, fmt.Errorf("%w: Invalid valid in buddyinfo: %f: %w", ErrFileParse, sizes[i], err) } } buddyInfo = append(buddyInfo, BuddyInfo{node, zone, sizes}) } return buddyInfo, scanner.Err() }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/net_dev.go
vendor/github.com/prometheus/procfs/net_dev.go
// Copyright 2018 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package procfs import ( "bufio" "errors" "os" "sort" "strconv" "strings" ) // NetDevLine is single line parsed from /proc/net/dev or /proc/[pid]/net/dev. type NetDevLine struct { Name string `json:"name"` // The name of the interface. RxBytes uint64 `json:"rx_bytes"` // Cumulative count of bytes received. RxPackets uint64 `json:"rx_packets"` // Cumulative count of packets received. RxErrors uint64 `json:"rx_errors"` // Cumulative count of receive errors encountered. RxDropped uint64 `json:"rx_dropped"` // Cumulative count of packets dropped while receiving. RxFIFO uint64 `json:"rx_fifo"` // Cumulative count of FIFO buffer errors. RxFrame uint64 `json:"rx_frame"` // Cumulative count of packet framing errors. RxCompressed uint64 `json:"rx_compressed"` // Cumulative count of compressed packets received by the device driver. RxMulticast uint64 `json:"rx_multicast"` // Cumulative count of multicast frames received by the device driver. TxBytes uint64 `json:"tx_bytes"` // Cumulative count of bytes transmitted. TxPackets uint64 `json:"tx_packets"` // Cumulative count of packets transmitted. TxErrors uint64 `json:"tx_errors"` // Cumulative count of transmit errors encountered. TxDropped uint64 `json:"tx_dropped"` // Cumulative count of packets dropped while transmitting. TxFIFO uint64 `json:"tx_fifo"` // Cumulative count of FIFO buffer errors. TxCollisions uint64 `json:"tx_collisions"` // Cumulative count of collisions detected on the interface. TxCarrier uint64 `json:"tx_carrier"` // Cumulative count of carrier losses detected by the device driver. TxCompressed uint64 `json:"tx_compressed"` // Cumulative count of compressed packets transmitted by the device driver. } // NetDev is parsed from /proc/net/dev or /proc/[pid]/net/dev. The map keys // are interface names. type NetDev map[string]NetDevLine // NetDev returns kernel/system statistics read from /proc/net/dev. func (fs FS) NetDev() (NetDev, error) { return newNetDev(fs.proc.Path("net/dev")) } // NetDev returns kernel/system statistics read from /proc/[pid]/net/dev. func (p Proc) NetDev() (NetDev, error) { return newNetDev(p.path("net/dev")) } // newNetDev creates a new NetDev from the contents of the given file. func newNetDev(file string) (NetDev, error) { f, err := os.Open(file) if err != nil { return NetDev{}, err } defer f.Close() netDev := NetDev{} s := bufio.NewScanner(f) for n := 0; s.Scan(); n++ { // Skip the 2 header lines. if n < 2 { continue } line, err := netDev.parseLine(s.Text()) if err != nil { return netDev, err } netDev[line.Name] = *line } return netDev, s.Err() } // parseLine parses a single line from the /proc/net/dev file. Header lines // must be filtered prior to calling this method. func (netDev NetDev) parseLine(rawLine string) (*NetDevLine, error) { idx := strings.LastIndex(rawLine, ":") if idx == -1 { return nil, errors.New("invalid net/dev line, missing colon") } fields := strings.Fields(strings.TrimSpace(rawLine[idx+1:])) var err error line := &NetDevLine{} // Interface Name line.Name = strings.TrimSpace(rawLine[:idx]) if line.Name == "" { return nil, errors.New("invalid net/dev line, empty interface name") } // RX line.RxBytes, err = strconv.ParseUint(fields[0], 10, 64) if err != nil { return nil, err } line.RxPackets, err = strconv.ParseUint(fields[1], 10, 64) if err != nil { return nil, err } line.RxErrors, err = strconv.ParseUint(fields[2], 10, 64) if err != nil { return nil, err } line.RxDropped, err = strconv.ParseUint(fields[3], 10, 64) if err != nil { return nil, err } line.RxFIFO, err = strconv.ParseUint(fields[4], 10, 64) if err != nil { return nil, err } line.RxFrame, err = strconv.ParseUint(fields[5], 10, 64) if err != nil { return nil, err } line.RxCompressed, err = strconv.ParseUint(fields[6], 10, 64) if err != nil { return nil, err } line.RxMulticast, err = strconv.ParseUint(fields[7], 10, 64) if err != nil { return nil, err } // TX line.TxBytes, err = strconv.ParseUint(fields[8], 10, 64) if err != nil { return nil, err } line.TxPackets, err = strconv.ParseUint(fields[9], 10, 64) if err != nil { return nil, err } line.TxErrors, err = strconv.ParseUint(fields[10], 10, 64) if err != nil { return nil, err } line.TxDropped, err = strconv.ParseUint(fields[11], 10, 64) if err != nil { return nil, err } line.TxFIFO, err = strconv.ParseUint(fields[12], 10, 64) if err != nil { return nil, err } line.TxCollisions, err = strconv.ParseUint(fields[13], 10, 64) if err != nil { return nil, err } line.TxCarrier, err = strconv.ParseUint(fields[14], 10, 64) if err != nil { return nil, err } line.TxCompressed, err = strconv.ParseUint(fields[15], 10, 64) if err != nil { return nil, err } return line, nil } // Total aggregates the values across interfaces and returns a new NetDevLine. // The Name field will be a sorted comma separated list of interface names. func (netDev NetDev) Total() NetDevLine { total := NetDevLine{} names := make([]string, 0, len(netDev)) for _, ifc := range netDev { names = append(names, ifc.Name) total.RxBytes += ifc.RxBytes total.RxPackets += ifc.RxPackets total.RxErrors += ifc.RxErrors total.RxDropped += ifc.RxDropped total.RxFIFO += ifc.RxFIFO total.RxFrame += ifc.RxFrame total.RxCompressed += ifc.RxCompressed total.RxMulticast += ifc.RxMulticast total.TxBytes += ifc.TxBytes total.TxPackets += ifc.TxPackets total.TxErrors += ifc.TxErrors total.TxDropped += ifc.TxDropped total.TxFIFO += ifc.TxFIFO total.TxCollisions += ifc.TxCollisions total.TxCarrier += ifc.TxCarrier total.TxCompressed += ifc.TxCompressed } sort.Strings(names) total.Name = strings.Join(names, ", ") return total }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/proc_io.go
vendor/github.com/prometheus/procfs/proc_io.go
// Copyright 2018 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package procfs import ( "fmt" "github.com/prometheus/procfs/internal/util" ) // ProcIO models the content of /proc/<pid>/io. type ProcIO struct { // Chars read. RChar uint64 // Chars written. WChar uint64 // Read syscalls. SyscR uint64 // Write syscalls. SyscW uint64 // Bytes read. ReadBytes uint64 // Bytes written. WriteBytes uint64 // Bytes written, but taking into account truncation. See // Documentation/filesystems/proc.txt in the kernel sources for // detailed explanation. CancelledWriteBytes int64 } // IO creates a new ProcIO instance from a given Proc instance. func (p Proc) IO() (ProcIO, error) { pio := ProcIO{} data, err := util.ReadFileNoStat(p.path("io")) if err != nil { return pio, err } ioFormat := "rchar: %d\nwchar: %d\nsyscr: %d\nsyscw: %d\n" + "read_bytes: %d\nwrite_bytes: %d\n" + "cancelled_write_bytes: %d\n" _, err = fmt.Sscanf(string(data), ioFormat, &pio.RChar, &pio.WChar, &pio.SyscR, &pio.SyscW, &pio.ReadBytes, &pio.WriteBytes, &pio.CancelledWriteBytes) return pio, err }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/net_route.go
vendor/github.com/prometheus/procfs/net_route.go
// Copyright 2023 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package procfs import ( "bufio" "bytes" "fmt" "io" "strconv" "strings" "github.com/prometheus/procfs/internal/util" ) const ( blackholeRepresentation string = "*" blackholeIfaceName string = "blackhole" routeLineColumns int = 11 ) // A NetRouteLine represents one line from net/route. type NetRouteLine struct { Iface string Destination uint32 Gateway uint32 Flags uint32 RefCnt uint32 Use uint32 Metric uint32 Mask uint32 MTU uint32 Window uint32 IRTT uint32 } func (fs FS) NetRoute() ([]NetRouteLine, error) { return readNetRoute(fs.proc.Path("net", "route")) } func readNetRoute(path string) ([]NetRouteLine, error) { b, err := util.ReadFileNoStat(path) if err != nil { return nil, err } routelines, err := parseNetRoute(bytes.NewReader(b)) if err != nil { return nil, fmt.Errorf("failed to read net route from %s: %w", path, err) } return routelines, nil } func parseNetRoute(r io.Reader) ([]NetRouteLine, error) { var routelines []NetRouteLine scanner := bufio.NewScanner(r) scanner.Scan() for scanner.Scan() { fields := strings.Fields(scanner.Text()) routeline, err := parseNetRouteLine(fields) if err != nil { return nil, err } routelines = append(routelines, *routeline) } return routelines, nil } func parseNetRouteLine(fields []string) (*NetRouteLine, error) { if len(fields) != routeLineColumns { return nil, fmt.Errorf("invalid routeline, num of digits: %d", len(fields)) } iface := fields[0] if iface == blackholeRepresentation { iface = blackholeIfaceName } destination, err := strconv.ParseUint(fields[1], 16, 32) if err != nil { return nil, err } gateway, err := strconv.ParseUint(fields[2], 16, 32) if err != nil { return nil, err } flags, err := strconv.ParseUint(fields[3], 10, 32) if err != nil { return nil, err } refcnt, err := strconv.ParseUint(fields[4], 10, 32) if err != nil { return nil, err } use, err := strconv.ParseUint(fields[5], 10, 32) if err != nil { return nil, err } metric, err := strconv.ParseUint(fields[6], 10, 32) if err != nil { return nil, err } mask, err := strconv.ParseUint(fields[7], 16, 32) if err != nil { return nil, err } mtu, err := strconv.ParseUint(fields[8], 10, 32) if err != nil { return nil, err } window, err := strconv.ParseUint(fields[9], 10, 32) if err != nil { return nil, err } irtt, err := strconv.ParseUint(fields[10], 10, 32) if err != nil { return nil, err } routeline := &NetRouteLine{ Iface: iface, Destination: uint32(destination), Gateway: uint32(gateway), Flags: uint32(flags), RefCnt: uint32(refcnt), Use: uint32(use), Metric: uint32(metric), Mask: uint32(mask), MTU: uint32(mtu), Window: uint32(window), IRTT: uint32(irtt), } return routeline, nil }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/net_tcp.go
vendor/github.com/prometheus/procfs/net_tcp.go
// Copyright 2020 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package procfs type ( // NetTCP represents the contents of /proc/net/tcp{,6} file without the header. NetTCP []*netIPSocketLine // NetTCPSummary provides already computed values like the total queue lengths or // the total number of used sockets. In contrast to NetTCP it does not collect // the parsed lines into a slice. NetTCPSummary NetIPSocketSummary ) // NetTCP returns the IPv4 kernel/networking statistics for TCP datagrams // read from /proc/net/tcp. func (fs FS) NetTCP() (NetTCP, error) { return newNetTCP(fs.proc.Path("net/tcp")) } // NetTCP6 returns the IPv6 kernel/networking statistics for TCP datagrams // read from /proc/net/tcp6. func (fs FS) NetTCP6() (NetTCP, error) { return newNetTCP(fs.proc.Path("net/tcp6")) } // NetTCPSummary returns already computed statistics like the total queue lengths // for TCP datagrams read from /proc/net/tcp. func (fs FS) NetTCPSummary() (*NetTCPSummary, error) { return newNetTCPSummary(fs.proc.Path("net/tcp")) } // NetTCP6Summary returns already computed statistics like the total queue lengths // for TCP datagrams read from /proc/net/tcp6. func (fs FS) NetTCP6Summary() (*NetTCPSummary, error) { return newNetTCPSummary(fs.proc.Path("net/tcp6")) } // newNetTCP creates a new NetTCP{,6} from the contents of the given file. func newNetTCP(file string) (NetTCP, error) { n, err := newNetIPSocket(file) n1 := NetTCP(n) return n1, err } func newNetTCPSummary(file string) (*NetTCPSummary, error) { n, err := newNetIPSocketSummary(file) if n == nil { return nil, err } n1 := NetTCPSummary(*n) return &n1, err }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/net_ip_socket.go
vendor/github.com/prometheus/procfs/net_ip_socket.go
// Copyright 2020 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package procfs import ( "bufio" "encoding/hex" "fmt" "io" "net" "os" "strconv" "strings" ) const ( // readLimit is used by io.LimitReader while reading the content of the // /proc/net/udp{,6} files. The number of lines inside such a file is dynamic // as each line represents a single used socket. // In theory, the number of available sockets is 65535 (2^16 - 1) per IP. // With e.g. 150 Byte per line and the maximum number of 65535, // the reader needs to handle 150 Byte * 65535 =~ 10 MB for a single IP. readLimit = 4294967296 // Byte -> 4 GiB ) // This contains generic data structures for both udp and tcp sockets. type ( // NetIPSocket represents the contents of /proc/net/{t,u}dp{,6} file without the header. NetIPSocket []*netIPSocketLine // NetIPSocketSummary provides already computed values like the total queue lengths or // the total number of used sockets. In contrast to NetIPSocket it does not collect // the parsed lines into a slice. NetIPSocketSummary struct { // TxQueueLength shows the total queue length of all parsed tx_queue lengths. TxQueueLength uint64 // RxQueueLength shows the total queue length of all parsed rx_queue lengths. RxQueueLength uint64 // UsedSockets shows the total number of parsed lines representing the // number of used sockets. UsedSockets uint64 // Drops shows the total number of dropped packets of all UPD sockets. Drops *uint64 } // netIPSocketLine represents the fields parsed from a single line // in /proc/net/{t,u}dp{,6}. Fields which are not used by IPSocket are skipped. // Drops is non-nil for udp{,6}, but nil for tcp{,6}. // For the proc file format details, see https://linux.die.net/man/5/proc. netIPSocketLine struct { Sl uint64 LocalAddr net.IP LocalPort uint64 RemAddr net.IP RemPort uint64 St uint64 TxQueue uint64 RxQueue uint64 UID uint64 Inode uint64 Drops *uint64 } ) func newNetIPSocket(file string) (NetIPSocket, error) { f, err := os.Open(file) if err != nil { return nil, err } defer f.Close() var netIPSocket NetIPSocket isUDP := strings.Contains(file, "udp") lr := io.LimitReader(f, readLimit) s := bufio.NewScanner(lr) s.Scan() // skip first line with headers for s.Scan() { fields := strings.Fields(s.Text()) line, err := parseNetIPSocketLine(fields, isUDP) if err != nil { return nil, err } netIPSocket = append(netIPSocket, line) } if err := s.Err(); err != nil { return nil, err } return netIPSocket, nil } // newNetIPSocketSummary creates a new NetIPSocket{,6} from the contents of the given file. func newNetIPSocketSummary(file string) (*NetIPSocketSummary, error) { f, err := os.Open(file) if err != nil { return nil, err } defer f.Close() var netIPSocketSummary NetIPSocketSummary var udpPacketDrops uint64 isUDP := strings.Contains(file, "udp") lr := io.LimitReader(f, readLimit) s := bufio.NewScanner(lr) s.Scan() // skip first line with headers for s.Scan() { fields := strings.Fields(s.Text()) line, err := parseNetIPSocketLine(fields, isUDP) if err != nil { return nil, err } netIPSocketSummary.TxQueueLength += line.TxQueue netIPSocketSummary.RxQueueLength += line.RxQueue netIPSocketSummary.UsedSockets++ if isUDP { udpPacketDrops += *line.Drops netIPSocketSummary.Drops = &udpPacketDrops } } if err := s.Err(); err != nil { return nil, err } return &netIPSocketSummary, nil } // the /proc/net/{t,u}dp{,6} files are network byte order for ipv4 and for ipv6 the address is four words consisting of four bytes each. In each of those four words the four bytes are written in reverse order. func parseIP(hexIP string) (net.IP, error) { var byteIP []byte byteIP, err := hex.DecodeString(hexIP) if err != nil { return nil, fmt.Errorf("%w: Cannot parse socket field in %q: %w", ErrFileParse, hexIP, err) } switch len(byteIP) { case 4: return net.IP{byteIP[3], byteIP[2], byteIP[1], byteIP[0]}, nil case 16: i := net.IP{ byteIP[3], byteIP[2], byteIP[1], byteIP[0], byteIP[7], byteIP[6], byteIP[5], byteIP[4], byteIP[11], byteIP[10], byteIP[9], byteIP[8], byteIP[15], byteIP[14], byteIP[13], byteIP[12], } return i, nil default: return nil, fmt.Errorf("%w: Unable to parse IP %s: %v", ErrFileParse, hexIP, nil) } } // parseNetIPSocketLine parses a single line, represented by a list of fields. func parseNetIPSocketLine(fields []string, isUDP bool) (*netIPSocketLine, error) { line := &netIPSocketLine{} if len(fields) < 10 { return nil, fmt.Errorf( "%w: Less than 10 columns found %q", ErrFileParse, strings.Join(fields, " "), ) } var err error // parse error // sl s := strings.Split(fields[0], ":") if len(s) != 2 { return nil, fmt.Errorf("%w: Unable to parse sl field in line %q", ErrFileParse, fields[0]) } if line.Sl, err = strconv.ParseUint(s[0], 0, 64); err != nil { return nil, fmt.Errorf("%w: Unable to parse sl field in %q: %w", ErrFileParse, line.Sl, err) } // local_address l := strings.Split(fields[1], ":") if len(l) != 2 { return nil, fmt.Errorf("%w: Unable to parse local_address field in %q", ErrFileParse, fields[1]) } if line.LocalAddr, err = parseIP(l[0]); err != nil { return nil, err } if line.LocalPort, err = strconv.ParseUint(l[1], 16, 64); err != nil { return nil, fmt.Errorf("%w: Unable to parse local_address port value line %q: %w", ErrFileParse, line.LocalPort, err) } // remote_address r := strings.Split(fields[2], ":") if len(r) != 2 { return nil, fmt.Errorf("%w: Unable to parse rem_address field in %q", ErrFileParse, fields[1]) } if line.RemAddr, err = parseIP(r[0]); err != nil { return nil, err } if line.RemPort, err = strconv.ParseUint(r[1], 16, 64); err != nil { return nil, fmt.Errorf("%w: Cannot parse rem_address port value in %q: %w", ErrFileParse, line.RemPort, err) } // st if line.St, err = strconv.ParseUint(fields[3], 16, 64); err != nil { return nil, fmt.Errorf("%w: Cannot parse st value in %q: %w", ErrFileParse, line.St, err) } // tx_queue and rx_queue q := strings.Split(fields[4], ":") if len(q) != 2 { return nil, fmt.Errorf( "%w: Missing colon for tx/rx queues in socket line %q", ErrFileParse, fields[4], ) } if line.TxQueue, err = strconv.ParseUint(q[0], 16, 64); err != nil { return nil, fmt.Errorf("%w: Cannot parse tx_queue value in %q: %w", ErrFileParse, line.TxQueue, err) } if line.RxQueue, err = strconv.ParseUint(q[1], 16, 64); err != nil { return nil, fmt.Errorf("%w: Cannot parse trx_queue value in %q: %w", ErrFileParse, line.RxQueue, err) } // uid if line.UID, err = strconv.ParseUint(fields[7], 0, 64); err != nil { return nil, fmt.Errorf("%w: Cannot parse UID value in %q: %w", ErrFileParse, line.UID, err) } // inode if line.Inode, err = strconv.ParseUint(fields[9], 0, 64); err != nil { return nil, fmt.Errorf("%w: Cannot parse inode value in %q: %w", ErrFileParse, line.Inode, err) } // drops if isUDP { drops, err := strconv.ParseUint(fields[12], 0, 64) if err != nil { return nil, fmt.Errorf("%w: Cannot parse drops value in %q: %w", ErrFileParse, drops, err) } line.Drops = &drops } return line, nil }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/cpuinfo_ppcx.go
vendor/github.com/prometheus/procfs/cpuinfo_ppcx.go
// Copyright 2020 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. //go:build linux && (ppc64 || ppc64le) // +build linux // +build ppc64 ppc64le package procfs var parseCPUInfo = parseCPUInfoPPC
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/arp.go
vendor/github.com/prometheus/procfs/arp.go
// Copyright 2019 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package procfs import ( "fmt" "net" "os" "strconv" "strings" ) // Learned from include/uapi/linux/if_arp.h. const ( // completed entry (ha valid). ATFComplete = 0x02 // permanent entry. ATFPermanent = 0x04 // Publish entry. ATFPublish = 0x08 // Has requested trailers. ATFUseTrailers = 0x10 // Obsoleted: Want to use a netmask (only for proxy entries). ATFNetmask = 0x20 // Don't answer this addresses. ATFDontPublish = 0x40 ) // ARPEntry contains a single row of the columnar data represented in // /proc/net/arp. type ARPEntry struct { // IP address IPAddr net.IP // MAC address HWAddr net.HardwareAddr // Name of the device Device string // Flags Flags byte } // GatherARPEntries retrieves all the ARP entries, parse the relevant columns, // and then return a slice of ARPEntry's. func (fs FS) GatherARPEntries() ([]ARPEntry, error) { data, err := os.ReadFile(fs.proc.Path("net/arp")) if err != nil { return nil, fmt.Errorf("%w: error reading arp %s: %w", ErrFileRead, fs.proc.Path("net/arp"), err) } return parseARPEntries(data) } func parseARPEntries(data []byte) ([]ARPEntry, error) { lines := strings.Split(string(data), "\n") entries := make([]ARPEntry, 0) var err error const ( expectedDataWidth = 6 expectedHeaderWidth = 9 ) for _, line := range lines { columns := strings.Fields(line) width := len(columns) if width == expectedHeaderWidth || width == 0 { continue } else if width == expectedDataWidth { entry, err := parseARPEntry(columns) if err != nil { return []ARPEntry{}, fmt.Errorf("%w: Failed to parse ARP entry: %v: %w", ErrFileParse, entry, err) } entries = append(entries, entry) } else { return []ARPEntry{}, fmt.Errorf("%w: %d columns found, but expected %d: %w", ErrFileParse, width, expectedDataWidth, err) } } return entries, err } func parseARPEntry(columns []string) (ARPEntry, error) { entry := ARPEntry{Device: columns[5]} ip := net.ParseIP(columns[0]) entry.IPAddr = ip if mac, err := net.ParseMAC(columns[3]); err == nil { entry.HWAddr = mac } else { return ARPEntry{}, err } if flags, err := strconv.ParseUint(columns[2], 0, 8); err == nil { entry.Flags = byte(flags) } else { return ARPEntry{}, err } return entry, nil } // IsComplete returns true if ARP entry is marked with complete flag. func (entry *ARPEntry) IsComplete() bool { return entry.Flags&ATFComplete != 0 }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/proc_cgroups.go
vendor/github.com/prometheus/procfs/proc_cgroups.go
// Copyright 2021 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package procfs import ( "bufio" "bytes" "fmt" "strconv" "strings" "github.com/prometheus/procfs/internal/util" ) // CgroupSummary models one line from /proc/cgroups. // This file contains information about the controllers that are compiled into the kernel. // // Also see http://man7.org/linux/man-pages/man7/cgroups.7.html type CgroupSummary struct { // The name of the controller. controller is also known as subsystem. SubsysName string // The unique ID of the cgroup hierarchy on which this controller is mounted. Hierarchy int // The number of control groups in this hierarchy using this controller. Cgroups int // This field contains the value 1 if this controller is enabled, or 0 if it has been disabled Enabled int } // parseCgroupSummary parses each line of the /proc/cgroup file // Line format is `subsys_name hierarchy num_cgroups enabled`. func parseCgroupSummaryString(CgroupSummaryStr string) (*CgroupSummary, error) { var err error fields := strings.Fields(CgroupSummaryStr) // require at least 4 fields if len(fields) < 4 { return nil, fmt.Errorf("%w: 4+ fields required, found %d fields in cgroup info string: %s", ErrFileParse, len(fields), CgroupSummaryStr) } CgroupSummary := &CgroupSummary{ SubsysName: fields[0], } CgroupSummary.Hierarchy, err = strconv.Atoi(fields[1]) if err != nil { return nil, fmt.Errorf("%w: Unable to parse hierarchy ID from %q", ErrFileParse, fields[1]) } CgroupSummary.Cgroups, err = strconv.Atoi(fields[2]) if err != nil { return nil, fmt.Errorf("%w: Unable to parse Cgroup Num from %q", ErrFileParse, fields[2]) } CgroupSummary.Enabled, err = strconv.Atoi(fields[3]) if err != nil { return nil, fmt.Errorf("%w: Unable to parse Enabled from %q", ErrFileParse, fields[3]) } return CgroupSummary, nil } // parseCgroupSummary reads each line of the /proc/cgroup file. func parseCgroupSummary(data []byte) ([]CgroupSummary, error) { var CgroupSummarys []CgroupSummary scanner := bufio.NewScanner(bytes.NewReader(data)) for scanner.Scan() { CgroupSummaryString := scanner.Text() // ignore comment lines if strings.HasPrefix(CgroupSummaryString, "#") { continue } CgroupSummary, err := parseCgroupSummaryString(CgroupSummaryString) if err != nil { return nil, err } CgroupSummarys = append(CgroupSummarys, *CgroupSummary) } err := scanner.Err() return CgroupSummarys, err } // CgroupSummarys returns information about current /proc/cgroups. func (fs FS) CgroupSummarys() ([]CgroupSummary, error) { data, err := util.ReadFileNoStat(fs.proc.Path("cgroups")) if err != nil { return nil, err } return parseCgroupSummary(data) }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/cpuinfo_others.go
vendor/github.com/prometheus/procfs/cpuinfo_others.go
// Copyright 2020 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. //go:build linux && !386 && !amd64 && !arm && !arm64 && !loong64 && !mips && !mips64 && !mips64le && !mipsle && !ppc64 && !ppc64le && !riscv64 && !s390x // +build linux,!386,!amd64,!arm,!arm64,!loong64,!mips,!mips64,!mips64le,!mipsle,!ppc64,!ppc64le,!riscv64,!s390x package procfs var parseCPUInfo = parseCPUInfoDummy
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/net_sockstat.go
vendor/github.com/prometheus/procfs/net_sockstat.go
// Copyright 2019 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package procfs import ( "bufio" "bytes" "fmt" "io" "strings" "github.com/prometheus/procfs/internal/util" ) // A NetSockstat contains the output of /proc/net/sockstat{,6} for IPv4 or IPv6, // respectively. type NetSockstat struct { // Used is non-nil for IPv4 sockstat results, but nil for IPv6. Used *int Protocols []NetSockstatProtocol } // A NetSockstatProtocol contains statistics about a given socket protocol. // Pointer fields indicate that the value may or may not be present on any // given protocol. type NetSockstatProtocol struct { Protocol string InUse int Orphan *int TW *int Alloc *int Mem *int Memory *int } // NetSockstat retrieves IPv4 socket statistics. func (fs FS) NetSockstat() (*NetSockstat, error) { return readSockstat(fs.proc.Path("net", "sockstat")) } // NetSockstat6 retrieves IPv6 socket statistics. // // If IPv6 is disabled on this kernel, the returned error can be checked with // os.IsNotExist. func (fs FS) NetSockstat6() (*NetSockstat, error) { return readSockstat(fs.proc.Path("net", "sockstat6")) } // readSockstat opens and parses a NetSockstat from the input file. func readSockstat(name string) (*NetSockstat, error) { // This file is small and can be read with one syscall. b, err := util.ReadFileNoStat(name) if err != nil { // Do not wrap this error so the caller can detect os.IsNotExist and // similar conditions. return nil, err } stat, err := parseSockstat(bytes.NewReader(b)) if err != nil { return nil, fmt.Errorf("%w: sockstats from %q: %w", ErrFileRead, name, err) } return stat, nil } // parseSockstat reads the contents of a sockstat file and parses a NetSockstat. func parseSockstat(r io.Reader) (*NetSockstat, error) { var stat NetSockstat s := bufio.NewScanner(r) for s.Scan() { // Expect a minimum of a protocol and one key/value pair. fields := strings.Split(s.Text(), " ") if len(fields) < 3 { return nil, fmt.Errorf("%w: Malformed sockstat line: %q", ErrFileParse, s.Text()) } // The remaining fields are key/value pairs. kvs, err := parseSockstatKVs(fields[1:]) if err != nil { return nil, fmt.Errorf("%w: sockstat key/value pairs from %q: %w", ErrFileParse, s.Text(), err) } // The first field is the protocol. We must trim its colon suffix. proto := strings.TrimSuffix(fields[0], ":") switch proto { case "sockets": // Special case: IPv4 has a sockets "used" key/value pair that we // embed at the top level of the structure. used := kvs["used"] stat.Used = &used default: // Parse all other lines as individual protocols. nsp := parseSockstatProtocol(kvs) nsp.Protocol = proto stat.Protocols = append(stat.Protocols, nsp) } } if err := s.Err(); err != nil { return nil, err } return &stat, nil } // parseSockstatKVs parses a string slice into a map of key/value pairs. func parseSockstatKVs(kvs []string) (map[string]int, error) { if len(kvs)%2 != 0 { return nil, fmt.Errorf("%w:: Odd number of fields in key/value pairs %q", ErrFileParse, kvs) } // Iterate two values at a time to gather key/value pairs. out := make(map[string]int, len(kvs)/2) for i := 0; i < len(kvs); i += 2 { vp := util.NewValueParser(kvs[i+1]) out[kvs[i]] = vp.Int() if err := vp.Err(); err != nil { return nil, err } } return out, nil } // parseSockstatProtocol parses a NetSockstatProtocol from the input kvs map. func parseSockstatProtocol(kvs map[string]int) NetSockstatProtocol { var nsp NetSockstatProtocol for k, v := range kvs { // Capture the range variable to ensure we get unique pointers for // each of the optional fields. v := v switch k { case "inuse": nsp.InUse = v case "orphan": nsp.Orphan = &v case "tw": nsp.TW = &v case "alloc": nsp.Alloc = &v case "mem": nsp.Mem = &v case "memory": nsp.Memory = &v } } return nsp }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/fs_statfs_type.go
vendor/github.com/prometheus/procfs/fs_statfs_type.go
// Copyright 2018 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. //go:build freebsd || linux // +build freebsd linux package procfs import ( "syscall" ) // isRealProc determines whether supplied mountpoint is really a proc filesystem. func isRealProc(mountPoint string) (bool, error) { stat := syscall.Statfs_t{} err := syscall.Statfs(mountPoint, &stat) if err != nil { return false, err } // 0x9fa0 is PROC_SUPER_MAGIC: https://elixir.bootlin.com/linux/v6.1/source/include/uapi/linux/magic.h#L87 return stat.Type == 0x9fa0, nil }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/cpuinfo_armx.go
vendor/github.com/prometheus/procfs/cpuinfo_armx.go
// Copyright 2020 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. //go:build linux && (arm || arm64) // +build linux // +build arm arm64 package procfs var parseCPUInfo = parseCPUInfoARM
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/kernel_random.go
vendor/github.com/prometheus/procfs/kernel_random.go
// Copyright 2020 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. //go:build !windows // +build !windows package procfs import ( "os" "github.com/prometheus/procfs/internal/util" ) // KernelRandom contains information about to the kernel's random number generator. type KernelRandom struct { // EntropyAvaliable gives the available entropy, in bits. EntropyAvaliable *uint64 // PoolSize gives the size of the entropy pool, in bits. PoolSize *uint64 // URandomMinReseedSeconds is the number of seconds after which the DRNG will be reseeded. URandomMinReseedSeconds *uint64 // WriteWakeupThreshold the number of bits of entropy below which we wake up processes // that do a select(2) or poll(2) for write access to /dev/random. WriteWakeupThreshold *uint64 // ReadWakeupThreshold is the number of bits of entropy required for waking up processes that sleep // waiting for entropy from /dev/random. ReadWakeupThreshold *uint64 } // KernelRandom returns values from /proc/sys/kernel/random. func (fs FS) KernelRandom() (KernelRandom, error) { random := KernelRandom{} for file, p := range map[string]**uint64{ "entropy_avail": &random.EntropyAvaliable, "poolsize": &random.PoolSize, "urandom_min_reseed_secs": &random.URandomMinReseedSeconds, "write_wakeup_threshold": &random.WriteWakeupThreshold, "read_wakeup_threshold": &random.ReadWakeupThreshold, } { val, err := util.ReadUintFromFile(fs.proc.Path("sys", "kernel", "random", file)) if os.IsNotExist(err) { continue } if err != nil { return random, err } *p = &val } return random, nil }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/net_conntrackstat.go
vendor/github.com/prometheus/procfs/net_conntrackstat.go
// Copyright 2020 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package procfs import ( "bufio" "bytes" "fmt" "io" "strings" "github.com/prometheus/procfs/internal/util" ) // A ConntrackStatEntry represents one line from net/stat/nf_conntrack // and contains netfilter conntrack statistics at one CPU core. type ConntrackStatEntry struct { Entries uint64 Searched uint64 Found uint64 New uint64 Invalid uint64 Ignore uint64 Delete uint64 DeleteList uint64 Insert uint64 InsertFailed uint64 Drop uint64 EarlyDrop uint64 SearchRestart uint64 } // ConntrackStat retrieves netfilter's conntrack statistics, split by CPU cores. func (fs FS) ConntrackStat() ([]ConntrackStatEntry, error) { return readConntrackStat(fs.proc.Path("net", "stat", "nf_conntrack")) } // Parses a slice of ConntrackStatEntries from the given filepath. func readConntrackStat(path string) ([]ConntrackStatEntry, error) { // This file is small and can be read with one syscall. b, err := util.ReadFileNoStat(path) if err != nil { // Do not wrap this error so the caller can detect os.IsNotExist and // similar conditions. return nil, err } stat, err := parseConntrackStat(bytes.NewReader(b)) if err != nil { return nil, fmt.Errorf("%w: Cannot read file: %v: %w", ErrFileRead, path, err) } return stat, nil } // Reads the contents of a conntrack statistics file and parses a slice of ConntrackStatEntries. func parseConntrackStat(r io.Reader) ([]ConntrackStatEntry, error) { var entries []ConntrackStatEntry scanner := bufio.NewScanner(r) scanner.Scan() for scanner.Scan() { fields := strings.Fields(scanner.Text()) conntrackEntry, err := parseConntrackStatEntry(fields) if err != nil { return nil, err } entries = append(entries, *conntrackEntry) } return entries, nil } // Parses a ConntrackStatEntry from given array of fields. func parseConntrackStatEntry(fields []string) (*ConntrackStatEntry, error) { entries, err := util.ParseHexUint64s(fields) if err != nil { return nil, fmt.Errorf("%w: Cannot parse entry: %d: %w", ErrFileParse, entries, err) } numEntries := len(entries) if numEntries < 16 || numEntries > 17 { return nil, fmt.Errorf("%w: invalid conntrackstat entry, invalid number of fields: %d", ErrFileParse, numEntries) } stats := &ConntrackStatEntry{ Entries: *entries[0], Searched: *entries[1], Found: *entries[2], New: *entries[3], Invalid: *entries[4], Ignore: *entries[5], Delete: *entries[6], DeleteList: *entries[7], Insert: *entries[8], InsertFailed: *entries[9], Drop: *entries[10], EarlyDrop: *entries[11], } // Ignore missing search_restart on Linux < 2.6.35. if numEntries == 17 { stats.SearchRestart = *entries[16] } return stats, nil }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/proc_interrupts.go
vendor/github.com/prometheus/procfs/proc_interrupts.go
// Copyright 2022 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package procfs import ( "bufio" "bytes" "errors" "fmt" "io" "strconv" "strings" "github.com/prometheus/procfs/internal/util" ) // Interrupt represents a single interrupt line. type Interrupt struct { // Info is the type of interrupt. Info string // Devices is the name of the device that is located at that IRQ Devices string // Values is the number of interrupts per CPU. Values []string } // Interrupts models the content of /proc/interrupts. Key is the IRQ number. // - https://access.redhat.com/documentation/en-us/red_hat_enterprise_linux/6/html/deployment_guide/s2-proc-interrupts // - https://raspberrypi.stackexchange.com/questions/105802/explanation-of-proc-interrupts-output type Interrupts map[string]Interrupt // Interrupts creates a new instance from a given Proc instance. func (p Proc) Interrupts() (Interrupts, error) { data, err := util.ReadFileNoStat(p.path("interrupts")) if err != nil { return nil, err } return parseInterrupts(bytes.NewReader(data)) } func parseInterrupts(r io.Reader) (Interrupts, error) { var ( interrupts = Interrupts{} scanner = bufio.NewScanner(r) ) if !scanner.Scan() { return nil, errors.New("interrupts empty") } cpuNum := len(strings.Fields(scanner.Text())) // one header per cpu for scanner.Scan() { parts := strings.Fields(scanner.Text()) if len(parts) == 0 { // skip empty lines continue } if len(parts) < 2 { return nil, fmt.Errorf("%w: Not enough fields in interrupts (expected 2+ fields but got %d): %s", ErrFileParse, len(parts), parts) } intName := parts[0][:len(parts[0])-1] // remove trailing : if len(parts) == 2 { interrupts[intName] = Interrupt{ Info: "", Devices: "", Values: []string{ parts[1], }, } continue } intr := Interrupt{ Values: parts[1 : cpuNum+1], } if _, err := strconv.Atoi(intName); err == nil { // numeral interrupt intr.Info = parts[cpuNum+1] intr.Devices = strings.Join(parts[cpuNum+2:], " ") } else { intr.Info = strings.Join(parts[cpuNum+1:], " ") } interrupts[intName] = intr } return interrupts, scanner.Err() }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/crypto.go
vendor/github.com/prometheus/procfs/crypto.go
// Copyright 2019 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package procfs import ( "bufio" "bytes" "fmt" "io" "strings" "github.com/prometheus/procfs/internal/util" ) // Crypto holds info parsed from /proc/crypto. type Crypto struct { Alignmask *uint64 Async bool Blocksize *uint64 Chunksize *uint64 Ctxsize *uint64 Digestsize *uint64 Driver string Geniv string Internal string Ivsize *uint64 Maxauthsize *uint64 MaxKeysize *uint64 MinKeysize *uint64 Module string Name string Priority *int64 Refcnt *int64 Seedsize *uint64 Selftest string Type string Walksize *uint64 } // Crypto parses an crypto-file (/proc/crypto) and returns a slice of // structs containing the relevant info. More information available here: // https://kernel.readthedocs.io/en/sphinx-samples/crypto-API.html func (fs FS) Crypto() ([]Crypto, error) { path := fs.proc.Path("crypto") b, err := util.ReadFileNoStat(path) if err != nil { return nil, fmt.Errorf("%w: Cannot read file %v: %w", ErrFileRead, b, err) } crypto, err := parseCrypto(bytes.NewReader(b)) if err != nil { return nil, fmt.Errorf("%w: Cannot parse %v: %w", ErrFileParse, crypto, err) } return crypto, nil } // parseCrypto parses a /proc/crypto stream into Crypto elements. func parseCrypto(r io.Reader) ([]Crypto, error) { var out []Crypto s := bufio.NewScanner(r) for s.Scan() { text := s.Text() switch { case strings.HasPrefix(text, "name"): // Each crypto element begins with its name. out = append(out, Crypto{}) case text == "": continue } kv := strings.Split(text, ":") if len(kv) != 2 { return nil, fmt.Errorf("%w: Cannot parse line: %q", ErrFileParse, text) } k := strings.TrimSpace(kv[0]) v := strings.TrimSpace(kv[1]) // Parse the key/value pair into the currently focused element. c := &out[len(out)-1] if err := c.parseKV(k, v); err != nil { return nil, err } } if err := s.Err(); err != nil { return nil, err } return out, nil } // parseKV parses a key/value pair into the appropriate field of c. func (c *Crypto) parseKV(k, v string) error { vp := util.NewValueParser(v) switch k { case "async": // Interpret literal yes as true. c.Async = v == "yes" case "blocksize": c.Blocksize = vp.PUInt64() case "chunksize": c.Chunksize = vp.PUInt64() case "digestsize": c.Digestsize = vp.PUInt64() case "driver": c.Driver = v case "geniv": c.Geniv = v case "internal": c.Internal = v case "ivsize": c.Ivsize = vp.PUInt64() case "maxauthsize": c.Maxauthsize = vp.PUInt64() case "max keysize": c.MaxKeysize = vp.PUInt64() case "min keysize": c.MinKeysize = vp.PUInt64() case "module": c.Module = v case "name": c.Name = v case "priority": c.Priority = vp.PInt64() case "refcnt": c.Refcnt = vp.PInt64() case "seedsize": c.Seedsize = vp.PUInt64() case "selftest": c.Selftest = v case "type": c.Type = v case "walksize": c.Walksize = vp.PUInt64() } return vp.Err() }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/netstat.go
vendor/github.com/prometheus/procfs/netstat.go
// Copyright 2020 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package procfs import ( "bufio" "os" "path/filepath" "strconv" "strings" ) // NetStat contains statistics for all the counters from one file. type NetStat struct { Stats map[string][]uint64 Filename string } // NetStat retrieves stats from `/proc/net/stat/`. func (fs FS) NetStat() ([]NetStat, error) { statFiles, err := filepath.Glob(fs.proc.Path("net/stat/*")) if err != nil { return nil, err } var netStatsTotal []NetStat for _, filePath := range statFiles { procNetstat, err := parseNetstat(filePath) if err != nil { return nil, err } procNetstat.Filename = filepath.Base(filePath) netStatsTotal = append(netStatsTotal, procNetstat) } return netStatsTotal, nil } // parseNetstat parses the metrics from `/proc/net/stat/` file // and returns a NetStat structure. func parseNetstat(filePath string) (NetStat, error) { netStat := NetStat{ Stats: make(map[string][]uint64), } file, err := os.Open(filePath) if err != nil { return netStat, err } defer file.Close() scanner := bufio.NewScanner(file) scanner.Scan() // First string is always a header for stats var headers []string headers = append(headers, strings.Fields(scanner.Text())...) // Other strings represent per-CPU counters for scanner.Scan() { for num, counter := range strings.Fields(scanner.Text()) { value, err := strconv.ParseUint(counter, 16, 64) if err != nil { return NetStat{}, err } netStat.Stats[headers[num]] = append(netStat.Stats[headers[num]], value) } } return netStat, nil }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/swaps.go
vendor/github.com/prometheus/procfs/swaps.go
// Copyright 2019 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package procfs import ( "bufio" "bytes" "fmt" "strconv" "strings" "github.com/prometheus/procfs/internal/util" ) // Swap represents an entry in /proc/swaps. type Swap struct { Filename string Type string Size int Used int Priority int } // Swaps returns a slice of all configured swap devices on the system. func (fs FS) Swaps() ([]*Swap, error) { data, err := util.ReadFileNoStat(fs.proc.Path("swaps")) if err != nil { return nil, err } return parseSwaps(data) } func parseSwaps(info []byte) ([]*Swap, error) { swaps := []*Swap{} scanner := bufio.NewScanner(bytes.NewReader(info)) scanner.Scan() // ignore header line for scanner.Scan() { swapString := scanner.Text() parsedSwap, err := parseSwapString(swapString) if err != nil { return nil, err } swaps = append(swaps, parsedSwap) } err := scanner.Err() return swaps, err } func parseSwapString(swapString string) (*Swap, error) { var err error swapFields := strings.Fields(swapString) swapLength := len(swapFields) if swapLength < 5 { return nil, fmt.Errorf("%w: too few fields in swap string: %s", ErrFileParse, swapString) } swap := &Swap{ Filename: swapFields[0], Type: swapFields[1], } swap.Size, err = strconv.Atoi(swapFields[2]) if err != nil { return nil, fmt.Errorf("%w: invalid swap size: %s: %w", ErrFileParse, swapFields[2], err) } swap.Used, err = strconv.Atoi(swapFields[3]) if err != nil { return nil, fmt.Errorf("%w: invalid swap used: %s: %w", ErrFileParse, swapFields[3], err) } swap.Priority, err = strconv.Atoi(swapFields[4]) if err != nil { return nil, fmt.Errorf("%w: invalid swap priority: %s: %w", ErrFileParse, swapFields[4], err) } return swap, nil }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/slab.go
vendor/github.com/prometheus/procfs/slab.go
// Copyright 2020 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package procfs import ( "bufio" "bytes" "fmt" "regexp" "strconv" "strings" "github.com/prometheus/procfs/internal/util" ) var ( slabSpace = regexp.MustCompile(`\s+`) slabVer = regexp.MustCompile(`slabinfo -`) slabHeader = regexp.MustCompile(`# name`) ) // Slab represents a slab pool in the kernel. type Slab struct { Name string ObjActive int64 ObjNum int64 ObjSize int64 ObjPerSlab int64 PagesPerSlab int64 // tunables Limit int64 Batch int64 SharedFactor int64 SlabActive int64 SlabNum int64 SharedAvail int64 } // SlabInfo represents info for all slabs. type SlabInfo struct { Slabs []*Slab } func shouldParseSlab(line string) bool { if slabVer.MatchString(line) { return false } if slabHeader.MatchString(line) { return false } return true } // parseV21SlabEntry is used to parse a line from /proc/slabinfo version 2.1. func parseV21SlabEntry(line string) (*Slab, error) { // First cleanup whitespace. l := slabSpace.ReplaceAllString(line, " ") s := strings.Split(l, " ") if len(s) != 16 { return nil, fmt.Errorf("%w: unable to parse: %q", ErrFileParse, line) } var err error i := &Slab{Name: s[0]} i.ObjActive, err = strconv.ParseInt(s[1], 10, 64) if err != nil { return nil, err } i.ObjNum, err = strconv.ParseInt(s[2], 10, 64) if err != nil { return nil, err } i.ObjSize, err = strconv.ParseInt(s[3], 10, 64) if err != nil { return nil, err } i.ObjPerSlab, err = strconv.ParseInt(s[4], 10, 64) if err != nil { return nil, err } i.PagesPerSlab, err = strconv.ParseInt(s[5], 10, 64) if err != nil { return nil, err } i.Limit, err = strconv.ParseInt(s[8], 10, 64) if err != nil { return nil, err } i.Batch, err = strconv.ParseInt(s[9], 10, 64) if err != nil { return nil, err } i.SharedFactor, err = strconv.ParseInt(s[10], 10, 64) if err != nil { return nil, err } i.SlabActive, err = strconv.ParseInt(s[13], 10, 64) if err != nil { return nil, err } i.SlabNum, err = strconv.ParseInt(s[14], 10, 64) if err != nil { return nil, err } i.SharedAvail, err = strconv.ParseInt(s[15], 10, 64) if err != nil { return nil, err } return i, nil } // parseSlabInfo21 is used to parse a slabinfo 2.1 file. func parseSlabInfo21(r *bytes.Reader) (SlabInfo, error) { scanner := bufio.NewScanner(r) s := SlabInfo{Slabs: []*Slab{}} for scanner.Scan() { line := scanner.Text() if !shouldParseSlab(line) { continue } slab, err := parseV21SlabEntry(line) if err != nil { return s, err } s.Slabs = append(s.Slabs, slab) } return s, nil } // SlabInfo reads data from `/proc/slabinfo`. func (fs FS) SlabInfo() (SlabInfo, error) { // TODO: Consider passing options to allow for parsing different // slabinfo versions. However, slabinfo 2.1 has been stable since // kernel 2.6.10 and later. data, err := util.ReadFileNoStat(fs.proc.Path("slabinfo")) if err != nil { return SlabInfo{}, err } return parseSlabInfo21(bytes.NewReader(data)) }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/fs_statfs_notype.go
vendor/github.com/prometheus/procfs/fs_statfs_notype.go
// Copyright 2018 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. //go:build !freebsd && !linux // +build !freebsd,!linux package procfs // isRealProc returns true on architectures that don't have a Type argument // in their Statfs_t struct func isRealProc(mountPoint string) (bool, error) { return true, nil }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/proc_snmp6.go
vendor/github.com/prometheus/procfs/proc_snmp6.go
// Copyright 2022 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package procfs import ( "bufio" "bytes" "errors" "io" "os" "strconv" "strings" "github.com/prometheus/procfs/internal/util" ) // ProcSnmp6 models the content of /proc/<pid>/net/snmp6. type ProcSnmp6 struct { // The process ID. PID int Ip6 Icmp6 Udp6 UdpLite6 } type Ip6 struct { // nolint:revive InReceives *float64 InHdrErrors *float64 InTooBigErrors *float64 InNoRoutes *float64 InAddrErrors *float64 InUnknownProtos *float64 InTruncatedPkts *float64 InDiscards *float64 InDelivers *float64 OutForwDatagrams *float64 OutRequests *float64 OutDiscards *float64 OutNoRoutes *float64 ReasmTimeout *float64 ReasmReqds *float64 ReasmOKs *float64 ReasmFails *float64 FragOKs *float64 FragFails *float64 FragCreates *float64 InMcastPkts *float64 OutMcastPkts *float64 InOctets *float64 OutOctets *float64 InMcastOctets *float64 OutMcastOctets *float64 InBcastOctets *float64 OutBcastOctets *float64 InNoECTPkts *float64 InECT1Pkts *float64 InECT0Pkts *float64 InCEPkts *float64 } type Icmp6 struct { InMsgs *float64 InErrors *float64 OutMsgs *float64 OutErrors *float64 InCsumErrors *float64 InDestUnreachs *float64 InPktTooBigs *float64 InTimeExcds *float64 InParmProblems *float64 InEchos *float64 InEchoReplies *float64 InGroupMembQueries *float64 InGroupMembResponses *float64 InGroupMembReductions *float64 InRouterSolicits *float64 InRouterAdvertisements *float64 InNeighborSolicits *float64 InNeighborAdvertisements *float64 InRedirects *float64 InMLDv2Reports *float64 OutDestUnreachs *float64 OutPktTooBigs *float64 OutTimeExcds *float64 OutParmProblems *float64 OutEchos *float64 OutEchoReplies *float64 OutGroupMembQueries *float64 OutGroupMembResponses *float64 OutGroupMembReductions *float64 OutRouterSolicits *float64 OutRouterAdvertisements *float64 OutNeighborSolicits *float64 OutNeighborAdvertisements *float64 OutRedirects *float64 OutMLDv2Reports *float64 InType1 *float64 InType134 *float64 InType135 *float64 InType136 *float64 InType143 *float64 OutType133 *float64 OutType135 *float64 OutType136 *float64 OutType143 *float64 } type Udp6 struct { // nolint:revive InDatagrams *float64 NoPorts *float64 InErrors *float64 OutDatagrams *float64 RcvbufErrors *float64 SndbufErrors *float64 InCsumErrors *float64 IgnoredMulti *float64 } type UdpLite6 struct { // nolint:revive InDatagrams *float64 NoPorts *float64 InErrors *float64 OutDatagrams *float64 RcvbufErrors *float64 SndbufErrors *float64 InCsumErrors *float64 } func (p Proc) Snmp6() (ProcSnmp6, error) { filename := p.path("net/snmp6") data, err := util.ReadFileNoStat(filename) if err != nil { // On systems with IPv6 disabled, this file won't exist. // Do nothing. if errors.Is(err, os.ErrNotExist) { return ProcSnmp6{PID: p.PID}, nil } return ProcSnmp6{PID: p.PID}, err } procSnmp6, err := parseSNMP6Stats(bytes.NewReader(data)) procSnmp6.PID = p.PID return procSnmp6, err } // parseSnmp6 parses the metrics from proc/<pid>/net/snmp6 file // and returns a map contains those metrics. func parseSNMP6Stats(r io.Reader) (ProcSnmp6, error) { var ( scanner = bufio.NewScanner(r) procSnmp6 = ProcSnmp6{} ) for scanner.Scan() { stat := strings.Fields(scanner.Text()) if len(stat) < 2 { continue } // Expect to have "6" in metric name, skip line otherwise if sixIndex := strings.Index(stat[0], "6"); sixIndex != -1 { protocol := stat[0][:sixIndex+1] key := stat[0][sixIndex+1:] value, err := strconv.ParseFloat(stat[1], 64) if err != nil { return procSnmp6, err } switch protocol { case "Ip6": switch key { case "InReceives": procSnmp6.Ip6.InReceives = &value case "InHdrErrors": procSnmp6.Ip6.InHdrErrors = &value case "InTooBigErrors": procSnmp6.Ip6.InTooBigErrors = &value case "InNoRoutes": procSnmp6.Ip6.InNoRoutes = &value case "InAddrErrors": procSnmp6.Ip6.InAddrErrors = &value case "InUnknownProtos": procSnmp6.Ip6.InUnknownProtos = &value case "InTruncatedPkts": procSnmp6.Ip6.InTruncatedPkts = &value case "InDiscards": procSnmp6.Ip6.InDiscards = &value case "InDelivers": procSnmp6.Ip6.InDelivers = &value case "OutForwDatagrams": procSnmp6.Ip6.OutForwDatagrams = &value case "OutRequests": procSnmp6.Ip6.OutRequests = &value case "OutDiscards": procSnmp6.Ip6.OutDiscards = &value case "OutNoRoutes": procSnmp6.Ip6.OutNoRoutes = &value case "ReasmTimeout": procSnmp6.Ip6.ReasmTimeout = &value case "ReasmReqds": procSnmp6.Ip6.ReasmReqds = &value case "ReasmOKs": procSnmp6.Ip6.ReasmOKs = &value case "ReasmFails": procSnmp6.Ip6.ReasmFails = &value case "FragOKs": procSnmp6.Ip6.FragOKs = &value case "FragFails": procSnmp6.Ip6.FragFails = &value case "FragCreates": procSnmp6.Ip6.FragCreates = &value case "InMcastPkts": procSnmp6.Ip6.InMcastPkts = &value case "OutMcastPkts": procSnmp6.Ip6.OutMcastPkts = &value case "InOctets": procSnmp6.Ip6.InOctets = &value case "OutOctets": procSnmp6.Ip6.OutOctets = &value case "InMcastOctets": procSnmp6.Ip6.InMcastOctets = &value case "OutMcastOctets": procSnmp6.Ip6.OutMcastOctets = &value case "InBcastOctets": procSnmp6.Ip6.InBcastOctets = &value case "OutBcastOctets": procSnmp6.Ip6.OutBcastOctets = &value case "InNoECTPkts": procSnmp6.Ip6.InNoECTPkts = &value case "InECT1Pkts": procSnmp6.Ip6.InECT1Pkts = &value case "InECT0Pkts": procSnmp6.Ip6.InECT0Pkts = &value case "InCEPkts": procSnmp6.Ip6.InCEPkts = &value } case "Icmp6": switch key { case "InMsgs": procSnmp6.Icmp6.InMsgs = &value case "InErrors": procSnmp6.Icmp6.InErrors = &value case "OutMsgs": procSnmp6.Icmp6.OutMsgs = &value case "OutErrors": procSnmp6.Icmp6.OutErrors = &value case "InCsumErrors": procSnmp6.Icmp6.InCsumErrors = &value case "InDestUnreachs": procSnmp6.Icmp6.InDestUnreachs = &value case "InPktTooBigs": procSnmp6.Icmp6.InPktTooBigs = &value case "InTimeExcds": procSnmp6.Icmp6.InTimeExcds = &value case "InParmProblems": procSnmp6.Icmp6.InParmProblems = &value case "InEchos": procSnmp6.Icmp6.InEchos = &value case "InEchoReplies": procSnmp6.Icmp6.InEchoReplies = &value case "InGroupMembQueries": procSnmp6.Icmp6.InGroupMembQueries = &value case "InGroupMembResponses": procSnmp6.Icmp6.InGroupMembResponses = &value case "InGroupMembReductions": procSnmp6.Icmp6.InGroupMembReductions = &value case "InRouterSolicits": procSnmp6.Icmp6.InRouterSolicits = &value case "InRouterAdvertisements": procSnmp6.Icmp6.InRouterAdvertisements = &value case "InNeighborSolicits": procSnmp6.Icmp6.InNeighborSolicits = &value case "InNeighborAdvertisements": procSnmp6.Icmp6.InNeighborAdvertisements = &value case "InRedirects": procSnmp6.Icmp6.InRedirects = &value case "InMLDv2Reports": procSnmp6.Icmp6.InMLDv2Reports = &value case "OutDestUnreachs": procSnmp6.Icmp6.OutDestUnreachs = &value case "OutPktTooBigs": procSnmp6.Icmp6.OutPktTooBigs = &value case "OutTimeExcds": procSnmp6.Icmp6.OutTimeExcds = &value case "OutParmProblems": procSnmp6.Icmp6.OutParmProblems = &value case "OutEchos": procSnmp6.Icmp6.OutEchos = &value case "OutEchoReplies": procSnmp6.Icmp6.OutEchoReplies = &value case "OutGroupMembQueries": procSnmp6.Icmp6.OutGroupMembQueries = &value case "OutGroupMembResponses": procSnmp6.Icmp6.OutGroupMembResponses = &value case "OutGroupMembReductions": procSnmp6.Icmp6.OutGroupMembReductions = &value case "OutRouterSolicits": procSnmp6.Icmp6.OutRouterSolicits = &value case "OutRouterAdvertisements": procSnmp6.Icmp6.OutRouterAdvertisements = &value case "OutNeighborSolicits": procSnmp6.Icmp6.OutNeighborSolicits = &value case "OutNeighborAdvertisements": procSnmp6.Icmp6.OutNeighborAdvertisements = &value case "OutRedirects": procSnmp6.Icmp6.OutRedirects = &value case "OutMLDv2Reports": procSnmp6.Icmp6.OutMLDv2Reports = &value case "InType1": procSnmp6.Icmp6.InType1 = &value case "InType134": procSnmp6.Icmp6.InType134 = &value case "InType135": procSnmp6.Icmp6.InType135 = &value case "InType136": procSnmp6.Icmp6.InType136 = &value case "InType143": procSnmp6.Icmp6.InType143 = &value case "OutType133": procSnmp6.Icmp6.OutType133 = &value case "OutType135": procSnmp6.Icmp6.OutType135 = &value case "OutType136": procSnmp6.Icmp6.OutType136 = &value case "OutType143": procSnmp6.Icmp6.OutType143 = &value } case "Udp6": switch key { case "InDatagrams": procSnmp6.Udp6.InDatagrams = &value case "NoPorts": procSnmp6.Udp6.NoPorts = &value case "InErrors": procSnmp6.Udp6.InErrors = &value case "OutDatagrams": procSnmp6.Udp6.OutDatagrams = &value case "RcvbufErrors": procSnmp6.Udp6.RcvbufErrors = &value case "SndbufErrors": procSnmp6.Udp6.SndbufErrors = &value case "InCsumErrors": procSnmp6.Udp6.InCsumErrors = &value case "IgnoredMulti": procSnmp6.Udp6.IgnoredMulti = &value } case "UdpLite6": switch key { case "InDatagrams": procSnmp6.UdpLite6.InDatagrams = &value case "NoPorts": procSnmp6.UdpLite6.NoPorts = &value case "InErrors": procSnmp6.UdpLite6.InErrors = &value case "OutDatagrams": procSnmp6.UdpLite6.OutDatagrams = &value case "RcvbufErrors": procSnmp6.UdpLite6.RcvbufErrors = &value case "SndbufErrors": procSnmp6.UdpLite6.SndbufErrors = &value case "InCsumErrors": procSnmp6.UdpLite6.InCsumErrors = &value } } } } return procSnmp6, scanner.Err() }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/ipvs.go
vendor/github.com/prometheus/procfs/ipvs.go
// Copyright 2018 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package procfs import ( "bufio" "bytes" "encoding/hex" "errors" "fmt" "io" "net" "os" "strconv" "strings" "github.com/prometheus/procfs/internal/util" ) // IPVSStats holds IPVS statistics, as exposed by the kernel in `/proc/net/ip_vs_stats`. type IPVSStats struct { // Total count of connections. Connections uint64 // Total incoming packages processed. IncomingPackets uint64 // Total outgoing packages processed. OutgoingPackets uint64 // Total incoming traffic. IncomingBytes uint64 // Total outgoing traffic. OutgoingBytes uint64 } // IPVSBackendStatus holds current metrics of one virtual / real address pair. type IPVSBackendStatus struct { // The local (virtual) IP address. LocalAddress net.IP // The remote (real) IP address. RemoteAddress net.IP // The local (virtual) port. LocalPort uint16 // The remote (real) port. RemotePort uint16 // The local firewall mark LocalMark string // The transport protocol (TCP, UDP). Proto string // The current number of active connections for this virtual/real address pair. ActiveConn uint64 // The current number of inactive connections for this virtual/real address pair. InactConn uint64 // The current weight of this virtual/real address pair. Weight uint64 } // IPVSStats reads the IPVS statistics from the specified `proc` filesystem. func (fs FS) IPVSStats() (IPVSStats, error) { data, err := util.ReadFileNoStat(fs.proc.Path("net/ip_vs_stats")) if err != nil { return IPVSStats{}, err } return parseIPVSStats(bytes.NewReader(data)) } // parseIPVSStats performs the actual parsing of `ip_vs_stats`. func parseIPVSStats(r io.Reader) (IPVSStats, error) { var ( statContent []byte statLines []string statFields []string stats IPVSStats ) statContent, err := io.ReadAll(r) if err != nil { return IPVSStats{}, err } statLines = strings.SplitN(string(statContent), "\n", 4) if len(statLines) != 4 { return IPVSStats{}, errors.New("ip_vs_stats corrupt: too short") } statFields = strings.Fields(statLines[2]) if len(statFields) != 5 { return IPVSStats{}, errors.New("ip_vs_stats corrupt: unexpected number of fields") } stats.Connections, err = strconv.ParseUint(statFields[0], 16, 64) if err != nil { return IPVSStats{}, err } stats.IncomingPackets, err = strconv.ParseUint(statFields[1], 16, 64) if err != nil { return IPVSStats{}, err } stats.OutgoingPackets, err = strconv.ParseUint(statFields[2], 16, 64) if err != nil { return IPVSStats{}, err } stats.IncomingBytes, err = strconv.ParseUint(statFields[3], 16, 64) if err != nil { return IPVSStats{}, err } stats.OutgoingBytes, err = strconv.ParseUint(statFields[4], 16, 64) if err != nil { return IPVSStats{}, err } return stats, nil } // IPVSBackendStatus reads and returns the status of all (virtual,real) server pairs from the specified `proc` filesystem. func (fs FS) IPVSBackendStatus() ([]IPVSBackendStatus, error) { file, err := os.Open(fs.proc.Path("net/ip_vs")) if err != nil { return nil, err } defer file.Close() return parseIPVSBackendStatus(file) } func parseIPVSBackendStatus(file io.Reader) ([]IPVSBackendStatus, error) { var ( status []IPVSBackendStatus scanner = bufio.NewScanner(file) proto string localMark string localAddress net.IP localPort uint16 err error ) for scanner.Scan() { fields := strings.Fields(scanner.Text()) if len(fields) == 0 { continue } switch { case fields[0] == "IP" || fields[0] == "Prot" || fields[1] == "RemoteAddress:Port": continue case fields[0] == "TCP" || fields[0] == "UDP": if len(fields) < 2 { continue } proto = fields[0] localMark = "" localAddress, localPort, err = parseIPPort(fields[1]) if err != nil { return nil, err } case fields[0] == "FWM": if len(fields) < 2 { continue } proto = fields[0] localMark = fields[1] localAddress = nil localPort = 0 case fields[0] == "->": if len(fields) < 6 { continue } remoteAddress, remotePort, err := parseIPPort(fields[1]) if err != nil { return nil, err } weight, err := strconv.ParseUint(fields[3], 10, 64) if err != nil { return nil, err } activeConn, err := strconv.ParseUint(fields[4], 10, 64) if err != nil { return nil, err } inactConn, err := strconv.ParseUint(fields[5], 10, 64) if err != nil { return nil, err } status = append(status, IPVSBackendStatus{ LocalAddress: localAddress, LocalPort: localPort, LocalMark: localMark, RemoteAddress: remoteAddress, RemotePort: remotePort, Proto: proto, Weight: weight, ActiveConn: activeConn, InactConn: inactConn, }) } } return status, nil } func parseIPPort(s string) (net.IP, uint16, error) { var ( ip net.IP err error ) switch len(s) { case 13: ip, err = hex.DecodeString(s[0:8]) if err != nil { return nil, 0, err } case 46: ip = net.ParseIP(s[1:40]) if ip == nil { return nil, 0, fmt.Errorf("%w: Invalid IPv6 addr %s: %w", ErrFileParse, s[1:40], err) } default: return nil, 0, fmt.Errorf("%w: Unexpected IP:Port %s: %w", ErrFileParse, s, err) } portString := s[len(s)-4:] if len(portString) != 4 { return nil, 0, fmt.Errorf("%w: Unexpected port string format %s: %w", ErrFileParse, portString, err) } port, err := strconv.ParseUint(portString, 16, 16) if err != nil { return nil, 0, err } return ip, uint16(port), nil }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/net_wireless.go
vendor/github.com/prometheus/procfs/net_wireless.go
// Copyright 2023 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package procfs import ( "bufio" "bytes" "fmt" "io" "strconv" "strings" "github.com/prometheus/procfs/internal/util" ) // Wireless models the content of /proc/net/wireless. type Wireless struct { Name string // Status is the current 4-digit hex value status of the interface. Status uint64 // QualityLink is the link quality. QualityLink int // QualityLevel is the signal gain (dBm). QualityLevel int // QualityNoise is the signal noise baseline (dBm). QualityNoise int // DiscardedNwid is the number of discarded packets with wrong nwid/essid. DiscardedNwid int // DiscardedCrypt is the number of discarded packets with wrong code/decode (WEP). DiscardedCrypt int // DiscardedFrag is the number of discarded packets that can't perform MAC reassembly. DiscardedFrag int // DiscardedRetry is the number of discarded packets that reached max MAC retries. DiscardedRetry int // DiscardedMisc is the number of discarded packets for other reasons. DiscardedMisc int // MissedBeacon is the number of missed beacons/superframe. MissedBeacon int } // Wireless returns kernel wireless statistics. func (fs FS) Wireless() ([]*Wireless, error) { b, err := util.ReadFileNoStat(fs.proc.Path("net/wireless")) if err != nil { return nil, err } m, err := parseWireless(bytes.NewReader(b)) if err != nil { return nil, fmt.Errorf("%w: wireless: %w", ErrFileParse, err) } return m, nil } // parseWireless parses the contents of /proc/net/wireless. /* Inter-| sta-| Quality | Discarded packets | Missed | WE face | tus | link level noise | nwid crypt frag retry misc | beacon | 22 eth1: 0000 5. -256. -10. 0 1 0 3 0 0 eth2: 0000 5. -256. -20. 0 2 0 4 0 0 */ func parseWireless(r io.Reader) ([]*Wireless, error) { var ( interfaces []*Wireless scanner = bufio.NewScanner(r) ) for n := 0; scanner.Scan(); n++ { // Skip the 2 header lines. if n < 2 { continue } line := scanner.Text() parts := strings.Split(line, ":") if len(parts) != 2 { return nil, fmt.Errorf("%w: expected 2 parts after splitting line by ':', got %d for line %q", ErrFileParse, len(parts), line) } name := strings.TrimSpace(parts[0]) stats := strings.Fields(parts[1]) if len(stats) < 10 { return nil, fmt.Errorf("%w: invalid number of fields in line %d, expected 10+, got %d: %q", ErrFileParse, n, len(stats), line) } status, err := strconv.ParseUint(stats[0], 16, 16) if err != nil { return nil, fmt.Errorf("%w: invalid status in line %d: %q", ErrFileParse, n, line) } qlink, err := strconv.Atoi(strings.TrimSuffix(stats[1], ".")) if err != nil { return nil, fmt.Errorf("%w: parse Quality:link as integer %q: %w", ErrFileParse, qlink, err) } qlevel, err := strconv.Atoi(strings.TrimSuffix(stats[2], ".")) if err != nil { return nil, fmt.Errorf("%w: Quality:level as integer %q: %w", ErrFileParse, qlevel, err) } qnoise, err := strconv.Atoi(strings.TrimSuffix(stats[3], ".")) if err != nil { return nil, fmt.Errorf("%w: Quality:noise as integer %q: %w", ErrFileParse, qnoise, err) } dnwid, err := strconv.Atoi(stats[4]) if err != nil { return nil, fmt.Errorf("%w: Discarded:nwid as integer %q: %w", ErrFileParse, dnwid, err) } dcrypt, err := strconv.Atoi(stats[5]) if err != nil { return nil, fmt.Errorf("%w: Discarded:crypt as integer %q: %w", ErrFileParse, dcrypt, err) } dfrag, err := strconv.Atoi(stats[6]) if err != nil { return nil, fmt.Errorf("%w: Discarded:frag as integer %q: %w", ErrFileParse, dfrag, err) } dretry, err := strconv.Atoi(stats[7]) if err != nil { return nil, fmt.Errorf("%w: Discarded:retry as integer %q: %w", ErrFileParse, dretry, err) } dmisc, err := strconv.Atoi(stats[8]) if err != nil { return nil, fmt.Errorf("%w: Discarded:misc as integer %q: %w", ErrFileParse, dmisc, err) } mbeacon, err := strconv.Atoi(stats[9]) if err != nil { return nil, fmt.Errorf("%w: Missed:beacon as integer %q: %w", ErrFileParse, mbeacon, err) } w := &Wireless{ Name: name, Status: status, QualityLink: qlink, QualityLevel: qlevel, QualityNoise: qnoise, DiscardedNwid: dnwid, DiscardedCrypt: dcrypt, DiscardedFrag: dfrag, DiscardedRetry: dretry, DiscardedMisc: dmisc, MissedBeacon: mbeacon, } interfaces = append(interfaces, w) } if err := scanner.Err(); err != nil { return nil, fmt.Errorf("%w: Failed to scan /proc/net/wireless: %w", ErrFileRead, err) } return interfaces, nil }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/fscache.go
vendor/github.com/prometheus/procfs/fscache.go
// Copyright 2019 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package procfs import ( "bufio" "bytes" "fmt" "io" "strconv" "strings" "github.com/prometheus/procfs/internal/util" ) // Fscacheinfo represents fscache statistics. type Fscacheinfo struct { // Number of index cookies allocated IndexCookiesAllocated uint64 // data storage cookies allocated DataStorageCookiesAllocated uint64 // Number of special cookies allocated SpecialCookiesAllocated uint64 // Number of objects allocated ObjectsAllocated uint64 // Number of object allocation failures ObjectAllocationsFailure uint64 // Number of objects that reached the available state ObjectsAvailable uint64 // Number of objects that reached the dead state ObjectsDead uint64 // Number of objects that didn't have a coherency check ObjectsWithoutCoherencyCheck uint64 // Number of objects that passed a coherency check ObjectsWithCoherencyCheck uint64 // Number of objects that needed a coherency data update ObjectsNeedCoherencyCheckUpdate uint64 // Number of objects that were declared obsolete ObjectsDeclaredObsolete uint64 // Number of pages marked as being cached PagesMarkedAsBeingCached uint64 // Number of uncache page requests seen UncachePagesRequestSeen uint64 // Number of acquire cookie requests seen AcquireCookiesRequestSeen uint64 // Number of acq reqs given a NULL parent AcquireRequestsWithNullParent uint64 // Number of acq reqs rejected due to no cache available AcquireRequestsRejectedNoCacheAvailable uint64 // Number of acq reqs succeeded AcquireRequestsSucceeded uint64 // Number of acq reqs rejected due to error AcquireRequestsRejectedDueToError uint64 // Number of acq reqs failed on ENOMEM AcquireRequestsFailedDueToEnomem uint64 // Number of lookup calls made on cache backends LookupsNumber uint64 // Number of negative lookups made LookupsNegative uint64 // Number of positive lookups made LookupsPositive uint64 // Number of objects created by lookup ObjectsCreatedByLookup uint64 // Number of lookups timed out and requeued LookupsTimedOutAndRequed uint64 InvalidationsNumber uint64 InvalidationsRunning uint64 // Number of update cookie requests seen UpdateCookieRequestSeen uint64 // Number of upd reqs given a NULL parent UpdateRequestsWithNullParent uint64 // Number of upd reqs granted CPU time UpdateRequestsRunning uint64 // Number of relinquish cookie requests seen RelinquishCookiesRequestSeen uint64 // Number of rlq reqs given a NULL parent RelinquishCookiesWithNullParent uint64 // Number of rlq reqs waited on completion of creation RelinquishRequestsWaitingCompleteCreation uint64 // Relinqs rtr RelinquishRetries uint64 // Number of attribute changed requests seen AttributeChangedRequestsSeen uint64 // Number of attr changed requests queued AttributeChangedRequestsQueued uint64 // Number of attr changed rejected -ENOBUFS AttributeChangedRejectDueToEnobufs uint64 // Number of attr changed failed -ENOMEM AttributeChangedFailedDueToEnomem uint64 // Number of attr changed ops given CPU time AttributeChangedOps uint64 // Number of allocation requests seen AllocationRequestsSeen uint64 // Number of successful alloc reqs AllocationOkRequests uint64 // Number of alloc reqs that waited on lookup completion AllocationWaitingOnLookup uint64 // Number of alloc reqs rejected -ENOBUFS AllocationsRejectedDueToEnobufs uint64 // Number of alloc reqs aborted -ERESTARTSYS AllocationsAbortedDueToErestartsys uint64 // Number of alloc reqs submitted AllocationOperationsSubmitted uint64 // Number of alloc reqs waited for CPU time AllocationsWaitedForCPU uint64 // Number of alloc reqs aborted due to object death AllocationsAbortedDueToObjectDeath uint64 // Number of retrieval (read) requests seen RetrievalsReadRequests uint64 // Number of successful retr reqs RetrievalsOk uint64 // Number of retr reqs that waited on lookup completion RetrievalsWaitingLookupCompletion uint64 // Number of retr reqs returned -ENODATA RetrievalsReturnedEnodata uint64 // Number of retr reqs rejected -ENOBUFS RetrievalsRejectedDueToEnobufs uint64 // Number of retr reqs aborted -ERESTARTSYS RetrievalsAbortedDueToErestartsys uint64 // Number of retr reqs failed -ENOMEM RetrievalsFailedDueToEnomem uint64 // Number of retr reqs submitted RetrievalsRequests uint64 // Number of retr reqs waited for CPU time RetrievalsWaitingCPU uint64 // Number of retr reqs aborted due to object death RetrievalsAbortedDueToObjectDeath uint64 // Number of storage (write) requests seen StoreWriteRequests uint64 // Number of successful store reqs StoreSuccessfulRequests uint64 // Number of store reqs on a page already pending storage StoreRequestsOnPendingStorage uint64 // Number of store reqs rejected -ENOBUFS StoreRequestsRejectedDueToEnobufs uint64 // Number of store reqs failed -ENOMEM StoreRequestsFailedDueToEnomem uint64 // Number of store reqs submitted StoreRequestsSubmitted uint64 // Number of store reqs granted CPU time StoreRequestsRunning uint64 // Number of pages given store req processing time StorePagesWithRequestsProcessing uint64 // Number of store reqs deleted from tracking tree StoreRequestsDeleted uint64 // Number of store reqs over store limit StoreRequestsOverStoreLimit uint64 // Number of release reqs against pages with no pending store ReleaseRequestsAgainstPagesWithNoPendingStorage uint64 // Number of release reqs against pages stored by time lock granted ReleaseRequestsAgainstPagesStoredByTimeLockGranted uint64 // Number of release reqs ignored due to in-progress store ReleaseRequestsIgnoredDueToInProgressStore uint64 // Number of page stores cancelled due to release req PageStoresCancelledByReleaseRequests uint64 VmscanWaiting uint64 // Number of times async ops added to pending queues OpsPending uint64 // Number of times async ops given CPU time OpsRunning uint64 // Number of times async ops queued for processing OpsEnqueued uint64 // Number of async ops cancelled OpsCancelled uint64 // Number of async ops rejected due to object lookup/create failure OpsRejected uint64 // Number of async ops initialised OpsInitialised uint64 // Number of async ops queued for deferred release OpsDeferred uint64 // Number of async ops released (should equal ini=N when idle) OpsReleased uint64 // Number of deferred-release async ops garbage collected OpsGarbageCollected uint64 // Number of in-progress alloc_object() cache ops CacheopAllocationsinProgress uint64 // Number of in-progress lookup_object() cache ops CacheopLookupObjectInProgress uint64 // Number of in-progress lookup_complete() cache ops CacheopLookupCompleteInPorgress uint64 // Number of in-progress grab_object() cache ops CacheopGrabObjectInProgress uint64 CacheopInvalidations uint64 // Number of in-progress update_object() cache ops CacheopUpdateObjectInProgress uint64 // Number of in-progress drop_object() cache ops CacheopDropObjectInProgress uint64 // Number of in-progress put_object() cache ops CacheopPutObjectInProgress uint64 // Number of in-progress attr_changed() cache ops CacheopAttributeChangeInProgress uint64 // Number of in-progress sync_cache() cache ops CacheopSyncCacheInProgress uint64 // Number of in-progress read_or_alloc_page() cache ops CacheopReadOrAllocPageInProgress uint64 // Number of in-progress read_or_alloc_pages() cache ops CacheopReadOrAllocPagesInProgress uint64 // Number of in-progress allocate_page() cache ops CacheopAllocatePageInProgress uint64 // Number of in-progress allocate_pages() cache ops CacheopAllocatePagesInProgress uint64 // Number of in-progress write_page() cache ops CacheopWritePagesInProgress uint64 // Number of in-progress uncache_page() cache ops CacheopUncachePagesInProgress uint64 // Number of in-progress dissociate_pages() cache ops CacheopDissociatePagesInProgress uint64 // Number of object lookups/creations rejected due to lack of space CacheevLookupsAndCreationsRejectedLackSpace uint64 // Number of stale objects deleted CacheevStaleObjectsDeleted uint64 // Number of objects retired when relinquished CacheevRetiredWhenReliquished uint64 // Number of objects culled CacheevObjectsCulled uint64 } // Fscacheinfo returns information about current fscache statistics. // See https://www.kernel.org/doc/Documentation/filesystems/caching/fscache.txt func (fs FS) Fscacheinfo() (Fscacheinfo, error) { b, err := util.ReadFileNoStat(fs.proc.Path("fs/fscache/stats")) if err != nil { return Fscacheinfo{}, err } m, err := parseFscacheinfo(bytes.NewReader(b)) if err != nil { return Fscacheinfo{}, fmt.Errorf("%w: Cannot parse %v: %w", ErrFileParse, m, err) } return *m, nil } func setFSCacheFields(fields []string, setFields ...*uint64) error { var err error if len(fields) < len(setFields) { return fmt.Errorf("%w: Expected %d, but got %d: %w", ErrFileParse, len(setFields), len(fields), err) } for i := range setFields { *setFields[i], err = strconv.ParseUint(strings.Split(fields[i], "=")[1], 0, 64) if err != nil { return err } } return nil } func parseFscacheinfo(r io.Reader) (*Fscacheinfo, error) { var m Fscacheinfo s := bufio.NewScanner(r) for s.Scan() { fields := strings.Fields(s.Text()) if len(fields) < 2 { return nil, fmt.Errorf("%w: malformed Fscacheinfo line: %q", ErrFileParse, s.Text()) } switch fields[0] { case "Cookies:": err := setFSCacheFields(fields[1:], &m.IndexCookiesAllocated, &m.DataStorageCookiesAllocated, &m.SpecialCookiesAllocated) if err != nil { return &m, err } case "Objects:": err := setFSCacheFields(fields[1:], &m.ObjectsAllocated, &m.ObjectAllocationsFailure, &m.ObjectsAvailable, &m.ObjectsDead) if err != nil { return &m, err } case "ChkAux": err := setFSCacheFields(fields[2:], &m.ObjectsWithoutCoherencyCheck, &m.ObjectsWithCoherencyCheck, &m.ObjectsNeedCoherencyCheckUpdate, &m.ObjectsDeclaredObsolete) if err != nil { return &m, err } case "Pages": err := setFSCacheFields(fields[2:], &m.PagesMarkedAsBeingCached, &m.UncachePagesRequestSeen) if err != nil { return &m, err } case "Acquire:": err := setFSCacheFields(fields[1:], &m.AcquireCookiesRequestSeen, &m.AcquireRequestsWithNullParent, &m.AcquireRequestsRejectedNoCacheAvailable, &m.AcquireRequestsSucceeded, &m.AcquireRequestsRejectedDueToError, &m.AcquireRequestsFailedDueToEnomem) if err != nil { return &m, err } case "Lookups:": err := setFSCacheFields(fields[1:], &m.LookupsNumber, &m.LookupsNegative, &m.LookupsPositive, &m.ObjectsCreatedByLookup, &m.LookupsTimedOutAndRequed) if err != nil { return &m, err } case "Invals": err := setFSCacheFields(fields[2:], &m.InvalidationsNumber, &m.InvalidationsRunning) if err != nil { return &m, err } case "Updates:": err := setFSCacheFields(fields[1:], &m.UpdateCookieRequestSeen, &m.UpdateRequestsWithNullParent, &m.UpdateRequestsRunning) if err != nil { return &m, err } case "Relinqs:": err := setFSCacheFields(fields[1:], &m.RelinquishCookiesRequestSeen, &m.RelinquishCookiesWithNullParent, &m.RelinquishRequestsWaitingCompleteCreation, &m.RelinquishRetries) if err != nil { return &m, err } case "AttrChg:": err := setFSCacheFields(fields[1:], &m.AttributeChangedRequestsSeen, &m.AttributeChangedRequestsQueued, &m.AttributeChangedRejectDueToEnobufs, &m.AttributeChangedFailedDueToEnomem, &m.AttributeChangedOps) if err != nil { return &m, err } case "Allocs": if strings.Split(fields[2], "=")[0] == "n" { err := setFSCacheFields(fields[2:], &m.AllocationRequestsSeen, &m.AllocationOkRequests, &m.AllocationWaitingOnLookup, &m.AllocationsRejectedDueToEnobufs, &m.AllocationsAbortedDueToErestartsys) if err != nil { return &m, err } } else { err := setFSCacheFields(fields[2:], &m.AllocationOperationsSubmitted, &m.AllocationsWaitedForCPU, &m.AllocationsAbortedDueToObjectDeath) if err != nil { return &m, err } } case "Retrvls:": if strings.Split(fields[1], "=")[0] == "n" { err := setFSCacheFields(fields[1:], &m.RetrievalsReadRequests, &m.RetrievalsOk, &m.RetrievalsWaitingLookupCompletion, &m.RetrievalsReturnedEnodata, &m.RetrievalsRejectedDueToEnobufs, &m.RetrievalsAbortedDueToErestartsys, &m.RetrievalsFailedDueToEnomem) if err != nil { return &m, err } } else { err := setFSCacheFields(fields[1:], &m.RetrievalsRequests, &m.RetrievalsWaitingCPU, &m.RetrievalsAbortedDueToObjectDeath) if err != nil { return &m, err } } case "Stores": if strings.Split(fields[2], "=")[0] == "n" { err := setFSCacheFields(fields[2:], &m.StoreWriteRequests, &m.StoreSuccessfulRequests, &m.StoreRequestsOnPendingStorage, &m.StoreRequestsRejectedDueToEnobufs, &m.StoreRequestsFailedDueToEnomem) if err != nil { return &m, err } } else { err := setFSCacheFields(fields[2:], &m.StoreRequestsSubmitted, &m.StoreRequestsRunning, &m.StorePagesWithRequestsProcessing, &m.StoreRequestsDeleted, &m.StoreRequestsOverStoreLimit) if err != nil { return &m, err } } case "VmScan": err := setFSCacheFields(fields[2:], &m.ReleaseRequestsAgainstPagesWithNoPendingStorage, &m.ReleaseRequestsAgainstPagesStoredByTimeLockGranted, &m.ReleaseRequestsIgnoredDueToInProgressStore, &m.PageStoresCancelledByReleaseRequests, &m.VmscanWaiting) if err != nil { return &m, err } case "Ops": if strings.Split(fields[2], "=")[0] == "pend" { err := setFSCacheFields(fields[2:], &m.OpsPending, &m.OpsRunning, &m.OpsEnqueued, &m.OpsCancelled, &m.OpsRejected) if err != nil { return &m, err } } else { err := setFSCacheFields(fields[2:], &m.OpsInitialised, &m.OpsDeferred, &m.OpsReleased, &m.OpsGarbageCollected) if err != nil { return &m, err } } case "CacheOp:": if strings.Split(fields[1], "=")[0] == "alo" { err := setFSCacheFields(fields[1:], &m.CacheopAllocationsinProgress, &m.CacheopLookupObjectInProgress, &m.CacheopLookupCompleteInPorgress, &m.CacheopGrabObjectInProgress) if err != nil { return &m, err } } else if strings.Split(fields[1], "=")[0] == "inv" { err := setFSCacheFields(fields[1:], &m.CacheopInvalidations, &m.CacheopUpdateObjectInProgress, &m.CacheopDropObjectInProgress, &m.CacheopPutObjectInProgress, &m.CacheopAttributeChangeInProgress, &m.CacheopSyncCacheInProgress) if err != nil { return &m, err } } else { err := setFSCacheFields(fields[1:], &m.CacheopReadOrAllocPageInProgress, &m.CacheopReadOrAllocPagesInProgress, &m.CacheopAllocatePageInProgress, &m.CacheopAllocatePagesInProgress, &m.CacheopWritePagesInProgress, &m.CacheopUncachePagesInProgress, &m.CacheopDissociatePagesInProgress) if err != nil { return &m, err } } case "CacheEv:": err := setFSCacheFields(fields[1:], &m.CacheevLookupsAndCreationsRejectedLackSpace, &m.CacheevStaleObjectsDeleted, &m.CacheevRetiredWhenReliquished, &m.CacheevObjectsCulled) if err != nil { return &m, err } } } return &m, nil }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/cpuinfo_x86.go
vendor/github.com/prometheus/procfs/cpuinfo_x86.go
// Copyright 2020 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. //go:build linux && (386 || amd64) // +build linux // +build 386 amd64 package procfs var parseCPUInfo = parseCPUInfoX86
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/proc_snmp.go
vendor/github.com/prometheus/procfs/proc_snmp.go
// Copyright 2022 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package procfs import ( "bufio" "bytes" "fmt" "io" "strconv" "strings" "github.com/prometheus/procfs/internal/util" ) // ProcSnmp models the content of /proc/<pid>/net/snmp. type ProcSnmp struct { // The process ID. PID int Ip Icmp IcmpMsg Tcp Udp UdpLite } type Ip struct { // nolint:revive Forwarding *float64 DefaultTTL *float64 InReceives *float64 InHdrErrors *float64 InAddrErrors *float64 ForwDatagrams *float64 InUnknownProtos *float64 InDiscards *float64 InDelivers *float64 OutRequests *float64 OutDiscards *float64 OutNoRoutes *float64 ReasmTimeout *float64 ReasmReqds *float64 ReasmOKs *float64 ReasmFails *float64 FragOKs *float64 FragFails *float64 FragCreates *float64 } type Icmp struct { // nolint:revive InMsgs *float64 InErrors *float64 InCsumErrors *float64 InDestUnreachs *float64 InTimeExcds *float64 InParmProbs *float64 InSrcQuenchs *float64 InRedirects *float64 InEchos *float64 InEchoReps *float64 InTimestamps *float64 InTimestampReps *float64 InAddrMasks *float64 InAddrMaskReps *float64 OutMsgs *float64 OutErrors *float64 OutDestUnreachs *float64 OutTimeExcds *float64 OutParmProbs *float64 OutSrcQuenchs *float64 OutRedirects *float64 OutEchos *float64 OutEchoReps *float64 OutTimestamps *float64 OutTimestampReps *float64 OutAddrMasks *float64 OutAddrMaskReps *float64 } type IcmpMsg struct { InType3 *float64 OutType3 *float64 } type Tcp struct { // nolint:revive RtoAlgorithm *float64 RtoMin *float64 RtoMax *float64 MaxConn *float64 ActiveOpens *float64 PassiveOpens *float64 AttemptFails *float64 EstabResets *float64 CurrEstab *float64 InSegs *float64 OutSegs *float64 RetransSegs *float64 InErrs *float64 OutRsts *float64 InCsumErrors *float64 } type Udp struct { // nolint:revive InDatagrams *float64 NoPorts *float64 InErrors *float64 OutDatagrams *float64 RcvbufErrors *float64 SndbufErrors *float64 InCsumErrors *float64 IgnoredMulti *float64 } type UdpLite struct { // nolint:revive InDatagrams *float64 NoPorts *float64 InErrors *float64 OutDatagrams *float64 RcvbufErrors *float64 SndbufErrors *float64 InCsumErrors *float64 IgnoredMulti *float64 } func (p Proc) Snmp() (ProcSnmp, error) { filename := p.path("net/snmp") data, err := util.ReadFileNoStat(filename) if err != nil { return ProcSnmp{PID: p.PID}, err } procSnmp, err := parseSnmp(bytes.NewReader(data), filename) procSnmp.PID = p.PID return procSnmp, err } // parseSnmp parses the metrics from proc/<pid>/net/snmp file // and returns a map contains those metrics (e.g. {"Ip": {"Forwarding": 2}}). func parseSnmp(r io.Reader, fileName string) (ProcSnmp, error) { var ( scanner = bufio.NewScanner(r) procSnmp = ProcSnmp{} ) for scanner.Scan() { nameParts := strings.Split(scanner.Text(), " ") scanner.Scan() valueParts := strings.Split(scanner.Text(), " ") // Remove trailing :. protocol := strings.TrimSuffix(nameParts[0], ":") if len(nameParts) != len(valueParts) { return procSnmp, fmt.Errorf("%w: mismatch field count mismatch in %s: %s", ErrFileParse, fileName, protocol) } for i := 1; i < len(nameParts); i++ { value, err := strconv.ParseFloat(valueParts[i], 64) if err != nil { return procSnmp, err } key := nameParts[i] switch protocol { case "Ip": switch key { case "Forwarding": procSnmp.Ip.Forwarding = &value case "DefaultTTL": procSnmp.Ip.DefaultTTL = &value case "InReceives": procSnmp.Ip.InReceives = &value case "InHdrErrors": procSnmp.Ip.InHdrErrors = &value case "InAddrErrors": procSnmp.Ip.InAddrErrors = &value case "ForwDatagrams": procSnmp.Ip.ForwDatagrams = &value case "InUnknownProtos": procSnmp.Ip.InUnknownProtos = &value case "InDiscards": procSnmp.Ip.InDiscards = &value case "InDelivers": procSnmp.Ip.InDelivers = &value case "OutRequests": procSnmp.Ip.OutRequests = &value case "OutDiscards": procSnmp.Ip.OutDiscards = &value case "OutNoRoutes": procSnmp.Ip.OutNoRoutes = &value case "ReasmTimeout": procSnmp.Ip.ReasmTimeout = &value case "ReasmReqds": procSnmp.Ip.ReasmReqds = &value case "ReasmOKs": procSnmp.Ip.ReasmOKs = &value case "ReasmFails": procSnmp.Ip.ReasmFails = &value case "FragOKs": procSnmp.Ip.FragOKs = &value case "FragFails": procSnmp.Ip.FragFails = &value case "FragCreates": procSnmp.Ip.FragCreates = &value } case "Icmp": switch key { case "InMsgs": procSnmp.Icmp.InMsgs = &value case "InErrors": procSnmp.Icmp.InErrors = &value case "InCsumErrors": procSnmp.Icmp.InCsumErrors = &value case "InDestUnreachs": procSnmp.Icmp.InDestUnreachs = &value case "InTimeExcds": procSnmp.Icmp.InTimeExcds = &value case "InParmProbs": procSnmp.Icmp.InParmProbs = &value case "InSrcQuenchs": procSnmp.Icmp.InSrcQuenchs = &value case "InRedirects": procSnmp.Icmp.InRedirects = &value case "InEchos": procSnmp.Icmp.InEchos = &value case "InEchoReps": procSnmp.Icmp.InEchoReps = &value case "InTimestamps": procSnmp.Icmp.InTimestamps = &value case "InTimestampReps": procSnmp.Icmp.InTimestampReps = &value case "InAddrMasks": procSnmp.Icmp.InAddrMasks = &value case "InAddrMaskReps": procSnmp.Icmp.InAddrMaskReps = &value case "OutMsgs": procSnmp.Icmp.OutMsgs = &value case "OutErrors": procSnmp.Icmp.OutErrors = &value case "OutDestUnreachs": procSnmp.Icmp.OutDestUnreachs = &value case "OutTimeExcds": procSnmp.Icmp.OutTimeExcds = &value case "OutParmProbs": procSnmp.Icmp.OutParmProbs = &value case "OutSrcQuenchs": procSnmp.Icmp.OutSrcQuenchs = &value case "OutRedirects": procSnmp.Icmp.OutRedirects = &value case "OutEchos": procSnmp.Icmp.OutEchos = &value case "OutEchoReps": procSnmp.Icmp.OutEchoReps = &value case "OutTimestamps": procSnmp.Icmp.OutTimestamps = &value case "OutTimestampReps": procSnmp.Icmp.OutTimestampReps = &value case "OutAddrMasks": procSnmp.Icmp.OutAddrMasks = &value case "OutAddrMaskReps": procSnmp.Icmp.OutAddrMaskReps = &value } case "IcmpMsg": switch key { case "InType3": procSnmp.IcmpMsg.InType3 = &value case "OutType3": procSnmp.IcmpMsg.OutType3 = &value } case "Tcp": switch key { case "RtoAlgorithm": procSnmp.Tcp.RtoAlgorithm = &value case "RtoMin": procSnmp.Tcp.RtoMin = &value case "RtoMax": procSnmp.Tcp.RtoMax = &value case "MaxConn": procSnmp.Tcp.MaxConn = &value case "ActiveOpens": procSnmp.Tcp.ActiveOpens = &value case "PassiveOpens": procSnmp.Tcp.PassiveOpens = &value case "AttemptFails": procSnmp.Tcp.AttemptFails = &value case "EstabResets": procSnmp.Tcp.EstabResets = &value case "CurrEstab": procSnmp.Tcp.CurrEstab = &value case "InSegs": procSnmp.Tcp.InSegs = &value case "OutSegs": procSnmp.Tcp.OutSegs = &value case "RetransSegs": procSnmp.Tcp.RetransSegs = &value case "InErrs": procSnmp.Tcp.InErrs = &value case "OutRsts": procSnmp.Tcp.OutRsts = &value case "InCsumErrors": procSnmp.Tcp.InCsumErrors = &value } case "Udp": switch key { case "InDatagrams": procSnmp.Udp.InDatagrams = &value case "NoPorts": procSnmp.Udp.NoPorts = &value case "InErrors": procSnmp.Udp.InErrors = &value case "OutDatagrams": procSnmp.Udp.OutDatagrams = &value case "RcvbufErrors": procSnmp.Udp.RcvbufErrors = &value case "SndbufErrors": procSnmp.Udp.SndbufErrors = &value case "InCsumErrors": procSnmp.Udp.InCsumErrors = &value case "IgnoredMulti": procSnmp.Udp.IgnoredMulti = &value } case "UdpLite": switch key { case "InDatagrams": procSnmp.UdpLite.InDatagrams = &value case "NoPorts": procSnmp.UdpLite.NoPorts = &value case "InErrors": procSnmp.UdpLite.InErrors = &value case "OutDatagrams": procSnmp.UdpLite.OutDatagrams = &value case "RcvbufErrors": procSnmp.UdpLite.RcvbufErrors = &value case "SndbufErrors": procSnmp.UdpLite.SndbufErrors = &value case "InCsumErrors": procSnmp.UdpLite.InCsumErrors = &value case "IgnoredMulti": procSnmp.UdpLite.IgnoredMulti = &value } } } } return procSnmp, scanner.Err() }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/doc.go
vendor/github.com/prometheus/procfs/doc.go
// Copyright 2014 Prometheus Team // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. // Package procfs provides functions to retrieve system, kernel and process // metrics from the pseudo-filesystem proc. // // Example: // // package main // // import ( // "fmt" // "log" // // "github.com/prometheus/procfs" // ) // // func main() { // p, err := procfs.Self() // if err != nil { // log.Fatalf("could not get process: %s", err) // } // // stat, err := p.Stat() // if err != nil { // log.Fatalf("could not get process stat: %s", err) // } // // fmt.Printf("command: %s\n", stat.Comm) // fmt.Printf("cpu time: %fs\n", stat.CPUTime()) // fmt.Printf("vsize: %dB\n", stat.VirtualMemory()) // fmt.Printf("rss: %dB\n", stat.ResidentMemory()) // } package procfs
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/net_udp.go
vendor/github.com/prometheus/procfs/net_udp.go
// Copyright 2020 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package procfs type ( // NetUDP represents the contents of /proc/net/udp{,6} file without the header. NetUDP []*netIPSocketLine // NetUDPSummary provides already computed values like the total queue lengths or // the total number of used sockets. In contrast to NetUDP it does not collect // the parsed lines into a slice. NetUDPSummary NetIPSocketSummary ) // NetUDP returns the IPv4 kernel/networking statistics for UDP datagrams // read from /proc/net/udp. func (fs FS) NetUDP() (NetUDP, error) { return newNetUDP(fs.proc.Path("net/udp")) } // NetUDP6 returns the IPv6 kernel/networking statistics for UDP datagrams // read from /proc/net/udp6. func (fs FS) NetUDP6() (NetUDP, error) { return newNetUDP(fs.proc.Path("net/udp6")) } // NetUDPSummary returns already computed statistics like the total queue lengths // for UDP datagrams read from /proc/net/udp. func (fs FS) NetUDPSummary() (*NetUDPSummary, error) { return newNetUDPSummary(fs.proc.Path("net/udp")) } // NetUDP6Summary returns already computed statistics like the total queue lengths // for UDP datagrams read from /proc/net/udp6. func (fs FS) NetUDP6Summary() (*NetUDPSummary, error) { return newNetUDPSummary(fs.proc.Path("net/udp6")) } // newNetUDP creates a new NetUDP{,6} from the contents of the given file. func newNetUDP(file string) (NetUDP, error) { n, err := newNetIPSocket(file) n1 := NetUDP(n) return n1, err } func newNetUDPSummary(file string) (*NetUDPSummary, error) { n, err := newNetIPSocketSummary(file) if n == nil { return nil, err } n1 := NetUDPSummary(*n) return &n1, err }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/net_unix.go
vendor/github.com/prometheus/procfs/net_unix.go
// Copyright 2018 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package procfs import ( "bufio" "fmt" "io" "os" "strconv" "strings" ) // For the proc file format details, // see https://elixir.bootlin.com/linux/v4.17/source/net/unix/af_unix.c#L2815 // and https://elixir.bootlin.com/linux/latest/source/include/uapi/linux/net.h#L48. // Constants for the various /proc/net/unix enumerations. // TODO: match against x/sys/unix or similar? const ( netUnixTypeStream = 1 netUnixTypeDgram = 2 netUnixTypeSeqpacket = 5 netUnixFlagDefault = 0 netUnixFlagListen = 1 << 16 netUnixStateUnconnected = 1 netUnixStateConnecting = 2 netUnixStateConnected = 3 netUnixStateDisconnected = 4 ) // NetUNIXType is the type of the type field. type NetUNIXType uint64 // NetUNIXFlags is the type of the flags field. type NetUNIXFlags uint64 // NetUNIXState is the type of the state field. type NetUNIXState uint64 // NetUNIXLine represents a line of /proc/net/unix. type NetUNIXLine struct { KernelPtr string RefCount uint64 Protocol uint64 Flags NetUNIXFlags Type NetUNIXType State NetUNIXState Inode uint64 Path string } // NetUNIX holds the data read from /proc/net/unix. type NetUNIX struct { Rows []*NetUNIXLine } // NetUNIX returns data read from /proc/net/unix. func (fs FS) NetUNIX() (*NetUNIX, error) { return readNetUNIX(fs.proc.Path("net/unix")) } // readNetUNIX reads data in /proc/net/unix format from the specified file. func readNetUNIX(file string) (*NetUNIX, error) { // This file could be quite large and a streaming read is desirable versus // reading the entire contents at once. f, err := os.Open(file) if err != nil { return nil, err } defer f.Close() return parseNetUNIX(f) } // parseNetUNIX creates a NetUnix structure from the incoming stream. func parseNetUNIX(r io.Reader) (*NetUNIX, error) { // Begin scanning by checking for the existence of Inode. s := bufio.NewScanner(r) s.Scan() // From the man page of proc(5), it does not contain an Inode field, // but in actually it exists. This code works for both cases. hasInode := strings.Contains(s.Text(), "Inode") // Expect a minimum number of fields, but Inode and Path are optional: // Num RefCount Protocol Flags Type St Inode Path minFields := 6 if hasInode { minFields++ } var nu NetUNIX for s.Scan() { line := s.Text() item, err := nu.parseLine(line, hasInode, minFields) if err != nil { return nil, fmt.Errorf("%w: /proc/net/unix encountered data %q: %w", ErrFileParse, line, err) } nu.Rows = append(nu.Rows, item) } if err := s.Err(); err != nil { return nil, fmt.Errorf("%w: /proc/net/unix encountered data: %w", ErrFileParse, err) } return &nu, nil } func (u *NetUNIX) parseLine(line string, hasInode bool, min int) (*NetUNIXLine, error) { fields := strings.Fields(line) l := len(fields) if l < min { return nil, fmt.Errorf("%w: expected at least %d fields but got %d", ErrFileParse, min, l) } // Field offsets are as follows: // Num RefCount Protocol Flags Type St Inode Path kernelPtr := strings.TrimSuffix(fields[0], ":") users, err := u.parseUsers(fields[1]) if err != nil { return nil, fmt.Errorf("%w: ref count %q: %w", ErrFileParse, fields[1], err) } flags, err := u.parseFlags(fields[3]) if err != nil { return nil, fmt.Errorf("%w: Unable to parse flags %q: %w", ErrFileParse, fields[3], err) } typ, err := u.parseType(fields[4]) if err != nil { return nil, fmt.Errorf("%w: Failed to parse type %q: %w", ErrFileParse, fields[4], err) } state, err := u.parseState(fields[5]) if err != nil { return nil, fmt.Errorf("%w: Failed to parse state %q: %w", ErrFileParse, fields[5], err) } var inode uint64 if hasInode { inode, err = u.parseInode(fields[6]) if err != nil { return nil, fmt.Errorf("%w failed to parse inode %q: %w", ErrFileParse, fields[6], err) } } n := &NetUNIXLine{ KernelPtr: kernelPtr, RefCount: users, Type: typ, Flags: flags, State: state, Inode: inode, } // Path field is optional. if l > min { // Path occurs at either index 6 or 7 depending on whether inode is // already present. pathIdx := 7 if !hasInode { pathIdx-- } n.Path = fields[pathIdx] } return n, nil } func (u NetUNIX) parseUsers(s string) (uint64, error) { return strconv.ParseUint(s, 16, 32) } func (u NetUNIX) parseType(s string) (NetUNIXType, error) { typ, err := strconv.ParseUint(s, 16, 16) if err != nil { return 0, err } return NetUNIXType(typ), nil } func (u NetUNIX) parseFlags(s string) (NetUNIXFlags, error) { flags, err := strconv.ParseUint(s, 16, 32) if err != nil { return 0, err } return NetUNIXFlags(flags), nil } func (u NetUNIX) parseState(s string) (NetUNIXState, error) { st, err := strconv.ParseInt(s, 16, 8) if err != nil { return 0, err } return NetUNIXState(st), nil } func (u NetUNIX) parseInode(s string) (uint64, error) { return strconv.ParseUint(s, 10, 64) } func (t NetUNIXType) String() string { switch t { case netUnixTypeStream: return "stream" case netUnixTypeDgram: return "dgram" case netUnixTypeSeqpacket: return "seqpacket" } return "unknown" } func (f NetUNIXFlags) String() string { switch f { case netUnixFlagListen: return "listen" default: return "default" } } func (s NetUNIXState) String() string { switch s { case netUnixStateUnconnected: return "unconnected" case netUnixStateConnecting: return "connecting" case netUnixStateConnected: return "connected" case netUnixStateDisconnected: return "disconnected" } return "unknown" }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/fs.go
vendor/github.com/prometheus/procfs/fs.go
// Copyright 2018 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package procfs import ( "github.com/prometheus/procfs/internal/fs" ) // FS represents the pseudo-filesystem sys, which provides an interface to // kernel data structures. type FS struct { proc fs.FS isReal bool } // DefaultMountPoint is the common mount point of the proc filesystem. const DefaultMountPoint = fs.DefaultProcMountPoint // NewDefaultFS returns a new proc FS mounted under the default proc mountPoint. // It will error if the mount point directory can't be read or is a file. func NewDefaultFS() (FS, error) { return NewFS(DefaultMountPoint) } // NewFS returns a new proc FS mounted under the given proc mountPoint. It will error // if the mount point directory can't be read or is a file. func NewFS(mountPoint string) (FS, error) { fs, err := fs.NewFS(mountPoint) if err != nil { return FS{}, err } isReal, err := isRealProc(mountPoint) if err != nil { return FS{}, err } return FS{fs, isReal}, nil }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/mdstat.go
vendor/github.com/prometheus/procfs/mdstat.go
// Copyright 2018 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package procfs import ( "fmt" "os" "regexp" "strconv" "strings" ) var ( statusLineRE = regexp.MustCompile(`(\d+) blocks .*\[(\d+)/(\d+)\] \[([U_]+)\]`) recoveryLineBlocksRE = regexp.MustCompile(`\((\d+/\d+)\)`) recoveryLinePctRE = regexp.MustCompile(`= (.+)%`) recoveryLineFinishRE = regexp.MustCompile(`finish=(.+)min`) recoveryLineSpeedRE = regexp.MustCompile(`speed=(.+)[A-Z]`) componentDeviceRE = regexp.MustCompile(`(.*)\[\d+\]`) ) // MDStat holds info parsed from /proc/mdstat. type MDStat struct { // Name of the device. Name string // activity-state of the device. ActivityState string // Number of active disks. DisksActive int64 // Total number of disks the device requires. DisksTotal int64 // Number of failed disks. DisksFailed int64 // Number of "down" disks. (the _ indicator in the status line) DisksDown int64 // Spare disks in the device. DisksSpare int64 // Number of blocks the device holds. BlocksTotal int64 // Number of blocks on the device that are in sync. BlocksSynced int64 // Number of blocks on the device that need to be synced. BlocksToBeSynced int64 // progress percentage of current sync BlocksSyncedPct float64 // estimated finishing time for current sync (in minutes) BlocksSyncedFinishTime float64 // current sync speed (in Kilobytes/sec) BlocksSyncedSpeed float64 // Name of md component devices Devices []string } // MDStat parses an mdstat-file (/proc/mdstat) and returns a slice of // structs containing the relevant info. More information available here: // https://raid.wiki.kernel.org/index.php/Mdstat func (fs FS) MDStat() ([]MDStat, error) { data, err := os.ReadFile(fs.proc.Path("mdstat")) if err != nil { return nil, err } mdstat, err := parseMDStat(data) if err != nil { return nil, fmt.Errorf("%w: Cannot parse %v: %w", ErrFileParse, fs.proc.Path("mdstat"), err) } return mdstat, nil } // parseMDStat parses data from mdstat file (/proc/mdstat) and returns a slice of // structs containing the relevant info. func parseMDStat(mdStatData []byte) ([]MDStat, error) { mdStats := []MDStat{} lines := strings.Split(string(mdStatData), "\n") for i, line := range lines { if strings.TrimSpace(line) == "" || line[0] == ' ' || strings.HasPrefix(line, "Personalities") || strings.HasPrefix(line, "unused") { continue } deviceFields := strings.Fields(line) if len(deviceFields) < 3 { return nil, fmt.Errorf("%w: Expected 3+ lines, got %q", ErrFileParse, line) } mdName := deviceFields[0] // mdx state := deviceFields[2] // active or inactive if len(lines) <= i+3 { return nil, fmt.Errorf("%w: Too few lines for md device: %q", ErrFileParse, mdName) } // Failed disks have the suffix (F) & Spare disks have the suffix (S). fail := int64(strings.Count(line, "(F)")) spare := int64(strings.Count(line, "(S)")) active, total, down, size, err := evalStatusLine(lines[i], lines[i+1]) if err != nil { return nil, fmt.Errorf("%w: Cannot parse md device lines: %v: %w", ErrFileParse, active, err) } syncLineIdx := i + 2 if strings.Contains(lines[i+2], "bitmap") { // skip bitmap line syncLineIdx++ } // If device is syncing at the moment, get the number of currently // synced bytes, otherwise that number equals the size of the device. blocksSynced := size blocksToBeSynced := size speed := float64(0) finish := float64(0) pct := float64(0) recovering := strings.Contains(lines[syncLineIdx], "recovery") resyncing := strings.Contains(lines[syncLineIdx], "resync") checking := strings.Contains(lines[syncLineIdx], "check") // Append recovery and resyncing state info. if recovering || resyncing || checking { if recovering { state = "recovering" } else if checking { state = "checking" } else { state = "resyncing" } // Handle case when resync=PENDING or resync=DELAYED. if strings.Contains(lines[syncLineIdx], "PENDING") || strings.Contains(lines[syncLineIdx], "DELAYED") { blocksSynced = 0 } else { blocksSynced, blocksToBeSynced, pct, finish, speed, err = evalRecoveryLine(lines[syncLineIdx]) if err != nil { return nil, fmt.Errorf("%w: Cannot parse sync line in md device: %q: %w", ErrFileParse, mdName, err) } } } mdStats = append(mdStats, MDStat{ Name: mdName, ActivityState: state, DisksActive: active, DisksFailed: fail, DisksDown: down, DisksSpare: spare, DisksTotal: total, BlocksTotal: size, BlocksSynced: blocksSynced, BlocksToBeSynced: blocksToBeSynced, BlocksSyncedPct: pct, BlocksSyncedFinishTime: finish, BlocksSyncedSpeed: speed, Devices: evalComponentDevices(deviceFields), }) } return mdStats, nil } func evalStatusLine(deviceLine, statusLine string) (active, total, down, size int64, err error) { statusFields := strings.Fields(statusLine) if len(statusFields) < 1 { return 0, 0, 0, 0, fmt.Errorf("%w: Unexpected statusline %q: %w", ErrFileParse, statusLine, err) } sizeStr := statusFields[0] size, err = strconv.ParseInt(sizeStr, 10, 64) if err != nil { return 0, 0, 0, 0, fmt.Errorf("%w: Unexpected statusline %q: %w", ErrFileParse, statusLine, err) } if strings.Contains(deviceLine, "raid0") || strings.Contains(deviceLine, "linear") { // In the device deviceLine, only disks have a number associated with them in []. total = int64(strings.Count(deviceLine, "[")) return total, total, 0, size, nil } if strings.Contains(deviceLine, "inactive") { return 0, 0, 0, size, nil } matches := statusLineRE.FindStringSubmatch(statusLine) if len(matches) != 5 { return 0, 0, 0, 0, fmt.Errorf("%w: Could not fild all substring matches %s: %w", ErrFileParse, statusLine, err) } total, err = strconv.ParseInt(matches[2], 10, 64) if err != nil { return 0, 0, 0, 0, fmt.Errorf("%w: Unexpected statusline %q: %w", ErrFileParse, statusLine, err) } active, err = strconv.ParseInt(matches[3], 10, 64) if err != nil { return 0, 0, 0, 0, fmt.Errorf("%w: Unexpected active %d: %w", ErrFileParse, active, err) } down = int64(strings.Count(matches[4], "_")) return active, total, down, size, nil } func evalRecoveryLine(recoveryLine string) (blocksSynced int64, blocksToBeSynced int64, pct float64, finish float64, speed float64, err error) { matches := recoveryLineBlocksRE.FindStringSubmatch(recoveryLine) if len(matches) != 2 { return 0, 0, 0, 0, 0, fmt.Errorf("%w: Unexpected recoveryLine blocks %s: %w", ErrFileParse, recoveryLine, err) } blocks := strings.Split(matches[1], "/") blocksSynced, err = strconv.ParseInt(blocks[0], 10, 64) if err != nil { return 0, 0, 0, 0, 0, fmt.Errorf("%w: Unable to parse recovery blocks synced %q: %w", ErrFileParse, matches[1], err) } blocksToBeSynced, err = strconv.ParseInt(blocks[1], 10, 64) if err != nil { return blocksSynced, 0, 0, 0, 0, fmt.Errorf("%w: Unable to parse recovery to be synced blocks %q: %w", ErrFileParse, matches[2], err) } // Get percentage complete matches = recoveryLinePctRE.FindStringSubmatch(recoveryLine) if len(matches) != 2 { return blocksSynced, blocksToBeSynced, 0, 0, 0, fmt.Errorf("%w: Unexpected recoveryLine matching percentage %s", ErrFileParse, recoveryLine) } pct, err = strconv.ParseFloat(strings.TrimSpace(matches[1]), 64) if err != nil { return blocksSynced, blocksToBeSynced, 0, 0, 0, fmt.Errorf("%w: Error parsing float from recoveryLine %q", ErrFileParse, recoveryLine) } // Get time expected left to complete matches = recoveryLineFinishRE.FindStringSubmatch(recoveryLine) if len(matches) != 2 { return blocksSynced, blocksToBeSynced, pct, 0, 0, fmt.Errorf("%w: Unexpected recoveryLine matching est. finish time: %s", ErrFileParse, recoveryLine) } finish, err = strconv.ParseFloat(matches[1], 64) if err != nil { return blocksSynced, blocksToBeSynced, pct, 0, 0, fmt.Errorf("%w: Unable to parse float from recoveryLine: %q", ErrFileParse, recoveryLine) } // Get recovery speed matches = recoveryLineSpeedRE.FindStringSubmatch(recoveryLine) if len(matches) != 2 { return blocksSynced, blocksToBeSynced, pct, finish, 0, fmt.Errorf("%w: Unexpected recoveryLine value: %s", ErrFileParse, recoveryLine) } speed, err = strconv.ParseFloat(matches[1], 64) if err != nil { return blocksSynced, blocksToBeSynced, pct, finish, 0, fmt.Errorf("%w: Error parsing float from recoveryLine: %q: %w", ErrFileParse, recoveryLine, err) } return blocksSynced, blocksToBeSynced, pct, finish, speed, nil } func evalComponentDevices(deviceFields []string) []string { mdComponentDevices := make([]string, 0) if len(deviceFields) > 3 { for _, field := range deviceFields[4:] { match := componentDeviceRE.FindStringSubmatch(field) if match == nil { continue } mdComponentDevices = append(mdComponentDevices, match[1]) } } return mdComponentDevices }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/cpuinfo_s390x.go
vendor/github.com/prometheus/procfs/cpuinfo_s390x.go
// Copyright 2020 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. //go:build linux // +build linux package procfs var parseCPUInfo = parseCPUInfoS390X
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/net_xfrm.go
vendor/github.com/prometheus/procfs/net_xfrm.go
// Copyright 2017 Prometheus Team // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package procfs import ( "bufio" "fmt" "os" "strconv" "strings" ) // XfrmStat models the contents of /proc/net/xfrm_stat. type XfrmStat struct { // All errors which are not matched by other XfrmInError int // No buffer is left XfrmInBufferError int // Header Error XfrmInHdrError int // No state found // i.e. either inbound SPI, address, or IPSEC protocol at SA is wrong XfrmInNoStates int // Transformation protocol specific error // e.g. SA Key is wrong XfrmInStateProtoError int // Transformation mode specific error XfrmInStateModeError int // Sequence error // e.g. sequence number is out of window XfrmInStateSeqError int // State is expired XfrmInStateExpired int // State has mismatch option // e.g. UDP encapsulation type is mismatched XfrmInStateMismatch int // State is invalid XfrmInStateInvalid int // No matching template for states // e.g. Inbound SAs are correct but SP rule is wrong XfrmInTmplMismatch int // No policy is found for states // e.g. Inbound SAs are correct but no SP is found XfrmInNoPols int // Policy discards XfrmInPolBlock int // Policy error XfrmInPolError int // All errors which are not matched by others XfrmOutError int // Bundle generation error XfrmOutBundleGenError int // Bundle check error XfrmOutBundleCheckError int // No state was found XfrmOutNoStates int // Transformation protocol specific error XfrmOutStateProtoError int // Transportation mode specific error XfrmOutStateModeError int // Sequence error // i.e sequence number overflow XfrmOutStateSeqError int // State is expired XfrmOutStateExpired int // Policy discads XfrmOutPolBlock int // Policy is dead XfrmOutPolDead int // Policy Error XfrmOutPolError int // Forward routing of a packet is not allowed XfrmFwdHdrError int // State is invalid, perhaps expired XfrmOutStateInvalid int // State hasn’t been fully acquired before use XfrmAcquireError int } // NewXfrmStat reads the xfrm_stat statistics. func NewXfrmStat() (XfrmStat, error) { fs, err := NewFS(DefaultMountPoint) if err != nil { return XfrmStat{}, err } return fs.NewXfrmStat() } // NewXfrmStat reads the xfrm_stat statistics from the 'proc' filesystem. func (fs FS) NewXfrmStat() (XfrmStat, error) { file, err := os.Open(fs.proc.Path("net/xfrm_stat")) if err != nil { return XfrmStat{}, err } defer file.Close() var ( x = XfrmStat{} s = bufio.NewScanner(file) ) for s.Scan() { fields := strings.Fields(s.Text()) if len(fields) != 2 { return XfrmStat{}, fmt.Errorf("%w: %q line %q", ErrFileParse, file.Name(), s.Text()) } name := fields[0] value, err := strconv.Atoi(fields[1]) if err != nil { return XfrmStat{}, err } switch name { case "XfrmInError": x.XfrmInError = value case "XfrmInBufferError": x.XfrmInBufferError = value case "XfrmInHdrError": x.XfrmInHdrError = value case "XfrmInNoStates": x.XfrmInNoStates = value case "XfrmInStateProtoError": x.XfrmInStateProtoError = value case "XfrmInStateModeError": x.XfrmInStateModeError = value case "XfrmInStateSeqError": x.XfrmInStateSeqError = value case "XfrmInStateExpired": x.XfrmInStateExpired = value case "XfrmInStateInvalid": x.XfrmInStateInvalid = value case "XfrmInTmplMismatch": x.XfrmInTmplMismatch = value case "XfrmInNoPols": x.XfrmInNoPols = value case "XfrmInPolBlock": x.XfrmInPolBlock = value case "XfrmInPolError": x.XfrmInPolError = value case "XfrmOutError": x.XfrmOutError = value case "XfrmInStateMismatch": x.XfrmInStateMismatch = value case "XfrmOutBundleGenError": x.XfrmOutBundleGenError = value case "XfrmOutBundleCheckError": x.XfrmOutBundleCheckError = value case "XfrmOutNoStates": x.XfrmOutNoStates = value case "XfrmOutStateProtoError": x.XfrmOutStateProtoError = value case "XfrmOutStateModeError": x.XfrmOutStateModeError = value case "XfrmOutStateSeqError": x.XfrmOutStateSeqError = value case "XfrmOutStateExpired": x.XfrmOutStateExpired = value case "XfrmOutPolBlock": x.XfrmOutPolBlock = value case "XfrmOutPolDead": x.XfrmOutPolDead = value case "XfrmOutPolError": x.XfrmOutPolError = value case "XfrmFwdHdrError": x.XfrmFwdHdrError = value case "XfrmOutStateInvalid": x.XfrmOutStateInvalid = value case "XfrmAcquireError": x.XfrmAcquireError = value } } return x, s.Err() }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/proc_ns.go
vendor/github.com/prometheus/procfs/proc_ns.go
// Copyright 2018 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package procfs import ( "fmt" "os" "strconv" "strings" ) // Namespace represents a single namespace of a process. type Namespace struct { Type string // Namespace type. Inode uint32 // Inode number of the namespace. If two processes are in the same namespace their inodes will match. } // Namespaces contains all of the namespaces that the process is contained in. type Namespaces map[string]Namespace // Namespaces reads from /proc/<pid>/ns/* to get the namespaces of which the // process is a member. func (p Proc) Namespaces() (Namespaces, error) { d, err := os.Open(p.path("ns")) if err != nil { return nil, err } defer d.Close() names, err := d.Readdirnames(-1) if err != nil { return nil, fmt.Errorf("%w: failed to read contents of ns dir: %w", ErrFileRead, err) } ns := make(Namespaces, len(names)) for _, name := range names { target, err := os.Readlink(p.path("ns", name)) if err != nil { return nil, err } fields := strings.SplitN(target, ":", 2) if len(fields) != 2 { return nil, fmt.Errorf("%w: namespace type and inode from %q", ErrFileParse, target) } typ := fields[0] inode, err := strconv.ParseUint(strings.Trim(fields[1], "[]"), 10, 32) if err != nil { return nil, fmt.Errorf("%w: inode from %q: %w", ErrFileParse, fields[1], err) } ns[name] = Namespace{typ, uint32(inode)} } return ns, nil }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/cpuinfo_mipsx.go
vendor/github.com/prometheus/procfs/cpuinfo_mipsx.go
// Copyright 2020 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. //go:build linux && (mips || mipsle || mips64 || mips64le) // +build linux // +build mips mipsle mips64 mips64le package procfs var parseCPUInfo = parseCPUInfoMips
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/proc_sys.go
vendor/github.com/prometheus/procfs/proc_sys.go
// Copyright 2022 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package procfs import ( "fmt" "strings" "github.com/prometheus/procfs/internal/util" ) func sysctlToPath(sysctl string) string { return strings.Replace(sysctl, ".", "/", -1) } func (fs FS) SysctlStrings(sysctl string) ([]string, error) { value, err := util.SysReadFile(fs.proc.Path("sys", sysctlToPath(sysctl))) if err != nil { return nil, err } return strings.Fields(value), nil } func (fs FS) SysctlInts(sysctl string) ([]int, error) { fields, err := fs.SysctlStrings(sysctl) if err != nil { return nil, err } values := make([]int, len(fields)) for i, f := range fields { vp := util.NewValueParser(f) values[i] = vp.Int() if err := vp.Err(); err != nil { return nil, fmt.Errorf("%w: field %d in sysctl %s is not a valid int: %w", ErrFileParse, i, sysctl, err) } } return values, nil }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/proc_maps.go
vendor/github.com/prometheus/procfs/proc_maps.go
// Copyright 2019 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. //go:build (aix || darwin || dragonfly || freebsd || linux || netbsd || openbsd || solaris) && !js // +build aix darwin dragonfly freebsd linux netbsd openbsd solaris // +build !js package procfs import ( "bufio" "fmt" "os" "strconv" "strings" "golang.org/x/sys/unix" ) // ProcMapPermissions contains permission settings read from `/proc/[pid]/maps`. type ProcMapPermissions struct { // mapping has the [R]ead flag set Read bool // mapping has the [W]rite flag set Write bool // mapping has the [X]ecutable flag set Execute bool // mapping has the [S]hared flag set Shared bool // mapping is marked as [P]rivate (copy on write) Private bool } // ProcMap contains the process memory-mappings of the process // read from `/proc/[pid]/maps`. type ProcMap struct { // The start address of current mapping. StartAddr uintptr // The end address of the current mapping EndAddr uintptr // The permissions for this mapping Perms *ProcMapPermissions // The current offset into the file/fd (e.g., shared libs) Offset int64 // Device owner of this mapping (major:minor) in Mkdev format. Dev uint64 // The inode of the device above Inode uint64 // The file or psuedofile (or empty==anonymous) Pathname string } // parseDevice parses the device token of a line and converts it to a dev_t // (mkdev) like structure. func parseDevice(s string) (uint64, error) { i := strings.Index(s, ":") if i == -1 { return 0, fmt.Errorf("%w: expected separator `:` in %s", ErrFileParse, s) } major, err := strconv.ParseUint(s[0:i], 16, 0) if err != nil { return 0, err } minor, err := strconv.ParseUint(s[i+1:], 16, 0) if err != nil { return 0, err } return unix.Mkdev(uint32(major), uint32(minor)), nil } // parseAddress converts a hex-string to a uintptr. func parseAddress(s string) (uintptr, error) { a, err := strconv.ParseUint(s, 16, 0) if err != nil { return 0, err } return uintptr(a), nil } // parseAddresses parses the start-end address. func parseAddresses(s string) (uintptr, uintptr, error) { idx := strings.Index(s, "-") if idx == -1 { return 0, 0, fmt.Errorf("%w: expected separator `-` in %s", ErrFileParse, s) } saddr, err := parseAddress(s[0:idx]) if err != nil { return 0, 0, err } eaddr, err := parseAddress(s[idx+1:]) if err != nil { return 0, 0, err } return saddr, eaddr, nil } // parsePermissions parses a token and returns any that are set. func parsePermissions(s string) (*ProcMapPermissions, error) { if len(s) < 4 { return nil, fmt.Errorf("%w: invalid permissions token", ErrFileParse) } perms := ProcMapPermissions{} for _, ch := range s { switch ch { case 'r': perms.Read = true case 'w': perms.Write = true case 'x': perms.Execute = true case 'p': perms.Private = true case 's': perms.Shared = true } } return &perms, nil } // parseProcMap will attempt to parse a single line within a proc/[pid]/maps // buffer. func parseProcMap(text string) (*ProcMap, error) { fields := strings.Fields(text) if len(fields) < 5 { return nil, fmt.Errorf("%w: truncated procmap entry", ErrFileParse) } saddr, eaddr, err := parseAddresses(fields[0]) if err != nil { return nil, err } perms, err := parsePermissions(fields[1]) if err != nil { return nil, err } offset, err := strconv.ParseInt(fields[2], 16, 0) if err != nil { return nil, err } device, err := parseDevice(fields[3]) if err != nil { return nil, err } inode, err := strconv.ParseUint(fields[4], 10, 0) if err != nil { return nil, err } pathname := "" if len(fields) >= 5 { pathname = strings.Join(fields[5:], " ") } return &ProcMap{ StartAddr: saddr, EndAddr: eaddr, Perms: perms, Offset: offset, Dev: device, Inode: inode, Pathname: pathname, }, nil } // ProcMaps reads from /proc/[pid]/maps to get the memory-mappings of the // process. func (p Proc) ProcMaps() ([]*ProcMap, error) { file, err := os.Open(p.path("maps")) if err != nil { return nil, err } defer file.Close() maps := []*ProcMap{} scan := bufio.NewScanner(file) for scan.Scan() { m, err := parseProcMap(scan.Text()) if err != nil { return nil, err } maps = append(maps, m) } return maps, nil }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/thread.go
vendor/github.com/prometheus/procfs/thread.go
// Copyright 2022 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package procfs import ( "fmt" "os" "strconv" fsi "github.com/prometheus/procfs/internal/fs" ) // Provide access to /proc/PID/task/TID files, for thread specific values. Since // such files have the same structure as /proc/PID/ ones, the data structures // and the parsers for the latter may be reused. // AllThreads returns a list of all currently available threads under /proc/PID. func AllThreads(pid int) (Procs, error) { fs, err := NewFS(DefaultMountPoint) if err != nil { return Procs{}, err } return fs.AllThreads(pid) } // AllThreads returns a list of all currently available threads for PID. func (fs FS) AllThreads(pid int) (Procs, error) { taskPath := fs.proc.Path(strconv.Itoa(pid), "task") d, err := os.Open(taskPath) if err != nil { return Procs{}, err } defer d.Close() names, err := d.Readdirnames(-1) if err != nil { return Procs{}, fmt.Errorf("%w: could not read %q: %w", ErrFileRead, d.Name(), err) } t := Procs{} for _, n := range names { tid, err := strconv.ParseInt(n, 10, 64) if err != nil { continue } t = append(t, Proc{PID: int(tid), fs: FS{fsi.FS(taskPath), fs.isReal}}) } return t, nil } // Thread returns a process for a given PID, TID. func (fs FS) Thread(pid, tid int) (Proc, error) { taskPath := fs.proc.Path(strconv.Itoa(pid), "task") if _, err := os.Stat(taskPath); err != nil { return Proc{}, err } return Proc{PID: tid, fs: FS{fsi.FS(taskPath), fs.isReal}}, nil } // Thread returns a process for a given TID of Proc. func (proc Proc) Thread(tid int) (Proc, error) { tfs := FS{fsi.FS(proc.path("task")), proc.fs.isReal} if _, err := os.Stat(tfs.proc.Path(strconv.Itoa(tid))); err != nil { return Proc{}, err } return Proc{PID: tid, fs: tfs}, nil }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/internal/util/readfile.go
vendor/github.com/prometheus/procfs/internal/util/readfile.go
// Copyright 2019 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package util import ( "io" "os" ) // ReadFileNoStat uses io.ReadAll to read contents of entire file. // This is similar to os.ReadFile but without the call to os.Stat, because // many files in /proc and /sys report incorrect file sizes (either 0 or 4096). // Reads a max file size of 1024kB. For files larger than this, a scanner // should be used. func ReadFileNoStat(filename string) ([]byte, error) { const maxBufferSize = 1024 * 1024 f, err := os.Open(filename) if err != nil { return nil, err } defer f.Close() reader := io.LimitReader(f, maxBufferSize) return io.ReadAll(reader) }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/internal/util/sysreadfile_compat.go
vendor/github.com/prometheus/procfs/internal/util/sysreadfile_compat.go
// Copyright 2019 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. //go:build (linux && appengine) || (!linux && !darwin) // +build linux,appengine !linux,!darwin package util import ( "fmt" ) // SysReadFile is here implemented as a noop for builds that do not support // the read syscall. For example Windows, or Linux on Google App Engine. func SysReadFile(file string) (string, error) { return "", fmt.Errorf("not supported on this platform") }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/internal/util/valueparser.go
vendor/github.com/prometheus/procfs/internal/util/valueparser.go
// Copyright 2019 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package util import ( "strconv" ) // TODO(mdlayher): util packages are an anti-pattern and this should be moved // somewhere else that is more focused in the future. // A ValueParser enables parsing a single string into a variety of data types // in a concise and safe way. The Err method must be invoked after invoking // any other methods to ensure a value was successfully parsed. type ValueParser struct { v string err error } // NewValueParser creates a ValueParser using the input string. func NewValueParser(v string) *ValueParser { return &ValueParser{v: v} } // Int interprets the underlying value as an int and returns that value. func (vp *ValueParser) Int() int { return int(vp.int64()) } // PInt64 interprets the underlying value as an int64 and returns a pointer to // that value. func (vp *ValueParser) PInt64() *int64 { if vp.err != nil { return nil } v := vp.int64() return &v } // int64 interprets the underlying value as an int64 and returns that value. // TODO: export if/when necessary. func (vp *ValueParser) int64() int64 { if vp.err != nil { return 0 } // A base value of zero makes ParseInt infer the correct base using the // string's prefix, if any. const base = 0 v, err := strconv.ParseInt(vp.v, base, 64) if err != nil { vp.err = err return 0 } return v } // PUInt64 interprets the underlying value as an uint64 and returns a pointer to // that value. func (vp *ValueParser) PUInt64() *uint64 { if vp.err != nil { return nil } // A base value of zero makes ParseInt infer the correct base using the // string's prefix, if any. const base = 0 v, err := strconv.ParseUint(vp.v, base, 64) if err != nil { vp.err = err return nil } return &v } // Err returns the last error, if any, encountered by the ValueParser. func (vp *ValueParser) Err() error { return vp.err }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/internal/util/sysreadfile.go
vendor/github.com/prometheus/procfs/internal/util/sysreadfile.go
// Copyright 2018 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. //go:build (linux || darwin) && !appengine // +build linux darwin // +build !appengine package util import ( "bytes" "os" "syscall" ) // SysReadFile is a simplified os.ReadFile that invokes syscall.Read directly. // https://github.com/prometheus/node_exporter/pull/728/files // // Note that this function will not read files larger than 128 bytes. func SysReadFile(file string) (string, error) { f, err := os.Open(file) if err != nil { return "", err } defer f.Close() // On some machines, hwmon drivers are broken and return EAGAIN. This causes // Go's os.ReadFile implementation to poll forever. // // Since we either want to read data or bail immediately, do the simplest // possible read using syscall directly. const sysFileBufferSize = 128 b := make([]byte, sysFileBufferSize) n, err := syscall.Read(int(f.Fd()), b) if err != nil { return "", err } return string(bytes.TrimSpace(b[:n])), nil }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/internal/util/parse.go
vendor/github.com/prometheus/procfs/internal/util/parse.go
// Copyright 2018 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package util import ( "os" "strconv" "strings" ) // ParseUint32s parses a slice of strings into a slice of uint32s. func ParseUint32s(ss []string) ([]uint32, error) { us := make([]uint32, 0, len(ss)) for _, s := range ss { u, err := strconv.ParseUint(s, 10, 32) if err != nil { return nil, err } us = append(us, uint32(u)) } return us, nil } // ParseUint64s parses a slice of strings into a slice of uint64s. func ParseUint64s(ss []string) ([]uint64, error) { us := make([]uint64, 0, len(ss)) for _, s := range ss { u, err := strconv.ParseUint(s, 10, 64) if err != nil { return nil, err } us = append(us, u) } return us, nil } // ParsePInt64s parses a slice of strings into a slice of int64 pointers. func ParsePInt64s(ss []string) ([]*int64, error) { us := make([]*int64, 0, len(ss)) for _, s := range ss { u, err := strconv.ParseInt(s, 10, 64) if err != nil { return nil, err } us = append(us, &u) } return us, nil } // Parses a uint64 from given hex in string. func ParseHexUint64s(ss []string) ([]*uint64, error) { us := make([]*uint64, 0, len(ss)) for _, s := range ss { u, err := strconv.ParseUint(s, 16, 64) if err != nil { return nil, err } us = append(us, &u) } return us, nil } // ReadUintFromFile reads a file and attempts to parse a uint64 from it. func ReadUintFromFile(path string) (uint64, error) { data, err := os.ReadFile(path) if err != nil { return 0, err } return strconv.ParseUint(strings.TrimSpace(string(data)), 10, 64) } // ReadIntFromFile reads a file and attempts to parse a int64 from it. func ReadIntFromFile(path string) (int64, error) { data, err := os.ReadFile(path) if err != nil { return 0, err } return strconv.ParseInt(strings.TrimSpace(string(data)), 10, 64) } // ParseBool parses a string into a boolean pointer. func ParseBool(b string) *bool { var truth bool switch b { case "enabled": truth = true case "disabled": truth = false default: return nil } return &truth }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false
kubev2v/forklift
https://github.com/kubev2v/forklift/blob/b3b4703e958c25d54c4d48138d9e80ae32fadac3/vendor/github.com/prometheus/procfs/internal/fs/fs.go
vendor/github.com/prometheus/procfs/internal/fs/fs.go
// Copyright 2019 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package fs import ( "fmt" "os" "path/filepath" ) const ( // DefaultProcMountPoint is the common mount point of the proc filesystem. DefaultProcMountPoint = "/proc" // DefaultSysMountPoint is the common mount point of the sys filesystem. DefaultSysMountPoint = "/sys" // DefaultConfigfsMountPoint is the common mount point of the configfs. DefaultConfigfsMountPoint = "/sys/kernel/config" ) // FS represents a pseudo-filesystem, normally /proc or /sys, which provides an // interface to kernel data structures. type FS string // NewFS returns a new FS mounted under the given mountPoint. It will error // if the mount point can't be read. func NewFS(mountPoint string) (FS, error) { info, err := os.Stat(mountPoint) if err != nil { return "", fmt.Errorf("could not read %q: %w", mountPoint, err) } if !info.IsDir() { return "", fmt.Errorf("mount point %q is not a directory", mountPoint) } return FS(mountPoint), nil } // Path appends the given path elements to the filesystem path, adding separators // as necessary. func (fs FS) Path(p ...string) string { return filepath.Join(append([]string{string(fs)}, p...)...) }
go
Apache-2.0
b3b4703e958c25d54c4d48138d9e80ae32fadac3
2026-01-07T09:44:30.792320Z
false