content
stringlengths
1
103k
path
stringlengths
8
216
filename
stringlengths
2
179
language
stringclasses
15 values
size_bytes
int64
2
189k
quality_score
float64
0.5
0.95
complexity
float64
0
1
documentation_ratio
float64
0
1
repository
stringclasses
5 values
stars
int64
0
1k
created_date
stringdate
2023-07-10 19:21:08
2025-07-09 19:11:45
license
stringclasses
4 values
is_test
bool
2 classes
file_hash
stringlengths
32
32
engines:\n gofmt:\n enabled: true\n golint:\n enabled: true\n govet:\n enabled: true\n\nexclude_patterns:\n- ".github/"\n- "vendor/"\n- "codegen/"\n- "*.yml"\n- ".*.yml"\n- "*.md"\n- "Gopkg.*"\n- "doc.go"\n- "type_specific_codegen_test.go"\n- "type_specific_codegen.go"\n- ".gitignore"\n- "LICENSE"\n
dataset_sample\yaml\kubernetes_kubernetes\vendor\github.com\stretchr\objx\.codeclimate.yml
.codeclimate.yml
YAML
291
0.7
0
0
awesome-app
701
2024-05-19T00:17:49.190138
Apache-2.0
false
e562ab2d6bdcdea92e1777334824ba2d
version: '3'\n\ntasks:\n default:\n deps: [test]\n\n lint:\n desc: Checks code style\n cmds:\n - gofmt -d -s *.go\n - go vet ./...\n silent: true\n\n lint-fix:\n desc: Fixes code style\n cmds:\n - gofmt -w -s *.go\n\n test:\n desc: Runs go tests\n cmds:\n - go test -race ./...\n\n test-coverage:\n desc: Runs go tests and calculates test coverage\n cmds:\n - go test -race -coverprofile=c.out ./...\n
dataset_sample\yaml\kubernetes_kubernetes\vendor\github.com\stretchr\objx\Taskfile.yml
Taskfile.yml
YAML
435
0.7
0
0
node-utils
122
2023-08-04T10:13:55.673901
Apache-2.0
false
3d1a878f7a54a73dde3ef6efec96e4b5
language: go\n\ngo:\n - 1.11.x\n\nenv:\n - GO111MODULE=on \n \nscript:\n - go test -short -coverprofile=coverage.txt -covermode=count ./...\n\nafter_success:\n - bash <(curl -s https://codecov.io/bash) \n
dataset_sample\yaml\kubernetes_kubernetes\vendor\github.com\x448\float16\.travis.yml
.travis.yml
YAML
197
0.8
0
0
node-utils
921
2024-04-18T10:53:40.126797
Apache-2.0
false
773d8d82e81a1db1741743c8ad3b9b7d
# See https://github.com/golangci/golangci-lint#config-file\nrun:\n issues-exit-code: 1 #Default\n tests: true #Default\n\nlinters:\n # Disable everything by default so upgrades to not include new "default\n # enabled" linters.\n disable-all: true\n # Specifically enable linters we want to use.\n enable:\n - asasalint\n - bodyclose\n - depguard\n - errcheck\n - errorlint\n - godot\n - gofumpt\n - goimports\n - gosec\n - gosimple\n - govet\n - ineffassign\n - misspell\n - perfsprint\n - revive\n - staticcheck\n - tenv\n - testifylint\n - typecheck\n - unconvert\n - unused\n - unparam\n - usestdlibvars\n\nissues:\n # Maximum issues count per one linter.\n # Set to 0 to disable.\n # Default: 50\n # Setting to unlimited so the linter only is run once to debug all issues.\n max-issues-per-linter: 0\n # Maximum count of issues with the same text.\n # Set to 0 to disable.\n # Default: 3\n # Setting to unlimited so the linter only is run once to debug all issues.\n max-same-issues: 0\n # Excluding configuration per-path, per-linter, per-text and per-source.\n exclude-rules:\n # TODO: Having appropriate comments for exported objects helps development,\n # even for objects in internal packages. Appropriate comments for all\n # exported objects should be added and this exclusion removed.\n - path: '.*internal/.*'\n text: "exported (method|function|type|const) (.+) should have comment or be unexported"\n linters:\n - revive\n # Yes, they are, but it's okay in a test.\n - path: _test\.go\n text: "exported func.*returns unexported type.*which can be annoying to use"\n linters:\n - revive\n # Example test functions should be treated like main.\n - path: example.*_test\.go\n text: "calls to (.+) only in main[(][)] or init[(][)] functions"\n linters:\n - revive\n # It's okay to not run gosec and perfsprint in a test.\n - path: _test\.go\n linters:\n - gosec\n - perfsprint\n # Ignoring gosec G404: Use of weak random number generator (math/rand instead of crypto/rand)\n # as we commonly use it in tests and examples.\n - text: "G404:"\n linters:\n - gosec\n # Ignoring gosec G402: TLS MinVersion too low\n # as the https://pkg.go.dev/crypto/tls#Config handles MinVersion default well.\n - text: "G402: TLS MinVersion too low."\n linters:\n - gosec\n include:\n # revive exported should have comment or be unexported.\n - EXC0012\n # revive package comment should be of the form ...\n - EXC0013\n\nlinters-settings:\n depguard:\n rules:\n non-tests:\n files:\n - "!$test"\n - "!**/*test/*.go"\n - "!**/internal/matchers/*.go"\n deny:\n - pkg: "testing"\n - pkg: "github.com/stretchr/testify"\n - pkg: "crypto/md5"\n - pkg: "crypto/sha1"\n - pkg: "crypto/**/pkix"\n auto/sdk:\n files:\n - "!internal/global/trace.go"\n - "~internal/global/trace_test.go"\n deny:\n - pkg: "go.opentelemetry.io/auto/sdk"\n desc: Do not use SDK from automatic instrumentation.\n otlp-internal:\n files:\n - "!**/exporters/otlp/internal/**/*.go"\n deny:\n - pkg: "go.opentelemetry.io/otel/exporters/otlp/internal"\n desc: Do not use cross-module internal packages.\n otlptrace-internal:\n files:\n - "!**/exporters/otlp/otlptrace/*.go"\n - "!**/exporters/otlp/otlptrace/internal/**.go"\n deny:\n - pkg: "go.opentelemetry.io/otel/exporters/otlp/otlptrace/internal"\n desc: Do not use cross-module internal packages.\n otlpmetric-internal:\n files:\n - "!**/exporters/otlp/otlpmetric/internal/*.go"\n - "!**/exporters/otlp/otlpmetric/internal/**/*.go"\n deny:\n - pkg: "go.opentelemetry.io/otel/exporters/otlp/otlpmetric/internal"\n desc: Do not use cross-module internal packages.\n otel-internal:\n files:\n - "**/sdk/*.go"\n - "**/sdk/**/*.go"\n - "**/exporters/*.go"\n - "**/exporters/**/*.go"\n - "**/schema/*.go"\n - "**/schema/**/*.go"\n - "**/metric/*.go"\n - "**/metric/**/*.go"\n - "**/bridge/*.go"\n - "**/bridge/**/*.go"\n - "**/trace/*.go"\n - "**/trace/**/*.go"\n - "**/log/*.go"\n - "**/log/**/*.go"\n deny:\n - pkg: "go.opentelemetry.io/otel/internal$"\n desc: Do not use cross-module internal packages.\n - pkg: "go.opentelemetry.io/otel/internal/attribute"\n desc: Do not use cross-module internal packages.\n - pkg: "go.opentelemetry.io/otel/internal/internaltest"\n desc: Do not use cross-module internal packages.\n - pkg: "go.opentelemetry.io/otel/internal/matchers"\n desc: Do not use cross-module internal packages.\n godot:\n exclude:\n # Exclude links.\n - '^ *\[[^]]+\]:'\n # Exclude sentence fragments for lists.\n - '^[ ]*[-•]'\n # Exclude sentences prefixing a list.\n - ':$'\n goimports:\n local-prefixes: go.opentelemetry.io\n misspell:\n locale: US\n ignore-words:\n - cancelled\n perfsprint:\n err-error: true\n errorf: true\n int-conversion: true\n sprintf1: true\n strconcat: true\n revive:\n # Sets the default failure confidence.\n # This means that linting errors with less than 0.8 confidence will be ignored.\n # Default: 0.8\n confidence: 0.01\n rules:\n # https://github.com/mgechev/revive/blob/master/RULES_DESCRIPTIONS.md#blank-imports\n - name: blank-imports\n disabled: false\n # https://github.com/mgechev/revive/blob/master/RULES_DESCRIPTIONS.md#bool-literal-in-expr\n - name: bool-literal-in-expr\n disabled: false\n # https://github.com/mgechev/revive/blob/master/RULES_DESCRIPTIONS.md#constant-logical-expr\n - name: constant-logical-expr\n disabled: false\n # https://github.com/mgechev/revive/blob/master/RULES_DESCRIPTIONS.md#context-as-argument\n # TODO (#3372) re-enable linter when it is compatible. https://github.com/golangci/golangci-lint/issues/3280\n - name: context-as-argument\n disabled: true\n arguments:\n allowTypesBefore: "*testing.T"\n # https://github.com/mgechev/revive/blob/master/RULES_DESCRIPTIONS.md#context-keys-type\n - name: context-keys-type\n disabled: false\n # https://github.com/mgechev/revive/blob/master/RULES_DESCRIPTIONS.md#deep-exit\n - name: deep-exit\n disabled: false\n # https://github.com/mgechev/revive/blob/master/RULES_DESCRIPTIONS.md#defer\n - name: defer\n disabled: false\n arguments:\n - ["call-chain", "loop"]\n # https://github.com/mgechev/revive/blob/master/RULES_DESCRIPTIONS.md#dot-imports\n - name: dot-imports\n disabled: false\n # https://github.com/mgechev/revive/blob/master/RULES_DESCRIPTIONS.md#duplicated-imports\n - name: duplicated-imports\n disabled: false\n # https://github.com/mgechev/revive/blob/master/RULES_DESCRIPTIONS.md#early-return\n - name: early-return\n disabled: false\n # https://github.com/mgechev/revive/blob/master/RULES_DESCRIPTIONS.md#empty-block\n - name: empty-block\n disabled: false\n # https://github.com/mgechev/revive/blob/master/RULES_DESCRIPTIONS.md#empty-lines\n - name: empty-lines\n disabled: false\n # https://github.com/mgechev/revive/blob/master/RULES_DESCRIPTIONS.md#error-naming\n - name: error-naming\n disabled: false\n # https://github.com/mgechev/revive/blob/master/RULES_DESCRIPTIONS.md#error-return\n - name: error-return\n disabled: false\n # https://github.com/mgechev/revive/blob/master/RULES_DESCRIPTIONS.md#error-strings\n - name: error-strings\n disabled: false\n # https://github.com/mgechev/revive/blob/master/RULES_DESCRIPTIONS.md#errorf\n - name: errorf\n disabled: false\n # https://github.com/mgechev/revive/blob/master/RULES_DESCRIPTIONS.md#exported\n - name: exported\n disabled: false\n arguments:\n - "sayRepetitiveInsteadOfStutters"\n # https://github.com/mgechev/revive/blob/master/RULES_DESCRIPTIONS.md#flag-parameter\n - name: flag-parameter\n disabled: false\n # https://github.com/mgechev/revive/blob/master/RULES_DESCRIPTIONS.md#identical-branches\n - name: identical-branches\n disabled: false\n # https://github.com/mgechev/revive/blob/master/RULES_DESCRIPTIONS.md#if-return\n - name: if-return\n disabled: false\n # https://github.com/mgechev/revive/blob/master/RULES_DESCRIPTIONS.md#increment-decrement\n - name: increment-decrement\n disabled: false\n # https://github.com/mgechev/revive/blob/master/RULES_DESCRIPTIONS.md#indent-error-flow\n - name: indent-error-flow\n disabled: false\n # https://github.com/mgechev/revive/blob/master/RULES_DESCRIPTIONS.md#import-shadowing\n - name: import-shadowing\n disabled: false\n # https://github.com/mgechev/revive/blob/master/RULES_DESCRIPTIONS.md#package-comments\n - name: package-comments\n disabled: false\n # https://github.com/mgechev/revive/blob/master/RULES_DESCRIPTIONS.md#range\n - name: range\n disabled: false\n # https://github.com/mgechev/revive/blob/master/RULES_DESCRIPTIONS.md#range-val-in-closure\n - name: range-val-in-closure\n disabled: false\n # https://github.com/mgechev/revive/blob/master/RULES_DESCRIPTIONS.md#range-val-address\n - name: range-val-address\n disabled: false\n # https://github.com/mgechev/revive/blob/master/RULES_DESCRIPTIONS.md#redefines-builtin-id\n - name: redefines-builtin-id\n disabled: false\n # https://github.com/mgechev/revive/blob/master/RULES_DESCRIPTIONS.md#string-format\n - name: string-format\n disabled: false\n arguments:\n - - panic\n - '/^[^\n]*$/'\n - must not contain line breaks\n # https://github.com/mgechev/revive/blob/master/RULES_DESCRIPTIONS.md#struct-tag\n - name: struct-tag\n disabled: false\n # https://github.com/mgechev/revive/blob/master/RULES_DESCRIPTIONS.md#superfluous-else\n - name: superfluous-else\n disabled: false\n # https://github.com/mgechev/revive/blob/master/RULES_DESCRIPTIONS.md#time-equal\n - name: time-equal\n disabled: false\n # https://github.com/mgechev/revive/blob/master/RULES_DESCRIPTIONS.md#var-naming\n - name: var-naming\n disabled: false\n arguments:\n - ["ID"] # AllowList\n - ["Otel", "Aws", "Gcp"] # DenyList\n # https://github.com/mgechev/revive/blob/master/RULES_DESCRIPTIONS.md#var-declaration\n - name: var-declaration\n disabled: false\n # https://github.com/mgechev/revive/blob/master/RULES_DESCRIPTIONS.md#unconditional-recursion\n - name: unconditional-recursion\n disabled: false\n # https://github.com/mgechev/revive/blob/master/RULES_DESCRIPTIONS.md#unexported-return\n - name: unexported-return\n disabled: false\n # https://github.com/mgechev/revive/blob/master/RULES_DESCRIPTIONS.md#unhandled-error\n - name: unhandled-error\n disabled: false\n arguments:\n - "fmt.Fprint"\n - "fmt.Fprintf"\n - "fmt.Fprintln"\n - "fmt.Print"\n - "fmt.Printf"\n - "fmt.Println"\n # https://github.com/mgechev/revive/blob/master/RULES_DESCRIPTIONS.md#unnecessary-stmt\n - name: unnecessary-stmt\n disabled: false\n # https://github.com/mgechev/revive/blob/master/RULES_DESCRIPTIONS.md#useless-break\n - name: useless-break\n disabled: false\n # https://github.com/mgechev/revive/blob/master/RULES_DESCRIPTIONS.md#waitgroup-by-value\n - name: waitgroup-by-value\n disabled: false\n testifylint:\n enable-all: true\n disable:\n - float-compare\n - go-require\n - require-error\n
dataset_sample\yaml\kubernetes_kubernetes\vendor\go.opentelemetry.io\otel\.golangci.yml
.golangci.yml
YAML
12,126
0.95
0.021538
0.223602
awesome-app
751
2025-06-17T05:35:43.881038
Apache-2.0
false
51bb55ddac75789a915f2ea3db558791
output:\n # Make output more digestible with quickfix in vim/emacs/etc.\n sort-results: true\n print-issued-lines: false\n\nlinters:\n enable:\n - gofumpt\n - nolintlint\n - revive\n\nlinters-settings:\n govet:\n # These govet checks are disabled by default, but they're useful.\n enable:\n - niliness\n - reflectvaluecompare\n - sortslice\n - unusedwrite\n\nissues:\n # Print all issues reported by all linters.\n max-issues-per-linter: 0\n max-same-issues: 0\n\n # Don't ignore some of the issues that golangci-lint considers okay.\n # This includes documenting all exported entities.\n exclude-use-default: false\n
dataset_sample\yaml\kubernetes_kubernetes\vendor\go.uber.org\goleak\.golangci.yml
.golangci.yml
YAML
635
0.8
0
0.208333
node-utils
927
2024-06-07T09:26:07.006286
Apache-2.0
false
0f7fbc2a02ff0c3d66ab60b306590bf7
coverage:\n range: 80..100\n round: down\n precision: 2\n\n status:\n project: # measuring the overall project coverage\n default: # context, you can create multiple ones with custom titles\n enabled: yes # must be yes|true to enable this status\n target: 100 # specify the target coverage for each commit status\n # option: "auto" (must increase from parent commit or pull request base)\n # option: "X%" a static target percentage to hit\n if_not_found: success # if parent is not found report status as success, error, or failure\n if_ci_failed: error # if ci fails report status as success, error, or failure\n\n
dataset_sample\yaml\kubernetes_kubernetes\vendor\go.uber.org\multierr\.codecov.yml
.codecov.yml
YAML
763
0.8
0.2
0.153846
react-lib
609
2025-01-02T00:13:04.199753
BSD-3-Clause
false
47a21b3ecf2c1abbbffc83074925a3c0
coverage:\n range: 80..100\n round: down\n precision: 2\n\n status:\n project: # measuring the overall project coverage\n default: # context, you can create multiple ones with custom titles\n enabled: yes # must be yes|true to enable this status\n target: 95% # specify the target coverage for each commit status\n # option: "auto" (must increase from parent commit or pull request base)\n # option: "X%" a static target percentage to hit\n if_not_found: success # if parent is not found report status as success, error, or failure\n if_ci_failed: error # if ci fails report status as success, error, or failure\nignore:\n - internal/readme/readme.go\n\n
dataset_sample\yaml\kubernetes_kubernetes\vendor\go.uber.org\zap\.codecov.yml
.codecov.yml
YAML
801
0.8
0.176471
0.133333
react-lib
602
2025-03-07T03:00:38.205055
BSD-3-Clause
false
322c63b6b124b9e5ab95491701bf47fb
output:\n # Make output more digestible with quickfix in vim/emacs/etc.\n sort-results: true\n print-issued-lines: false\n\nlinters:\n # We'll track the golangci-lint default linters manually\n # instead of letting them change without our control.\n disable-all: true\n enable:\n # golangci-lint defaults:\n - errcheck\n - gosimple\n - govet\n - ineffassign\n - staticcheck\n - unused\n\n # Our own extras:\n - gofumpt\n - nolintlint # lints nolint directives\n - revive\n\nlinters-settings:\n govet:\n # These govet checks are disabled by default, but they're useful.\n enable:\n - niliness\n - reflectvaluecompare\n - sortslice\n - unusedwrite\n\n errcheck:\n exclude-functions:\n # These methods can not fail.\n # They operate on an in-memory buffer.\n - (*go.uber.org/zap/buffer.Buffer).Write\n - (*go.uber.org/zap/buffer.Buffer).WriteByte\n - (*go.uber.org/zap/buffer.Buffer).WriteString\n\n - (*go.uber.org/zap/zapio.Writer).Close\n - (*go.uber.org/zap/zapio.Writer).Sync\n - (*go.uber.org/zap/zapio.Writer).Write\n # Write to zapio.Writer cannot fail,\n # so io.WriteString on it cannot fail.\n - io.WriteString(*go.uber.org/zap/zapio.Writer)\n\n # Writing a plain string to a fmt.State cannot fail.\n - io.WriteString(fmt.State)\n\nissues:\n # Print all issues reported by all linters.\n max-issues-per-linter: 0\n max-same-issues: 0\n\n # Don't ignore some of the issues that golangci-lint considers okay.\n # This includes documenting all exported entities.\n exclude-use-default: false\n\n exclude-rules:\n # Don't warn on unused parameters.\n # Parameter names are useful; replacing them with '_' is undesirable.\n - linters: [revive]\n text: 'unused-parameter: parameter \S+ seems to be unused, consider removing or renaming it as _'\n\n # staticcheck already has smarter checks for empty blocks.\n # revive's empty-block linter has false positives.\n # For example, as of writing this, the following is not allowed.\n # for foo() { }\n - linters: [revive]\n text: 'empty-block: this block is empty, you can remove it'\n\n # Ignore logger.Sync() errcheck failures in example_test.go\n # since those are intended to be uncomplicated examples.\n - linters: [errcheck]\n path: example_test.go\n text: 'Error return value of `logger.Sync` is not checked'\n
dataset_sample\yaml\kubernetes_kubernetes\vendor\go.uber.org\zap\.golangci.yml
.golangci.yml
YAML
2,386
0.95
0.025974
0.333333
react-lib
440
2024-11-15T19:13:19.080005
Apache-2.0
false
b126debe5c78b49b8d216daa58b8968a
language: go\n\ngo:\n - tip\n\ninstall:\n - export GOPATH="$HOME/gopath"\n - mkdir -p "$GOPATH/src/golang.org/x"\n - mv "$TRAVIS_BUILD_DIR" "$GOPATH/src/golang.org/x/oauth2"\n - go get -v -t -d golang.org/x/oauth2/...\n\nscript:\n - go test -v golang.org/x/oauth2/...\n
dataset_sample\yaml\kubernetes_kubernetes\vendor\golang.org\x\oauth2\.travis.yml
.travis.yml
YAML
262
0.7
0
0
awesome-app
301
2024-08-11T17:08:51.935320
Apache-2.0
false
1a68e7809045200cda2e2ed6fc4f5890
language: go\n\nsudo: false\n\nmatrix:\n fast_finish: true\n allow_failures:\n - go: tip\n\ngo:\n- '1.14.x'\n- '1.15.x'\n- tip\n\ngo_import_path: gopkg.in/square/go-jose.v2\n\nbefore_script:\n- export PATH=$HOME/.local/bin:$PATH\n\nbefore_install:\n# Install encrypted gitcookies to get around bandwidth-limits\n# that is causing Travis-CI builds to fail. For more info, see\n# https://github.com/golang/go/issues/12933\n- openssl aes-256-cbc -K $encrypted_1528c3c2cafd_key -iv $encrypted_1528c3c2cafd_iv -in .gitcookies.sh.enc -out .gitcookies.sh -d || true\n- bash .gitcookies.sh || true\n- go get github.com/wadey/gocovmerge\n- go get github.com/mattn/goveralls\n- go get github.com/stretchr/testify/assert\n- go get github.com/stretchr/testify/require\n- go get github.com/google/go-cmp/cmp\n- go get golang.org/x/tools/cmd/cover || true\n- go get code.google.com/p/go.tools/cmd/cover || true\n- pip install cram --user\n\nscript:\n- go test . -v -covermode=count -coverprofile=profile.cov\n- go test ./cipher -v -covermode=count -coverprofile=cipher/profile.cov\n- go test ./jwt -v -covermode=count -coverprofile=jwt/profile.cov\n- go test ./json -v # no coverage for forked encoding/json package\n- cd jose-util && go build && PATH=$PWD:$PATH cram -v jose-util.t # cram tests jose-util\n- cd ..\n\nafter_success:\n- gocovmerge *.cov */*.cov > merged.coverprofile\n- $HOME/gopath/bin/goveralls -coverprofile merged.coverprofile -service=travis-ci\n
dataset_sample\yaml\kubernetes_kubernetes\vendor\gopkg.in\go-jose\go-jose.v2\.travis.yml
.travis.yml
YAML
1,413
0.95
0.022222
0.081081
awesome-app
136
2023-09-17T15:33:55.664875
MIT
false
61ce2ebc37cff1e819f020395ee5c6f1
language: go\narch: arm64\ndist: focal\ngo: 1.15.x\nscript:\n - diff -u <(echo -n) <(gofmt -d *.go)\n - diff -u <(echo -n) <(golint $(go list -e ./...) | grep -v YAMLToJSON)\n - GO111MODULE=on go vet .\n - GO111MODULE=on go test -v -race ./...\n - git diff --exit-code\ninstall:\n - GO111MODULE=off go get golang.org/x/lint/golint\n
dataset_sample\yaml\kubernetes_kubernetes\vendor\sigs.k8s.io\yaml\.travis.yml
.travis.yml
YAML
326
0.7
0
0
react-lib
832
2025-05-17T20:41:27.826288
BSD-3-Clause
false
80c3e7b46ca42256453e99bc0225f319
comment:\n layout: "reach, diff, flags, files"\n behavior: default\n require_changes: false # if true: only post the comment if coverage changes\n require_base: no # [yes :: must have a base report to post]\n require_head: yes # [yes :: must have a head report to post]\n branches: null # branch names that can post comment
dataset_sample\yaml\kubernetes_minikube\.codecov.yml
.codecov.yml
YAML
346
0.95
0.333333
0
vue-tools
698
2023-12-11T10:35:52.571720
MIT
false
a07a5040835b30940c9072bb7e41c8de
header:\n schema-version: 1.0.0\n expiration-date: '2024-12-17T01:00:00.000Z'\n last-updated: '2023-12-17'\n last-reviewed: '2023-12-17'\n commit-hash: 8220a6eb95f0a4d75f7f2d7b14cef975f050512d\n project-url: https://github.com/kubernetes/minikube\n project-release: '1.32.0'\n changelog: https://github.com/kubernetes/minikube/blob/master/CHANGELOG.md\n license: https://github.com/kubernetes/minikube/blob/master/LICENSE\nproject-lifecycle:\n status: active\n roadmap: https://minikube.sigs.k8s.io/docs/contrib/roadmap/\n bug-fixes-only: false\n core-maintainers:\n - https://github.com/kubernetes/minikube/blob/master/OWNERS\n release-cycle: https://minikube.sigs.k8s.io/docs/contrib/release_schedule/\n release-process: https://minikube.sigs.k8s.io/docs/contrib/releasing/\ncontribution-policy:\n accepts-pull-requests: true\n accepts-automated-pull-requests: true\n automated-tools-list:\n - automated-tool: dependabot\n action: allowed\n path:\n - /\n - automated-tool: minikube-bot\n action: allowed\n path:\n - /\n - automated-tool: k8s-ci-robot\n action: allowed\n path:\n - /\n contributing-policy: https://minikube.sigs.k8s.io/docs/contrib/guide/\n code-of-conduct: https://github.com/kubernetes/minikube/blob/master/code-of-conduct.md\ndocumentation:\n - https://minikube.sigs.k8s.io/docs/\ndistribution-points:\n - https://github.com/kubernetes/minikube/releases\nsecurity-artifacts:\n threat-model:\n threat-model-created: false\n self-assessment:\n self-assessment-created: false\nsecurity-testing:\n - tool-type: sca\n tool-name: Dependabot\n tool-version: "2"\n tool-url: https://github.com/dependabot\n integration:\n ad-hoc: false\n ci: true\n before-release: true\n tool-rulesets:\n - https://github.com/kubernetes/minikube/blob/master/.github/dependabot.yml\n - tool-type: sca\n tool-name: minikube-bot\n tool-version: latest\n tool-url: https://github.com/minikube-bot\n tool-rulesets:\n - built-in\n integration:\n ad-hoc: false\n ci: true\n before-release: true\nsecurity-contacts:\n - type: email\n value: security@kubernetes.io\n primary: true\nvulnerability-reporting:\n accepts-vulnerability-reports: true\n email-contact: security@kubernetes.io\n security-policy: https://github.com/kubernetes/minikube/blob/master/SECURITY.md\n bug-bounty-available: true\n bug-bounty-url: https://hackerone.com/kubernetes\ndependencies:\n third-party-packages: true\n dependencies-lists:\n - https://github.com/kubernetes/minikube/blob/master/go.mod\n
dataset_sample\yaml\kubernetes_minikube\SECURITY-INSIGHTS.yml
SECURITY-INSIGHTS.yml
YAML
2,542
0.8
0
0
python-kit
734
2025-02-25T20:05:51.729805
Apache-2.0
false
f6f216a37a4ff44e083980fa389869fe
services:\n volumes-provisioner:\n image: hasnat/volumes-provisioner\n environment:\n PROVISION_DIRECTORIES: "1000:1000:0755:/tmp/certs"\n volumes:\n - "./certs:/tmp/certs"\n network_mode: none\n\n cert-gen:\n image: eventstore/es-gencert-cli:1.0.2\n entrypoint: bash\n user: "1000:1000"\n command: >\n -c "mkdir -p ./certs && cd /certs\n && es-gencert-cli create-ca\n && es-gencert-cli create-node -out ./node1 -ip-addresses 127.0.0.1,172.30.240.11 -dns-names localhost\n && es-gencert-cli create-node -out ./node2 -ip-addresses 127.0.0.1,172.30.240.12 -dns-names localhost\n && es-gencert-cli create-node -out ./node3 -ip-addresses 127.0.0.1,172.30.240.13 -dns-names localhost\n && find . -type f -print0 | xargs -0 chmod 666"\n volumes:\n - "./certs:/certs"\n depends_on:\n - volumes-provisioner\n\n esdb-node1:\n build: \n context: .\n args:\n RUNTIME: ${RUNTIME}\n env_file:\n - shared.env\n environment:\n - KURRENTDB_GOSSIP_SEED=172.30.240.12:2113,172.30.240.13:2113\n - KURRENTDB_REPLICATION_IP=172.30.240.11\n - KURRENTDB_CERTIFICATE_FILE=/etc/kurrentdb/certs/node1/node.crt\n - KURRENTDB_CERTIFICATE_PRIVATE_KEY_FILE=/etc/kurrentdb/certs/node1/node.key\n - KURRENTDB_ADVERTISE_HOST_TO_CLIENT_AS=127.0.0.1\n - KURRENTDB_ADVERTISE_NODE_PORT_TO_CLIENT_AS=2111\n ports:\n - 2111:2113\n networks:\n clusternetwork:\n ipv4_address: 172.30.240.11\n volumes:\n - ./certs:/etc/kurrentdb/certs\n restart: unless-stopped\n depends_on:\n - cert-gen\n\n esdb-node2:\n build: \n context: .\n args:\n RUNTIME: ${RUNTIME}\n env_file:\n - shared.env\n environment:\n - KURRENTDB_GOSSIP_SEED=172.30.240.11:2113,172.30.240.13:2113\n - KURRENTDB_REPLICATION_IP=172.30.240.12\n - KURRENTDB_CERTIFICATE_FILE=/etc/kurrentdb/certs/node2/node.crt\n - KURRENTDB_CERTIFICATE_PRIVATE_KEY_FILE=/etc/kurrentdb/certs/node2/node.key\n - KURRENTDB_ADVERTISE_HOST_TO_CLIENT_AS=127.0.0.1\n - KURRENTDB_ADVERTISE_NODE_PORT_TO_CLIENT_AS=2112\n ports:\n - 2112:2113\n networks:\n clusternetwork:\n ipv4_address: 172.30.240.12\n volumes:\n - ./certs:/etc/kurrentdb/certs\n restart: unless-stopped\n depends_on:\n - cert-gen\n\n esdb-node3:\n build: \n context: .\n args:\n RUNTIME: ${RUNTIME}\n env_file:\n - shared.env\n environment:\n - KURRENTDB_GOSSIP_SEED=172.30.240.11:2113,172.30.240.12:2113\n - KURRENTDB_REPLICATION_IP=172.30.240.13\n - KURRENTDB_CERTIFICATE_FILE=/etc/kurrentdb/certs/node3/node.crt\n - KURRENTDB_CERTIFICATE_PRIVATE_KEY_FILE=/etc/kurrentdb/certs/node3/node.key\n - KURRENTDB_ADVERTISE_HOST_TO_CLIENT_AS=127.0.0.1\n - KURRENTDB_ADVERTISE_NODE_PORT_TO_CLIENT_AS=2113\n ports:\n - 2113:2113\n networks:\n clusternetwork:\n ipv4_address: 172.30.240.13\n volumes:\n - ./certs:/etc/kurrentdb/certs\n restart: unless-stopped\n depends_on:\n - cert-gen\n\nnetworks:\n clusternetwork:\n name: kurrentdb.local\n driver: bridge\n ipam:\n driver: default\n config:\n - subnet: 172.30.240.0/24\n
dataset_sample\yaml\kurrent-io_KurrentDB\docker-compose.yml
docker-compose.yml
YAML
3,193
0.7
0
0
node-utils
867
2024-01-29T23:43:24.582950
BSD-3-Clause
false
ac5132d8d6fbe511ab5cce5992db6feb
php:\n preset: laravel\n version: 8.2\n finder:\n not-name:\n - bad-syntax-strategy.php\njs:\n finder:\n exclude:\n - src/Illuminate/Foundation/resources/exceptions/renderer/dist\n not-name:\n - webpack.mix.js\ncss:\n finder:\n exclude:\n - src/Illuminate/Foundation/resources/exceptions/renderer/dist\n
dataset_sample\yaml\laravel_framework\.styleci.yml
.styleci.yml
YAML
324
0.7
0
0
python-kit
380
2024-06-13T00:33:43.941916
Apache-2.0
false
c3e44af62a8e6a2e1d3ae457a79a5197
version: '3'\nservices:\n dynamodb:\n image: amazon/dynamodb-local:2.0.0\n ports:\n - "8000:8000"\n command: ["-jar", "DynamoDBLocal.jar", "-sharedDb", "-inMemory"]\n memcached:\n image: memcached:1.6-alpine\n ports:\n - "11211:11211"\n restart: always\n mysql:\n image: mysql/mysql-server:5.7\n environment:\n MYSQL_ALLOW_EMPTY_PASSWORD: 1\n MYSQL_ROOT_PASSWORD: ""\n MYSQL_DATABASE: "forge"\n MYSQL_ROOT_HOST: "%"\n ports:\n - "3306:3306"\n restart: always\n # postgres:\n # image: postgres:15\n # environment:\n # POSTGRES_PASSWORD: "secret"\n # POSTGRES_DB: "forge"\n # ports:\n # - "5432:5432"\n # restart: always\n # mariadb:\n # image: mariadb:11\n # environment:\n # MARIADB_ALLOW_EMPTY_ROOT_PASSWORD: "yes"\n # MARIADB_ROOT_PASSWORD: ""\n # MARIADB_DATABASE: "forge"\n # MARIADB_ROOT_HOST: "%"\n # ports:\n # - "3306:3306"\n # restart: always\n # mssql:\n # image: mcr.microsoft.com/mssql/server:2019-latest\n # environment:\n # ACCEPT_EULA: "Y"\n # SA_PASSWORD: "Forge123"\n # ports:\n # - "1433:1433"\n # restart: always\n redis:\n image: redis:7.0-alpine\n ports:\n - "6379:6379"\n restart: always\n # redis-cluster-0:\n # image: redis:7.0-alpine\n # ports:\n # - "7000:7000"\n # restart: always\n # command: redis-server --port 7000 --appendonly yes --cluster-enabled yes\n # redis-cluster-1:\n # image: redis:7.0-alpine\n # ports:\n # - "7001:7001"\n # restart: always\n # command: redis-server --port 7001 --appendonly yes --cluster-enabled yes\n # redis-cluster-2:\n # image: redis:7.0-alpine\n # ports:\n # - "7002:7002"\n # restart: always\n # command: redis-server --port 7002 --appendonly yes --cluster-enabled yes\n # redis-cluster-creator:\n # image: redis:7.0-alpine\n # depends_on:\n # - redis-cluster-0\n # - redis-cluster-1\n # - redis-cluster-2\n # command: sh -c 'redis-cli --cluster create redis-cluster-0:7000 redis-cluster-1:7001 redis-cluster-2:7002 --cluster-replicas 0 --cluster-yes || true'\n # restart: no\n
dataset_sample\yaml\laravel_framework\docker-compose.yml
docker-compose.yml
YAML
2,123
0.8
0
0.658228
node-utils
149
2025-06-01T10:04:48.390978
BSD-3-Clause
false
6e499f4e2c8169e26fac381ac03baa19
namespace: Tests\nsupport_namespace: Support\npaths:\n tests: tests\n output: tests/_output\n data: tests/Support/Data\n support: tests/Support\nactor_suffix: Tester\nextensions:\n enabled:\n - Codeception\Extension\RunFailed\nsettings:\n memory_limit: 1024M\n colors: true\n error_level: E_ALL & ~E_DEPRECATED\n\ncoverage:\n enabled: true\n include:\n - app/*\n\n
dataset_sample\yaml\Leantime_leantime\codeception.yml
codeception.yml
YAML
359
0.8
0
0
node-utils
956
2024-02-22T12:09:29.237395
GPL-3.0
false
491920a2ed0d87e78d9da1cc00615cf8
changelog_type: 'commit_message'\nheader_prefix: 'Version:'\ncommit_changelog: true\ncomment_changelog: true\ninclude_unlabeled_changes: true\nunlabeled_group_title: 'Unlabeled Changes'\npull_request_title_regex: '^Release'\nversion_regex: 'v?([0-9]{1,2})+[.]+([0-9]{1,2})+[.]+([0-9]{1,2})\s\(\d{1,2}-\d{1,2}-\d{4}\)'\nexclude_labels:\n - bot\n - dependabot\n - ci\ngroup_config:\n - title: Bug Fixes\n labels:\n - bug\n - bugfix\n - fix\n - title: Code Improvements\n labels:\n - improvements\n - enhancement\n - cleanup\n - title: New Features\n labels:\n - feature\n - title: Documentation Updates\n labels:\n - docs\n - documentation\n - doc\n
dataset_sample\yaml\Leantime_leantime\.github\changelogConfig.yml
changelogConfig.yml
YAML
685
0.7
0
0
python-kit
725
2024-10-30T10:43:13.753615
Apache-2.0
false
f14657ff7603b3b97654be1ba6411a42
changelog:\n exclude:\n labels:\n - ignore-for-release\n categories:\n - title: 🛠 Breaking Changes \n labels:\n - Breaking-Change\n - title: 🎉 New Features \n labels:\n - Feature\n - title: 🐞 Bug Fixes\n labels:\n - Bug\n - title: Other Changes\n labels:\n - "*"\n
dataset_sample\yaml\Leantime_leantime\.github\release.yml
release.yml
YAML
326
0.7
0.058824
0
react-lib
696
2025-02-01T07:13:47.188642
GPL-3.0
false
ef76964f63b248b07df5bde30087ca9a
name: Bug Report\ndescription: File a bug report.\ntitle: "Enter a title of your bug"\ntype: "Bug"\nassignees: \n - marcelfolaron\nbody:\n - type: markdown\n attributes:\n value: "Thank you for taking the time to submit a bug report. If you have problems with the installation or other questions please use the [discussion forum](https://github.com/Leantime/leantime/discussions) or our [discord server](https://discord.gg/4zMzJtAq9z)"\n - type: dropdown\n attributes:\n label: What is your set up? \n options:\n - Cloud Hosted\n - Self Hosted Server\n - Self Hosted Docker\n validations:\n required: true\n - type: input\n id: version\n attributes:\n label: Version\n description: "Which Leantime version are you using (see footer bottom right)"\n placeholder: "3.x.x"\n validations:\n required: true\n - type: textarea\n attributes:\n label: Describe the issue\n placeholder: A clear and concise description of what the bug is.\n validations:\n required: true\n - type: textarea\n id: repro\n attributes:\n label: Reproduction steps\n description: "How do you trigger this bug? Please walk us through it step by step."\n value: |\n 1.\n 2.\n 3.\n ...\n validations:\n required: true\n - type: textarea\n id: context\n attributes:\n label: Error Logs (LEANTIMEFOLDER/storage/logs)\n description: "Please past the contents of your error logs and any other logs you may have."\n validations:\n required: false\n
dataset_sample\yaml\Leantime_leantime\.github\ISSUE_TEMPLATE\bug_report.yml
bug_report.yml
YAML
1,546
0.95
0.019231
0
awesome-app
192
2024-04-07T22:29:27.207531
MIT
false
3c16c5a2ddf49f1e4fe0dec8963ecf9d
name: Feature request\ndescription: Suggest an idea for this project.\ntitle: "Enter title of your feature suggestion"\ntype: "Feature"\nassignees: \n - marcelfolaron\nbody:\n - type: markdown\n attributes:\n value: "Thank you for taking the time to submit a feature request. If you have problems with the installation or other questions please use the [discussion forum](https://github.com/Leantime/leantime/discussions) or our [discord server](https://discord.gg/4zMzJtAq9z)"\n - type: dropdown\n attributes:\n label: What is your set up? \n options:\n - Cloud Hosted\n - Self Hosted Server\n - Self Hosted Docker\n validations:\n required: true\n - type: textarea\n attributes:\n label: Is your feature request related to a problem? Please describe\n description: A clear and concise description of what the problem is. \n placeholder: Ex. I'm always frustrated when [...]\n - type: textarea\n attributes:\n label: Describe the solution you'd like\n description: A clear and concise description of what you want to happen.\n placeholder:\n validations:\n required: true\n - type: textarea\n attributes:\n label: Additional context\n description: "Add any other context or screenshots about the feature request here."\n validations:\n
dataset_sample\yaml\Leantime_leantime\.github\ISSUE_TEMPLATE\feature_request.yml
feature_request.yml
YAML
1,316
0.95
0.055556
0
react-lib
25
2023-07-26T11:58:16.588189
Apache-2.0
false
2f3588efd1820ff7c8b6663a6ebd0dac
name: Code Style Analysis\n\non:\n workflow_dispatch:\n push:\n branches: [ "master", "*.*-dev" ]\n pull_request:\n branches: [ "master", "*.*-dev" ]\n\njobs:\n pint:\n\n runs-on: ubuntu-24.04\n\n steps:\n - uses: actions/checkout@v3\n\n - name: Install Dependencies\n run: make build-dev\n\n - name: Run Laravel Pint Analysis\n run: make test-code-style\n
dataset_sample\yaml\Leantime_leantime\.github\workflows\codeStyleAnalysis.yml
codeStyleAnalysis.yml
YAML
369
0.7
0
0
vue-tools
962
2024-08-31T23:30:10.902031
BSD-3-Clause
false
b68cf74aebecef2463f9d08f081e95d0
name: Makefile CI\n\non:\n workflow_dispatch:\n push:\n branches: [ "master", "*.*-dev" ]\n pull_request:\n branches: [ "master", "*.*-dev" ]\n\njobs:\n build:\n\n runs-on: ubuntu-24.04\n\n steps:\n - uses: actions/checkout@v3\n\n - name: Install dependencies\n run: make build\n\n - name: Run distcheck\n run: make clean\n
dataset_sample\yaml\Leantime_leantime\.github\workflows\makefile.yml
makefile.yml
YAML
336
0.7
0
0
node-utils
6
2024-06-15T06:25:19.373909
GPL-3.0
false
d02b884e2917775ca7a06d89838070b8
name: Create Release\n\non:\n workflow_dispatch:\n push:\n tags:\n - '*'\n\njobs:\n build_release:\n name: build_release\n runs-on: ubuntu-24.04\n steps:\n - name: checkout\n uses: actions/checkout@v3\n with:\n fetch-depth: 0\n\n - name: version\n run: echo "version=$(make get-version)" >> $GITHUB_OUTPUT\n id: version\n\n - name: Get Hash\n id: hash\n run: echo "git_hash=$(git rev-parse --short "$GITHUB_SHA")" >> $GITHUB_OUTPUT\n - name: build artifacts\n run: make package\n\n - name: Create Changelog\n uses: saadmk11/changelog-ci@v1.1.2\n id: changelog_ci\n with:\n changelog_filename: CHANGELOG.md\n release_version: ${{ steps.version.outputs.version }}\n config_file: .github/changelogConfig.yml\n github_token: ${{ secrets.GITHUB_TOKEN }}\n \n - name: Extract Repo Attributes\n id: attrs\n uses: ibnesayeed/repo-attrs@master\n \n - name: release\n uses: ncipollo/release-action@v1\n id: create_release\n with:\n draft: false\n prerelease: false\n name: Leantime v${{ steps.version.outputs.version }}\n tag: v${{ steps.version.outputs.version }}\n generateReleaseNotes: false\n makeLatest: true\n body: |\n ## Changes in this Release\n\n History from `${{ steps.attrs.outputs.tail }}` to `${{ steps.attrs.outputs.head }}`\n\n ### Commits\n\n ${{ steps.attrs.outputs.commits }}\n\n ### Contributors\n\n ${{ steps.attrs.outputs.contributors }}\n\n ### Files\n\n ```\n ${{ steps.attrs.outputs.files }}\n ```\n env:\n GITHUB_TOKEN: ${{ github.token }}\n \n - name: upload zip\n uses: actions/upload-release-asset@v1\n env:\n GITHUB_TOKEN: ${{ github.token }}\n with:\n upload_url: ${{ steps.create_release.outputs.upload_url }}\n asset_path: ./target/Leantime-v${{ steps.version.outputs.version }}.zip\n asset_name: Leantime-v${{ steps.version.outputs.version }}.zip\n asset_content_type: application/gzip\n \n - name: upload tar\n uses: actions/upload-release-asset@v1\n env:\n GITHUB_TOKEN: ${{ github.token }}\n with:\n upload_url: ${{ steps.create_release.outputs.upload_url }}\n asset_path: ./target/Leantime-v${{ steps.version.outputs.version }}.tar.gz\n asset_name: Leantime-v${{ steps.version.outputs.version }}.tar.gz\n asset_content_type: application/gzip\n \n - name: Run latest-tag\n uses: EndBug/latest-tag@latest\n with:\n # You can change the name of the tag or branch with this input.\n # Default: 'latest'\n ref: latest\n \n # If a description is provided, the action will use it to create an annotated tag. If none is given, the action will create a lightweight tag.\n # Default: ''\n description: Latest Release of Leantime. \n
dataset_sample\yaml\Leantime_leantime\.github\workflows\release.yml
release.yml
YAML
3,090
0.8
0
0.095238
react-lib
19
2023-11-10T20:07:14.642944
GPL-3.0
false
7cf63cac91edaef87bf392a27b9266fa
name: Static Analysis\n\non:\n workflow_dispatch:\n push:\n branches: [ "master", "*.*-dev" ]\n pull_request:\n branches: [ "master", "*.*-dev" ]\n\njobs:\n phpstan:\n runs-on: ubuntu-24.04\n steps:\n - uses: actions/checkout@v3\n\n - name: Install Dependencies\n run: make build-dev\n\n - name: Run PHPStan Analysis\n run: make phpstan\n
dataset_sample\yaml\Leantime_leantime\.github\workflows\staticAnalysis.yml
staticAnalysis.yml
YAML
353
0.7
0
0
node-utils
860
2025-02-02T09:23:37.310478
BSD-3-Clause
false
3e9fe79483d18c099f0e03f5c0c6692a
# Codeception Test Suite Configuration\n#\n# Suite for acceptance tests.\n# Perform tests in browser using the WebDriver or PhpBrowser.\n# If you need both WebDriver and PHPBrowser tests - create a separate suite.\nactor: AcceptanceTester\nbootstrap: bootstrap.php\nmodules:\n enabled:\n - \Tests\Support\Helper\Acceptance\n - REST:\n url: 'https://leantime-dev'\n depends: PhpBrowser\n - WebDriver:\n url: 'https://leantime-dev'\n host: selenium\n browser: chrome\n wait: 20\n log_js_errors: true\n window_size: "2560x1440"\n capabilities:\n acceptInsecureCerts: true\n goog:chromeOptions:\n args: [ "--headless" ]\n - Db:\n dsn: "mysql:host=leantime-db;port=3306;dbname=leantime_test"\n user: root\n password: leantime\n populate: true\n cleanup: true\n\nstep_decorators:\n - Codeception\Step\ConditionalAssertion\n - Codeception\Step\TryTo\n - Codeception\Step\Retry\n
dataset_sample\yaml\Leantime_leantime\tests\Acceptance.suite.yml
Acceptance.suite.yml
YAML
979
0.8
0.028571
0.147059
react-lib
252
2025-01-23T12:46:37.526697
MIT
true
66f07b013ee60b4b099493bb96389980
# Codeception Test Suite Configuration\n\n# The actor attribute identifies the testing scenario. In this case, it's a `UnitTester`\nactor: UnitTester\n\nerror_level: E_ALL & ~E_STRICT & ~E_DEPRECATED\n\nmodules:\n enabled:\n - Asserts\n\ncoverage:\n enabled: true\n
dataset_sample\yaml\Leantime_leantime\tests\Unit.suite.yml
Unit.suite.yml
YAML
257
0.8
0
0.222222
node-utils
275
2024-03-23T02:14:23.558181
GPL-3.0
true
dacb6354493a253782891cccf3aad5a9
build:\n environment:\n php: '5.6.0'\n\nbefore_commands:\n - "composer install --prefer-source"\n\ntools:\n external_code_coverage:\n enabled: true\n timeout: 300\n filter:\n excluded_paths:\n - 'docs/*'\n - 'examples/*'\n - 'extras/*'\n - 'test/*'\n - 'vendor/*'\n\n php_code_coverage:\n enabled: false\n filter:\n excluded_paths:\n - 'docs/*'\n - 'examples/*'\n - 'extras/*'\n - 'test/*'\n - 'vendor/*'\n\n php_code_sniffer:\n enabled: true\n config:\n standard: PSR2\n sniffs:\n generic:\n files:\n one_class_per_file_sniff: false\n filter:\n excluded_paths:\n - 'docs/*'\n - 'examples/*'\n - 'extras/*'\n - 'test/*'\n - 'vendor/*'\n\n # Copy/Paste Detector\n php_cpd:\n enabled: true\n excluded_dirs:\n - docs\n - examples\n - extras\n - test\n - vendor\n\n # PHP CS Fixer (http://http://cs.sensiolabs.org/).\n php_cs_fixer:\n enabled: true\n config:\n level: psr2\n filter:\n excluded_paths:\n - 'docs/*'\n - 'examples/*'\n - 'extras/*'\n - 'test/*'\n - 'vendor/*'\n\n # Analyzes the size and structure of a PHP project.\n php_loc:\n enabled: true\n excluded_dirs:\n - docs\n - examples\n - extras\n - test\n - vendor\n\n # PHP Mess Detector (http://phpmd.org).\n php_mess_detector:\n enabled: true\n config:\n rulesets:\n - codesize\n - unusedcode\n - naming\n - design\n naming_rules:\n short_variable: { minimum: 2 }\n filter:\n excluded_paths:\n - 'docs/*'\n - 'examples/*'\n - 'extras/*'\n - 'test/*'\n - 'vendor/*'\n\n # Analyzes the size and structure of a PHP project.\n php_pdepend:\n enabled: true\n excluded_dirs:\n - docs\n - examples\n - extras\n - test\n - vendor\n\n # Runs Scrutinizer's PHP Analyzer Tool\n # https://scrutinizer-ci.com/docs/tools/php/php-analyzer/config_reference\n php_analyzer:\n enabled: true\n config:\n checkstyle:\n enabled: true\n naming:\n enabled: true\n property_name: ^[_a-zA-Z][a-zA-Z0-9_]*$ #Allow underscores & caps\n method_name: ^(?:[_a-zA-Z]|__)[a-zA-Z0-9_]*$ #Allow underscores & caps\n parameter_name: ^[a-z][a-zA-Z0-9_]*$ # Allow underscores\n local_variable: ^[a-zA-Z][a-zA-Z0-9_]*$ #Allow underscores & caps\n exception_name: ^[a-zA-Z][a-zA-Z0-9]*Exception$\n isser_method_name: ^(?:[_a-zA-Z]|__)[a-zA-Z0-9]*$ #Allow underscores & caps\n filter:\n excluded_paths:\n - 'docs/*'\n - 'examples/*'\n - 'extras/*'\n - 'test/*'\n - 'vendor/*'\n\n # Security Advisory Checker\n sensiolabs_security_checker: true\n
dataset_sample\yaml\leokhoa_laragon\etc\pear\PHPMailer\.scrutinizer.yml
.scrutinizer.yml
YAML
3,627
0.8
0
0.066116
awesome-app
819
2025-05-03T00:07:24.532259
MIT
false
88da7dc05a597333f532affd20f0151c
language: php\nphp:\n - 7.0\n - 5.6\n - 5.5\n - 5.4\n - 5.3\n - hhvm\n \nmatrix:\n allow_failures:\n - php: hhvm\n\nbefore_install:\n - sudo apt-get update -qq\n - sudo apt-get install -y -qq postfix\nbefore_script:\n - sudo service postfix stop\n - smtp-sink -d "%d.%H.%M.%S" localhost:2500 1000 &\n - mkdir -p build/logs\n - cd test\n - cp testbootstrap-dist.php testbootstrap.php\n - chmod +x fakesendmail.sh\n - sudo mkdir -p /var/qmail/bin\n - sudo cp fakesendmail.sh /var/qmail/bin/sendmail\n - sudo cp fakesendmail.sh /usr/sbin/sendmail\n - echo 'sendmail_path = "/usr/sbin/sendmail -t -i "' > $(php --ini|grep -m 1 "ini files in:"|cut -d ":" -f 2)/sendmail.ini\nscript:\n - phpunit --configuration ../travis.phpunit.xml.dist\nafter_script:\n - wget https://scrutinizer-ci.com/ocular.phar\n - php ocular.phar code-coverage:upload --format=php-clover ../build/logs/clover.xml\n
dataset_sample\yaml\leokhoa_laragon\etc\pear\PHPMailer\.travis.yml
.travis.yml
YAML
876
0.8
0
0
python-kit
182
2024-01-02T20:19:56.588064
BSD-3-Clause
false
223fa0902945c58ec03b009f23091e47
name: Docker Compose\n\non:\n push:\n branches: [ main, 'devnet_*', 'testnet_*' ]\n pull_request:\n branches:\n - "**"\n paths:\n - 'docker/**'\n - 'Cargo.toml'\n - '.github/workflows/docker-compose.yml'\n - 'toolchains/**'\n workflow_dispatch:\n\n# This allows a subsequently queued workflow run to interrupt previous runs on pull-requests\nconcurrency:\n group: '${{ github.workflow }} @ ${{ github.event.pull_request.head.label || github.head_ref || github.run_id }}'\n cancel-in-progress: true\n\nenv:\n CARGO_TERM_COLOR: always\n CARGO_INCREMENTAL: 0\n CARGO_NET_RETRY: 10\n RUST_BACKTRACE: short\n # We allow redundant explicit links because `cargo rdme` doesn't know how to resolve implicit intra-crate links.\n RUSTDOCFLAGS: -A rustdoc::redundant_explicit_links -D warnings\n RUSTFLAGS: -D warnings\n RUSTUP_MAX_RETRIES: 10\n RUST_LOG: warn\n DOCKER_COMPOSE_WAIT: "true"\n\npermissions:\n contents: read\n\njobs:\n compose:\n runs-on: ubuntu-latest-16-cores\n timeout-minutes: 40\n\n steps:\n - uses: actions/checkout@v3\n with:\n fetch-depth: 0\n - uses: actions-rust-lang/setup-rust-toolchain@v1\n - name: Install Protoc\n uses: arduino/setup-protoc@v1\n with:\n repo-token: ${{ secrets.GITHUB_TOKEN }}\n # TODO(#2709): Uncomment once this workflow runs in a custom runner\n # - name: Update aio-max-nr\n # run: echo 1048576 > /proc/sys/fs/aio-max-nr\n - name: Build client binary\n run: |\n cargo install --path linera-service --bin linera --bin linera-server --debug\n # TODO(#2709): Remove this step once this workflow runs in a custom runner\n - name: Patch Docker compose file to run ScyllaDB in developer mode\n run: |\n sed -i -e '/SCYLLA_AUTO_CONF/{N;s/command:/ SCYLLA_ENABLE_EXPERIMENTAL: 1/p;N;N;N;N;d}' \\n docker/docker-compose.yml\n - name: Run Compose\n run: |\n cd docker\n ./compose.sh\n # TODO(#2709): Uncomment once this workflow runs in a custom runner\n # - name: Setup upterm session\n # uses: lhotari/action-upterm@v1\n - name: Sync\n uses: nick-fields/retry@v2\n with:\n # Docker compose can take some time to start\n max_attempts: 10\n timeout_minutes: 2\n retry_wait_seconds: 10\n command: linera --wallet docker/wallet.json --storage rocksdb:docker/linera.db sync-balance\n
dataset_sample\yaml\linera-io_linera-protocol\.github\workflows\docker-compose.yml
docker-compose.yml
YAML
2,436
0.8
0
0.144928
python-kit
843
2024-02-20T07:40:30.600292
GPL-3.0
false
71f2f6f4b13581472e53b8e95ab97660
name: Docker Image\n\non:\n push:\n branches: [ 'devnet_*', 'testnet_*' ]\n workflow_dispatch:\n\npermissions:\n contents: read\n\njobs:\n build-and-push:\n runs-on: ubuntu-latest\n timeout-minutes: 40\n\n steps:\n - uses: actions/checkout@v3\n with:\n fetch-depth: 0\n - uses: actions-rust-lang/setup-rust-toolchain@v1\n - name: Install Protoc\n uses: arduino/setup-protoc@v1\n with:\n repo-token: ${{ secrets.GITHUB_TOKEN }}\n - name: Auth service account\n uses: google-github-actions/auth@v1\n with:\n credentials_json: ${{ secrets.GCP_SA_ACTIONS_RUNNER_KEY }}\n - name: Set up Google Cloud SDK\n uses: google-github-actions/setup-gcloud@v1\n \n - name: Authenticate Docker with GCP\n run: |\n gcloud auth configure-docker us-docker.pkg.dev\n \n - name: Set env variables\n run: |\n echo "GIT_COMMIT_SHORT=${GITHUB_SHA:0:7}" >> $GITHUB_ENV\n echo "GIT_COMMIT_LONG=${GITHUB_SHA}" >> $GITHUB_ENV\n if [ -n "${{ github.head_ref }}" ]; then\n echo "BRANCH_NAME=${{ github.head_ref }}" >> $GITHUB_ENV\n else\n echo "BRANCH_NAME=${{ github.ref_name }}" >> $GITHUB_ENV\n fi\n\n - name: Set docker image env variable\n run: |\n echo "DOCKER_IMAGE=us-docker.pkg.dev/linera-io-dev/linera-public-registry/linera" >> $GITHUB_ENV\n \n - name: Build Docker image\n run: |\n docker build --build-arg git_commit=${{ env.GIT_COMMIT_LONG }} \\n -f docker/Dockerfile . \\n -t ${{ env.DOCKER_IMAGE }}:${{ env.BRANCH_NAME }} \\n -t ${{ env.DOCKER_IMAGE }}:${{ env.GIT_COMMIT_SHORT }} \\n -t ${{ env.DOCKER_IMAGE }}:${{ env.GIT_COMMIT_LONG }}\n \n - name: Push Docker image to Google Artifact Registry\n run: |\n docker push ${{ env.DOCKER_IMAGE }}:${{ env.BRANCH_NAME }}\n docker push ${{ env.DOCKER_IMAGE }}:${{ env.GIT_COMMIT_SHORT }}\n docker push ${{ env.DOCKER_IMAGE }}:${{ env.GIT_COMMIT_LONG }}\n
dataset_sample\yaml\linera-io_linera-protocol\.github\workflows\docker_image.yml
docker_image.yml
YAML
2,065
0.7
0.016129
0
vue-tools
898
2024-10-09T19:44:00.615050
BSD-3-Clause
false
fcd4c29e80ce7fdf6e6bc9849ee0edf2
name: Documentation\n\non:\n push:\n branches: [ main, 'devnet_*', 'testnet_*' ]\n paths-ignore:\n - '*.md'\n - 'docker/**'\n - 'docker_scylla/**'\n - 'configuration/**'\n - 'kubernetes/**'\n - 'scripts/**'\n workflow_dispatch:\n\n# This allows a subsequently queued workflow run to interrupt previous runs on pull-requests\nconcurrency:\n group: '${{ github.workflow }} ${{ github.ref }}'\n cancel-in-progress: true\n\nenv:\n CARGO_TERM_COLOR: always\n CARGO_INCREMENTAL: 0\n CARGO_NET_RETRY: 10\n RUST_BACKTRACE: short\n RUSTFLAGS: "-D warnings"\n RUSTUP_MAX_RETRIES: 10\n LINERA_PACKAGES: "packages.txt"\n\npermissions:\n contents: read\n\njobs:\n\n publish-docs:\n runs-on: ubuntu-latest\n timeout-minutes: 20\n\n # We only publish docs for the main branch.\n if: github.ref == 'refs/heads/main'\n\n steps:\n - uses: actions/checkout@v3\n - uses: actions-rust-lang/setup-rust-toolchain@v1\n - name: Install Protoc\n uses: arduino/setup-protoc@v1\n with:\n repo-token: ${{ secrets.GITHUB_TOKEN }}\n - name: Compute docs\n run: |\n cargo doc --all-features\n - name: Deploy\n uses: peaceiris/actions-gh-pages@v3\n with:\n deploy_key: ${{ secrets.ACTIONS_DEPLOY_KEY }}\n publish_branch: gh-pages\n publish_dir: ./target/doc\n force_orphan: true\n\n test-crates-and-docrs:\n runs-on: ubuntu-latest\n timeout-minutes: 90\n\n steps:\n - uses: actions/checkout@v3\n - uses: actions-rust-lang/setup-rust-toolchain@v1\n - name: Install Protoc\n uses: arduino/setup-protoc@v1\n with:\n repo-token: ${{ secrets.GITHUB_TOKEN }}\n - name: Install cargo tools\n run: |\n cargo install --locked cargo-index\n cargo install --locked cargo-local-registry\n - name: Setup git committer information\n run: |\n git config --global user.email "ci@linera.io"\n git config --global user.name "CI"\n - name: Generate packages\n run: |\n scripts/test_publish.sh "$LINERA_PACKAGES" ${{ runner.temp }}/registry\n # TODO(#2593): fix docs.rs CI test\n # - name: Extract packaged crates\n # run: |\n # cd "${{ runner.temp }}/registry"\n # grep -v '^#' "$LINERA_PACKAGES" | while read LINE; do\n # CRATE="${LINE%% *}"\n # tar -xzvf "${CRATE}-*.crate"\n # done\n # - name: Checkout docs.rs tool repository\n # run: |\n # cd ${{ runner.temp }}\n # git clone https://github.com/rust-lang/docs.rs docsrs\n # cd docsrs\n # git submodule update --init\n # - name: Build documentation using docs.rs\n # run: |\n # cd ${{ runner.temp }}/docsrs\n # cp .env.sample .env\n # mkdir -p ignored/cratesfyi-prefix/crates.io-index\n # . .env\n # SQLX_OFFLINE=1 cargo build\n # docker compose up -d db s3\n # cargo run -- database migrate\n # grep -v ^# "$LINERA_PACKAGES" | while read LINE; do\n # CRATE="${LINE%% *}"\n # cargo run -- build crate --local "${{ runner.temp }}/registry/${CRATE}-*/"\n # done\n
dataset_sample\yaml\linera-io_linera-protocol\.github\workflows\documentation.yml
documentation.yml
YAML
3,080
0.8
0.037383
0.298969
node-utils
124
2024-04-24T12:26:00.469762
MIT
false
549d38cb02e66f18f9ceda1b1656f941
name: DynamoDB tests\n\non:\n push:\n branches: [ main ]\n pull_request:\n branches:\n - "**"\n paths:\n - '.github/workflows/dynamodb.yml'\n - 'toolchains/**'\n - 'linera-views/**'\n - 'linera-storage/**'\n workflow_dispatch:\n\n# This allows a subsequently queued workflow run to interrupt previous runs on pull-requests\nconcurrency:\n group: '${{ github.workflow }} @ ${{ github.event.pull_request.head.label || github.head_ref || github.run_id }}'\n cancel-in-progress: true\n\nenv:\n CARGO_TERM_COLOR: always\n CARGO_INCREMENTAL: 0\n CARGO_NET_RETRY: 10\n RUST_BACKTRACE: short\n RUSTFLAGS: "-D warnings"\n RUSTUP_MAX_RETRIES: 10\n RUST_LOG: warn\n\npermissions:\n contents: read\n\njobs:\n\n test:\n runs-on: ubuntu-latest-8-cores\n timeout-minutes: 40\n\n steps:\n - uses: actions/checkout@v3\n - uses: actions-rust-lang/setup-rust-toolchain@v1\n - name: Install Protoc\n uses: arduino/setup-protoc@v1\n with:\n repo-token: ${{ secrets.GITHUB_TOKEN }}\n - name: Build example applications\n run: |\n cd examples\n cargo build --locked --release --target wasm32-unknown-unknown\n - name: Build\n run: |\n cargo build --locked --features dynamodb\n - name: Setup local DynamoDB instance\n run: |\n docker run --rm -d --name local-dynamodb -p 8000:8000/tcp amazon/dynamodb-local\n - name: Run DynamoDB tests\n env:\n AWS_REGION: us-east-1\n AWS_ACCESS_KEY_ID: test\n AWS_SECRET_ACCESS_KEY: test\n DYNAMODB_LOCAL_ENDPOINT: http://localhost:8000\n run: |\n cargo test --locked --features dynamodb -- dynamo\n
dataset_sample\yaml\linera-io_linera-protocol\.github\workflows\dynamodb.yml
dynamodb.yml
YAML
1,631
0.8
0
0.017857
node-utils
737
2024-06-26T07:38:31.565877
MIT
false
155c915e53d5dd4bdfa64a6b3ea6686d
name: Explorer tests\n\non:\n push:\n branches: [ main ]\n pull_request:\n branches:\n - "**"\n paths-ignore:\n - 'CONTRIBUTING.md'\n - 'INSTALL.md'\n - 'docker/**'\n - 'kubernetes/**'\n workflow_dispatch:\n\n# This allows a subsequently queued workflow run to interrupt previous runs on pull-requests\nconcurrency:\n group: '${{ github.workflow }} @ ${{ github.event.pull_request.head.label || github.head_ref || github.run_id }}'\n cancel-in-progress: true\n\nenv:\n CARGO_TERM_COLOR: always\n CARGO_INCREMENTAL: 0\n CARGO_NET_RETRY: 10\n RUST_BACKTRACE: short\n RUSTFLAGS: "-D warnings"\n RUSTUP_MAX_RETRIES: 10\n\npermissions:\n contents: read\n\njobs:\n\n test:\n runs-on: ubuntu-latest\n timeout-minutes: 10\n\n steps:\n - uses: actions/checkout@v3\n - uses: actions/setup-node@v3\n - uses: actions-rust-lang/setup-rust-toolchain@v1\n - name: Check toolchain symlinks\n run: |\n cd linera-explorer\n cat rust-toolchain.toml\n - name: Build\n run: |\n cd linera-explorer\n npm cache clean --force\n npm ci\n npm run full\n - name: Test\n run: |\n cd linera-explorer\n npm run test -- --run\n
dataset_sample\yaml\linera-io_linera-protocol\.github\workflows\explorer.yml
explorer.yml
YAML
1,186
0.8
0
0.020833
awesome-app
854
2025-06-08T06:27:53.583397
MIT
false
863d438ace5c60b90a7006d8f11a2f54
name: Performance Summary\n\non:\n workflow_run:\n workflows:\n - Rust\n types:\n - completed\n workflow_dispatch:\n\n# This allows a subsequently queued workflow run to interrupt previous runs on pull-requests\nconcurrency:\n group: '${{ github.workflow }} @ ${{ github.head_ref || github.event.workflow_run.head_commit.id || github.event.workflow_run.id || github.run_id }}'\n cancel-in-progress: true\n\npermissions:\n contents: read\n pull-requests: write\n\njobs:\n performance-summary:\n runs-on: ubuntu-latest\n if: ${{ github.event.workflow_run.conclusion == 'success' }}\n steps:\n - uses: actions/checkout@v3\n - uses: actions-rust-lang/setup-rust-toolchain@v1\n - name: Post Performance Summary comment on PR\n if: ${{ github.event.workflow_run.event == 'pull_request' }}\n env:\n GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}\n GITHUB_REPOSITORY: ${{ github.repository }}\n GITHUB_PR_NUMBER: ${{ github.event.workflow_run.pull_requests[0].number }}\n GITHUB_BASE_BRANCH: ${{ github.event.workflow_run.pull_requests[0].base.ref}}\n GITHUB_PR_BRANCH: ${{ github.event.workflow_run.pull_requests[0].head.ref }}\n GITHUB_PR_COMMIT_HASH: ${{ github.event.workflow_run.pull_requests[0].head.sha }}\n # The list of workflows that trigger this need to be hardcoded above.\n # Make sure you have the same comma separated list of workflows here.\n run: cargo run -p linera-summary -- --workflows "Rust" ci >> $GITHUB_STEP_SUMMARY\n
dataset_sample\yaml\linera-io_linera-protocol\.github\workflows\performance_summary.yml
performance_summary.yml
YAML
1,527
0.8
0.052632
0.088235
awesome-app
705
2025-04-11T05:05:53.494956
MIT
false
04a5e3b5bfc79785f0dc614752c463b2
name: Create Network Release\n\non:\n push:\n tags:\n - 'testnet-*-v[0-9]+.[0-9]+.[0-9]+'\n - 'devnet-*-v[0-9]+.[0-9]+.[0-9]+'\n\njobs:\n release:\n runs-on: ubuntu-latest\n steps:\n - uses: actions/checkout@v4\n\n - name: Extract Tag Name\n id: tag\n run: echo "TAG_NAME=${GITHUB_REF#refs/tags/}" >> $GITHUB_OUTPUT\n\n - name: Create GitHub Release\n uses: softprops/action-gh-release@v2\n with:\n tag_name: ${{ steps.tag.outputs.TAG_NAME }}\n name: ${{ steps.tag.outputs.TAG_NAME }}\n generate_release_notes: false\n env:\n GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}\n
dataset_sample\yaml\linera-io_linera-protocol\.github\workflows\release.yml
release.yml
YAML
650
0.8
0
0
awesome-app
717
2023-11-25T00:40:11.657033
Apache-2.0
false
1091fc8943812d3996d86a8e333a511a
name: Rust\n\non:\n push:\n branches: [ main, 'devnet_*', 'testnet_*' ]\n pull_request:\n branches:\n - "**"\n paths-ignore:\n - 'CONTRIBUTING.md'\n - 'INSTALL.md'\n - 'docker/**'\n - 'docker_scylla/**'\n - 'configuration/**'\n - 'kubernetes/**'\n workflow_dispatch:\n\n# This allows a subsequently queued workflow run to interrupt previous runs on pull-requests\nconcurrency:\n group: '${{ github.workflow }} @ ${{ github.event.pull_request.head.label || github.head_ref || github.run_id }}'\n cancel-in-progress: true\n\nenv:\n CARGO_TERM_COLOR: always\n CARGO_INCREMENTAL: 0\n CARGO_NET_RETRY: 10\n RUST_BACKTRACE: full\n # We allow redundant explicit links because `cargo rdme` doesn't know how to resolve implicit intra-crate links.\n RUSTDOCFLAGS: -A rustdoc::redundant_explicit_links -D warnings\n RUSTFLAGS: -D warnings\n RUSTUP_MAX_RETRIES: 10\n RUST_LOG: linera=debug\n RUST_LOG_FORMAT: plain\n LINERA_STORAGE_SERVICE: 127.0.0.1:1235\n LINERA_WALLET: /tmp/local-linera-net/wallet_0.json\n LINERA_STORAGE: rocksdb:/tmp/local-linera-net/client_0.db\n LINERA_FAUCET_URL: http://localhost:8079\n\npermissions:\n contents: read\n\njobs:\n remote-net-test:\n runs-on: ubuntu-latest-8-cores\n timeout-minutes: 40\n\n steps:\n - uses: actions/checkout@v3\n - uses: actions-rust-lang/setup-rust-toolchain@v1\n - name: Install Protoc\n uses: arduino/setup-protoc@v1\n with:\n repo-token: ${{ secrets.GITHUB_TOKEN }}\n - name: Run the storage-service instance\n run: |\n cargo run --release -p linera-storage-service -- memory --endpoint $LINERA_STORAGE_SERVICE &\n - name: Run the validators\n run: |\n cargo build --features storage-service\n mkdir /tmp/local-linera-net\n cargo run --features storage-service --bin linera -- net up --storage service:tcp:$LINERA_STORAGE_SERVICE:table --policy-config testnet --path /tmp/local-linera-net --validators 4 --shards 4 &\n - name: Create two epochs and run the faucet\n run: |\n cargo build --bin linera\n cargo run --bin linera -- resource-control-policy --block 0.0000001\n cargo run --bin linera -- resource-control-policy --block 0.000000\n cargo run --bin linera -- faucet --amount 1000 --port 8079 a3edc33d8e951a1139333be8a4b56646b5598a8f51216e86592d881808972b07 &\n - name: Run the remote-net tests\n run: |\n cargo test -p linera-service remote_net_grpc --features remote-net\n\n execution-wasmtime-test:\n runs-on: ubuntu-latest\n timeout-minutes: 10\n\n steps:\n - uses: actions/checkout@v3\n - uses: actions-rust-lang/setup-rust-toolchain@v1\n - name: Run some extra execution tests with wasmtime\n run: |\n cargo test --locked -p linera-execution --features wasmtime\n\n metrics-test:\n runs-on: ubuntu-latest\n timeout-minutes: 10\n\n steps:\n - uses: actions/checkout@v3\n - uses: actions-rust-lang/setup-rust-toolchain@v1\n - name: Run metrics tests\n run: |\n cargo test --locked -p linera-base --features metrics\n\n wasm-application-test:\n runs-on: ubuntu-latest\n timeout-minutes: 15\n\n steps:\n - uses: actions/checkout@v3\n - uses: actions-rust-lang/setup-rust-toolchain@v1\n - name: Install Protoc\n uses: arduino/setup-protoc@v1\n with:\n repo-token: ${{ secrets.GITHUB_TOKEN }}\n - name: Run Wasm application tests\n run: |\n cd examples\n cargo test --locked\n\n default-features-and-witty-integration-test:\n runs-on: ubuntu-latest\n timeout-minutes: 40\n\n steps:\n - uses: actions/checkout@v3\n - uses: actions-rust-lang/setup-rust-toolchain@v1\n - name: Install Protoc\n uses: arduino/setup-protoc@v1\n with:\n repo-token: ${{ secrets.GITHUB_TOKEN }}\n - name: Compile Wasm test modules for Witty integration tests\n run: |\n cargo build -p linera-witty-test-modules --target wasm32-unknown-unknown\n - name: Run all tests using the default features (except storage-service)\n run: |\n # TODO(#2764): Actually link this to the default features\n cargo test --no-default-features --features fs,macros,wasmer,rocksdb --locked\n - name: Run Witty integration tests\n run: |\n cargo test -p linera-witty --features wasmer,wasmtime\n\n check-outdated-cli-md:\n runs-on: ubuntu-latest\n timeout-minutes: 10\n\n steps:\n - uses: actions/checkout@v3\n - uses: actions-rust-lang/setup-rust-toolchain@v1\n - name: Install Protoc\n uses: arduino/setup-protoc@v1\n with:\n repo-token: ${{ secrets.GITHUB_TOKEN }}\n - name: Check for outdated CLI.md\n run: |\n if ! diff CLI.md <(cargo run --bin linera -- help-markdown)\n then\n echo '`CLI.md` differs from the output of `linera help-markdown`'\n echo 'Run `linera help-markdown > CLI.md` to update it.'\n exit 1\n fi\n\n benchmark-test:\n runs-on: ubuntu-latest\n timeout-minutes: 30\n\n steps:\n - uses: actions/checkout@v3\n - uses: actions-rust-lang/setup-rust-toolchain@v1\n - name: Install Protoc\n uses: arduino/setup-protoc@v1\n with:\n repo-token: ${{ secrets.GITHUB_TOKEN }}\n - name: Run the storage-service instance\n run: |\n cargo run --release -p linera-storage-service -- memory --endpoint $LINERA_STORAGE_SERVICE &\n - name: Run the benchmark test\n run: |\n cargo build --locked -p linera-service --bin linera-benchmark --features benchmark,storage-service\n cargo test --locked -p linera-service --features benchmark,storage-service benchmark\n\n ethereum-tests:\n runs-on: ubuntu-latest\n timeout-minutes: 30\n\n steps:\n - uses: actions/checkout@v3\n - uses: actions-rust-lang/setup-rust-toolchain@v1\n - uses: foundry-rs/foundry-toolchain@v1.2.0\n - name: Ensure Solc Directory Exists\n run: mkdir -p /home/runner/.solc\n - name: Cache Solc\n id: cache-solc\n uses: actions/cache@v4\n with:\n path: /home/runner/.solc\n key: solc-v0.8.25\n restore-keys: solc-\n - name: Get Solc\n if: ${{ steps.cache-solc.outputs.cache-hit != 'true' }}\n uses: pontem-network/get-solc@master\n with:\n version: v0.8.25\n - name: Add Solc to PATH\n run: echo "/home/runner/.solc" >> $GITHUB_PATH\n - name: List the content of the Solc cache\n run: ls /home/runner/.solc/v0.8.25/solc-static-linux\n - name: Create a Solc symbolic link\n run: ln -sf /home/runner/.solc/v0.8.25/solc-static-linux /home/runner/.solc/solc\n - name: Check if Solc is in PATH\n run: which solc || echo "Solc not found in PATH"\n - name: Install Protoc\n uses: arduino/setup-protoc@v1\n with:\n repo-token: ${{ secrets.GITHUB_TOKEN }}\n - name: Run the storage-service instance\n run: |\n cargo run --release -p linera-storage-service -- memory --endpoint $LINERA_STORAGE_SERVICE &\n - name: Run Ethereum tests\n run: |\n cargo test -p linera-ethereum --features ethereum\n cargo test test_wasm_end_to_end_ethereum_tracker --features ethereum,storage-service\n - name: Run REVM test\n run: |\n cargo test -p linera-execution test_fuel_for_counter_revm_application --features revm\n cargo test test_evm_end_to_end_counter --features revm,storage-service\n\n storage-service-tests:\n runs-on: ubuntu-latest-16-cores\n timeout-minutes: 40\n steps:\n - uses: actions/checkout@v3\n - uses: actions-rust-lang/setup-rust-toolchain@v1\n - name: Install Protoc\n uses: arduino/setup-protoc@v1\n with:\n repo-token: ${{ secrets.GITHUB_TOKEN }}\n - name: Build example applications\n run: |\n cd examples\n cargo build --locked --release --target wasm32-unknown-unknown\n - name: Run the storage-service instance and the storage-service tests\n run: |\n cargo run --release -p linera-storage-service -- memory --endpoint $LINERA_STORAGE_SERVICE &\n cargo test --features storage-service -- storage_service --nocapture\n\n test-readme-scripts:\n runs-on: ubuntu-latest-16-cores\n timeout-minutes: 40\n steps:\n - uses: actions/checkout@v3\n - uses: actions-rust-lang/setup-rust-toolchain@v1\n - name: Install Protoc\n uses: arduino/setup-protoc@v1\n with:\n repo-token: ${{ secrets.GITHUB_TOKEN }}\n - name: Build example applications\n run: |\n cd examples\n cargo build --locked --release --target wasm32-unknown-unknown\n cd ..\n cargo build --locked -p linera-storage-service\n - name: Run the storage-service instance and the README script tests\n run: |\n cargo test --locked -p linera-service --features storage-service -- test_script_in_readme --nocapture\n\n web:\n runs-on: ubuntu-latest\n timeout-minutes: 20\n\n steps:\n - uses: arduino/setup-protoc@v3\n with:\n repo-token: ${{ secrets.GITHUB_TOKEN }}\n - uses: actions/checkout@v3\n - name: Switch to nightly Rust toolchain\n run: |\n ln -sf toolchains/nightly/rust-toolchain.toml\n - uses: actions-rust-lang/setup-rust-toolchain@v1\n - name: Compile `linera-client` for the browser\n run: |\n cargo build -p linera-client \\n --locked \\n --target wasm32-unknown-unknown \\n --no-default-features \\n --features web-default\n - name: Install chromedriver\n uses: nanasess/setup-chromedriver@v2\n - name: Install wasm-pack\n uses: jetli/wasm-pack-action@v0.4.0\n with:\n version: 'latest'\n - name: Run the browser tests\n run: |\n cd linera-views\n WASM_BINDGEN_TEST_TIMEOUT=300 wasm-pack test --chrome --headless -- --features web-default\n\n check-wit-files:\n runs-on: ubuntu-latest-16-cores\n steps:\n - uses: actions/checkout@v3\n - uses: actions-rust-lang/setup-rust-toolchain@v1\n - name: Check WIT files\n run: |\n cargo run --bin wit-generator -- -c\n\n lint-unexpected-chain-load-operations:\n runs-on: ubuntu-latest\n timeout-minutes: 2\n\n steps:\n - uses: actions/checkout@v3\n - name: Check for unexpected chain load operations\n run: |\n ./scripts/check_chain_loads.sh\n\n lint-check-copyright-headers:\n runs-on: ubuntu-latest\n timeout-minutes: 2\n\n steps:\n - uses: actions/checkout@v3\n - name: Build check_copyright_header script\n run: |\n cd ./scripts/check_copyright_header\n cargo build --locked --release\n - name: Check Copyright headers\n run: >\n find linera-* examples -name '*.rs' -a -not -wholename '*/target/*' -print0\n | xargs -0 scripts/target/release/check_copyright_header\n\n lint-cargo-machete:\n runs-on: ubuntu-latest\n timeout-minutes: 2\n\n steps:\n - uses: actions/checkout@v3\n - name: Put lint toolchain file in place\n run: |\n ln -sf toolchains/nightly/rust-toolchain.toml\n - uses: actions-rust-lang/setup-rust-toolchain@v1\n - name: Install cargo-machete\n run: |\n cargo install cargo-machete@0.7.0 --locked\n - name: Check for unused dependencies\n run: |\n cargo machete\n\n lint-cargo-fmt:\n runs-on: ubuntu-latest\n timeout-minutes: 2\n\n steps:\n - uses: actions/checkout@v3\n - name: Put lint toolchain file in place\n run: |\n ln -sf toolchains/nightly/rust-toolchain.toml\n - uses: actions-rust-lang/setup-rust-toolchain@v1\n - name: Check formatting\n run: |\n cargo fmt -- --check\n\n lint-taplo-fmt:\n runs-on: ubuntu-latest\n timeout-minutes: 5\n\n steps:\n - uses: actions/checkout@v3\n - uses: actions-rust-lang/setup-rust-toolchain@v1\n - name: Install `taplo-cli`\n run: RUSTFLAGS='' cargo install taplo-cli@0.9.3 --locked\n - name: Check if `Cargo.toml` files are formatted\n run: taplo fmt --check --diff\n\n lint-check-for-outdated-readme:\n runs-on: ubuntu-latest\n timeout-minutes: 5\n\n steps:\n - uses: actions/checkout@v3\n - name: Put lint toolchain file in place\n run: |\n ln -sf toolchains/nightly/rust-toolchain.toml\n - uses: actions-rust-lang/setup-rust-toolchain@v1\n - name: Install cargo-rdme\n run: |\n cargo install cargo-rdme --locked\n - name: Check for outdated README.md\n run: |\n (set -e; for I in linera-*; do if [ -d "$I" ]; then echo $I; cargo rdme --check --no-fail-on-warnings -w $I; fi; done)\n\n lint-wasm-applications:\n runs-on: ubuntu-latest\n timeout-minutes: 10\n\n steps:\n - uses: actions/checkout@v3\n - name: Put lint toolchain file in place\n run: |\n ln -sf toolchains/nightly/rust-toolchain.toml\n - uses: actions-rust-lang/setup-rust-toolchain@v1\n - name: Install Protoc\n uses: arduino/setup-protoc@v1\n with:\n repo-token: ${{ secrets.GITHUB_TOKEN }}\n - name: Run Wasm application lints\n run: |\n cd examples\n cargo fmt -- --check\n cargo clippy --all-targets --all-features --target wasm32-unknown-unknown --locked\n cargo clippy --all-targets --all-features --target x86_64-unknown-linux-gnu --locked\n cd agent\n cargo fmt -- --check\n\n lint-cargo-clippy:\n runs-on: ubuntu-latest\n timeout-minutes: 20\n\n steps:\n - uses: actions/checkout@v3\n - name: Put lint toolchain file in place\n run: |\n ln -sf toolchains/nightly/rust-toolchain.toml\n - uses: actions-rust-lang/setup-rust-toolchain@v1\n - name: Install Protoc\n uses: arduino/setup-protoc@v1\n with:\n repo-token: ${{ secrets.GITHUB_TOKEN }}\n - name: Run clippy\n run: |\n cargo clippy --locked --all-targets --all-features\n cargo clippy --locked --no-default-features --features kubernetes\n cargo clippy --locked --no-default-features\n cargo clippy --locked --target wasm32-unknown-unknown --no-default-features --features web-default \\n -p linera-client \\n -p linera-rpc \\n -p linera-views\n\n lint-cargo-doc:\n runs-on: ubuntu-latest\n timeout-minutes: 20\n\n steps:\n - uses: actions/checkout@v3\n - name: Put lint toolchain file in place\n run: |\n ln -sf toolchains/nightly/rust-toolchain.toml\n - uses: actions-rust-lang/setup-rust-toolchain@v1\n - name: Install Protoc\n uses: arduino/setup-protoc@v1\n with:\n repo-token: ${{ secrets.GITHUB_TOKEN }}\n - name: Run cargo doc\n run: |\n cargo doc --locked --all-features\n\n lint-check-linera-service-graphql-schema:\n runs-on: ubuntu-latest\n timeout-minutes: 15\n\n steps:\n - uses: actions/checkout@v3\n - uses: actions-rust-lang/setup-rust-toolchain@v1\n - name: Install Protoc\n uses: arduino/setup-protoc@v1\n with:\n repo-token: ${{ secrets.GITHUB_TOKEN }}\n - name: Check linera-service GraphQL schema\n run: |\n diff <(cargo run --locked --bin linera-schema-export) linera-service-graphql-client/gql/service_schema.graphql\n\n lint-check-all-features:\n runs-on: ubuntu-latest-16-cores\n timeout-minutes: 40\n\n steps:\n - uses: actions/checkout@v3\n - name: Put lint toolchain file in place\n run: |\n ln -sf toolchains/nightly/rust-toolchain.toml\n - uses: actions-rust-lang/setup-rust-toolchain@v1\n - name: Install cargo-all-features\n run: |\n cargo install --git https://github.com/ma2bd/cargo-all-features --branch workspace_metadata --locked\n - name: Install Protoc\n uses: arduino/setup-protoc@v1\n with:\n repo-token: ${{ secrets.GITHUB_TOKEN }}\n - name: Run cargo check-all-features\n run: |\n cargo check-all-features --all-targets\n
dataset_sample\yaml\linera-io_linera-protocol\.github\workflows\rust.yml
rust.yml
YAML
15,583
0.8
0.027083
0.006928
node-utils
835
2023-12-11T01:53:32.370773
MIT
false
9e9fe771e6e715e0705f6e16a6f0f514
name: ScyllaDB tests\n\non:\n push:\n branches: [ main, 'devnet_*', 'testnet_*' ]\n pull_request:\n branches:\n - "**"\n paths:\n - '.github/workflows/scylladb.yml'\n - 'toolchains/**'\n - 'linera-views/**'\n - 'linera-storage/**'\n - 'docker_scylla/**'\n workflow_dispatch:\n\n# This allows a subsequently queued workflow run to interrupt previous runs on pull-requests\nconcurrency:\n group: '${{ github.workflow }} @ ${{ github.event.pull_request.head.label || github.head_ref || github.run_id }}'\n cancel-in-progress: true\n\nenv:\n CARGO_TERM_COLOR: always\n CARGO_INCREMENTAL: 0\n CARGO_NET_RETRY: 10\n RUST_BACKTRACE: short\n RUSTFLAGS: "-D warnings"\n RUSTUP_MAX_RETRIES: 10\n RUST_LOG: warn\n\npermissions:\n contents: read\n\njobs:\n\n test:\n runs-on: ubuntu-latest-8-cores\n timeout-minutes: 40\n\n steps:\n - uses: actions/checkout@v3\n - uses: actions-rust-lang/setup-rust-toolchain@v1\n - name: Install Protoc\n uses: arduino/setup-protoc@v1\n with:\n repo-token: ${{ secrets.GITHUB_TOKEN }}\n - name: Build example applications\n run: |\n cd examples\n cargo build --locked --release --target wasm32-unknown-unknown\n - name: Build\n run: |\n cargo build --locked --features scylladb\n - name: Setup local ScyllaDB instance\n run: |\n docker run --name my_scylla_container -d -p 9042:9042 scylladb/scylla:6.1\n - name: Run ScyllaDB tests\n run: |\n RUST_LOG=linera=info cargo test --locked --features scylladb -- scylla --nocapture\n
dataset_sample\yaml\linera-io_linera-protocol\.github\workflows\scylladb.yml
scylladb.yml
YAML
1,546
0.8
0
0.019231
node-utils
92
2024-09-15T07:31:59.879827
GPL-3.0
false
4f16ba898cd78457a042b78b571103d3
services:\n scylla:\n image: scylladb/scylla:6.1.3\n container_name: scylla\n volumes:\n - linera-scylla-data:/var/lib/scylla\n environment:\n SCYLLA_AUTO_CONF: 1\n command:\n - "--developer-mode"\n - "0"\n - "--overprovisioned"\n - "1"\n\n proxy:\n image: "${LINERA_IMAGE:-linera}"\n container_name: proxy\n ports:\n - "19100:19100"\n command: [ "./compose-proxy-entrypoint.sh" ]\n volumes:\n - .:/config\n labels:\n com.centurylinklabs.watchtower.enable: "true"\n depends_on:\n shard-init:\n condition: service_completed_successfully\n\n shard:\n image: "${LINERA_IMAGE:-linera}"\n deploy:\n replicas: 4\n command: [ "./compose-server-entrypoint.sh" ]\n volumes:\n - .:/config\n labels:\n com.centurylinklabs.watchtower.enable: "true"\n depends_on:\n shard-init:\n condition: service_completed_successfully\n\n shard-init:\n image: "${LINERA_IMAGE:-linera}"\n container_name: shard-init\n command: [ "./compose-server-init.sh" ]\n volumes:\n - .:/config\n depends_on:\n - scylla\n\n prometheus:\n image: prom/prometheus:latest\n container_name: prometheus\n volumes:\n - ./prometheus.yml:/etc/prometheus/prometheus.yml\n ports:\n - "9090:9090"\n\n grafana:\n image: grafana/grafana:latest\n container_name: grafana\n environment:\n - GF_SECURITY_ADMIN_PASSWORD=admin\n ports:\n - "3000:3000"\n volumes:\n - grafana-storage:/var/lib/grafana\n - ./provisioning/dashboards:/etc/grafana/provisioning/dashboards\n - ./dashboards:/var/lib/grafana/dashboards\n\n watchtower:\n image: containrrr/watchtower:latest\n container_name: watchtower\n volumes:\n - /var/run/docker.sock:/var/run/docker.sock\n command: [ "--interval", "30"]\n\nvolumes:\n linera-scylla-data:\n driver: local\n grafana-storage:\n
dataset_sample\yaml\linera-io_linera-protocol\docker\docker-compose.yml
docker-compose.yml
YAML
1,870
0.7
0
0
awesome-app
531
2024-10-31T20:12:30.985533
Apache-2.0
false
d5aeb4df3f9a7dd3e821d53c821abcb8
global:\n scrape_interval: 15s\n\nscrape_configs:\n - job_name: 'metrics-server'\n static_configs:\n - targets: ['proxy:21100', 'shard:21100']\n
dataset_sample\yaml\linera-io_linera-protocol\docker\prometheus.yml
prometheus.yml
YAML
147
0.7
0
0
vue-tools
386
2023-08-19T23:18:45.229648
BSD-3-Clause
false
47421750f997e1f7451a0235daab570d
apiVersion: 1\n\nproviders:\n - name: 'linera'\n orgId: 1\n folder: ''\n type: file\n disableDeletion: false\n updateIntervalSeconds: 10\n options:\n path: /var/lib/grafana/dashboards\n
dataset_sample\yaml\linera-io_linera-protocol\docker\provisioning\dashboards\dashboards.yml
dashboards.yml
YAML
198
0.7
0
0
python-kit
463
2025-05-23T22:40:22.846976
GPL-3.0
false
f535fba8ef4efa05608308936e21b358
version: 2\nupdates:\n\n - package-ecosystem: "github-actions"\n directory: "/"\n open-pull-requests-limit: 50\n schedule:\n interval: "daily"\n labels:\n - "sdou"\n - "dependencies"\n\n - package-ecosystem: "maven"\n directory: "/"\n open-pull-requests-limit: 50\n schedule:\n interval: "daily"\n pull-request-branch-name:\n # Separate sections of the branch name with a hyphen\n # for example, `dependabot-npm_and_yarn-next_js-acorn-6.4.1`\n separator: "-"\n ignore:\n - dependency-name: "actions/delete-package-versions"\n update-types: ["version-update:semver-major"] \n
dataset_sample\yaml\liquibase_liquibase\.github\dependabot.yml
dependabot.yml
YAML
627
0.8
0.041667
0.090909
awesome-app
292
2024-04-18T07:58:53.151669
BSD-3-Clause
false
7cac89b3175928d7d81f68eeaf36ab7c
blank_issues_enabled: false\ncontact_links:\n - name: Liquibase Forums\n url: https://forum.liquibase.org/\n about: Please use the forums to ask open-ended questions and connect with other Liquibase Users and Developers.\n - name: StackOverflow\n url: https://stackoverflow.com/questions/tagged/liquibase\n about: For specific questions about Liquibase that will have specific answers, please use StackOverflow and tag the question `liquibase`.\n
dataset_sample\yaml\liquibase_liquibase\.github\ISSUE_TEMPLATE\config.yml
config.yml
YAML
452
0.8
0
0
vue-tools
323
2023-10-23T19:56:38.863879
MIT
false
def764472b738b4536e017d81de93014
name: Build Azure Uber jar\n\non:\n workflow_call:\n inputs:\n branch:\n description: 'branch to check out'\n required: true\n type: string\n liquibase-version:\n description: 'liquibase version'\n required: true\n type: string\n workflow_dispatch:\n inputs:\n branch:\n description: 'branch to check out'\n required: true\n type: string\n liquibase-version:\n description: 'liquibase version'\n required: true\n type: string\n\nenv:\n MAVEN_VERSION: '3.9.2'\n\njobs:\n build:\n name: Build\n runs-on: ubuntu-22.04\n env:\n GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}\n steps:\n - name: Checkout liquibase-pro\n uses: actions/checkout@v4\n with:\n repository: liquibase/liquibase-pro\n ref: ${{ inputs.branch }}\n token: ${{ secrets.BOT_TOKEN }}\n path: liquibase-pro\n\n - name: Set up JDK 17\n uses: actions/setup-java@v4\n with:\n java-version: 17\n distribution: 'temurin'\n cache: 'maven'\n\n - name: maven-settings-xml-action\n uses: whelk-io/maven-settings-xml-action@v22\n with:\n repositories: |\n [\n {\n "id": "liquibase",\n "url": "https://maven.pkg.github.com/liquibase/liquibase",\n "releases": {\n "enabled": "false"\n },\n "snapshots": {\n "enabled": "true",\n "updatePolicy": "always"\n }\n },\n {\n "id": "liquibase-pro",\n "url": "https://maven.pkg.github.com/liquibase/liquibase-pro",\n "releases": {\n "enabled": "false"\n },\n "snapshots": {\n "enabled": "true",\n "updatePolicy": "always"\n }\n }\n ]\n plugin_repositories: |\n [\n {\n "id": "file-replacer-maven-plugin",\n "url": "https://maven.pkg.github.com/StevenMassaro/file-replacer-maven-plugin",\n "releases": {\n "updatePolicy": "always",\n "enabled": "true"\n },\n "snapshots": {\n "enabled": "true",\n "updatePolicy": "always"\n }\n }\n ]\n servers: |\n [\n {\n "id": "liquibase",\n "username": "liquibot",\n "password": "${{ secrets.LIQUIBOT_PAT }}"\n },\n {\n "id": "liquibase-pro",\n "username": "liquibot",\n "password": "${{ secrets.LIQUIBOT_PAT }}"\n },\n {\n "id": "file-replacer-maven-plugin",\n "username": "liquibot",\n "password": "${{ secrets.LIQUIBOT_PAT }}"\n }\n ]\n\n - name: Set up Maven\n uses: stCarolas/setup-maven@v5\n with:\n maven-version: ${{ env.MAVEN_VERSION }}\n\n - name: Set version\n run: |\n cd liquibase-pro/liquibase-azure-deps && mvn versions:set -DnewVersion=${{ inputs.liquibase-version }}\n\n - name: Build & Test\n run: |\n cd liquibase-pro/liquibase-azure-deps && mvn -B clean package\n\n - name: Archive Modules\n uses: actions/upload-artifact@v4\n with:\n name: liquibase-pro-azure-artifacts\n path: liquibase-pro/liquibase-azure-deps/target/liquibase-azure-deps-${{ inputs.liquibase-version }}.jar\n\n
dataset_sample\yaml\liquibase_liquibase\.github\workflows\build-azure-uber-jar.yml
build-azure-uber-jar.yml
YAML
3,689
0.95
0
0
awesome-app
381
2025-02-20T06:16:16.044194
BSD-3-Clause
false
a6881922920224f07d9207a242e7648b
name: Build Branch SNAPSHOT\n\n# concurrent runs for pull requests (both internal and forked) will be canceled when a new run is triggered.\n# It does not specifically target or exclude the 'master' branch; it cancels concurrent runs whenever a pull request is triggered.\nconcurrency:\n group: ${{ github.workflow }}-${{ github.head_ref || github.ref }}\n cancel-in-progress: ${{ github.event_name == 'pull_request' }}\n\non:\n workflow_dispatch:\n\n pull_request_target:\n types:\n - opened\n - reopened\n - synchronize\n - labeled\n\n push:\n branches:\n - github-action-**\n paths-ignore:\n - '**.md'\n\njobs:\n\n authorize:\n environment: ${{ github.event_name == 'pull_request_target' && github.event.pull_request.head.repo.full_name != github.repository && 'external' || 'internal' }}\n runs-on: ubuntu-latest\n steps:\n - run: true\n\n setup:\n name: Setup\n needs: authorize\n runs-on: ubuntu-22.04\n outputs:\n timeStamp: ${{ steps.get-timestamp.outputs.timestamp }}\n thisBranchName: ${{ steps.get-branch-name.outputs.thisBranchName }}\n latestMergeSha: ${{ steps.get-sha.outputs.latestMergeSha }}\n setupSuccessful: "true"\n\n steps:\n - uses: actions/checkout@v4\n with:\n ref: ${{ github.event.pull_request.head.sha || github.event.after}}\n\n - name: Get Latest Merge Commit SHA\n id: get-sha\n run: |\n latest_merge_sha=`(git rev-parse HEAD)`\n echo "latestMergeSha=${latest_merge_sha}" >> $GITHUB_OUTPUT\n\n - name: Get Timestamp\n id: get-timestamp\n run: |\n timeStamp=$(date +'%Y-%m-%d %H:%M:%S %Z')\n echo "timestamp=${timeStamp}" >> $GITHUB_OUTPUT\n\n - name: Get Current BranchName\n id: get-branch-name\n run: |\n # this logic checks if the branch is from a forked repository PR or not. Where -n is the inverse of -z (not empty)\n if [ -n "${GITHUB_HEAD_REF}" ];\n then\n branch_name=${GITHUB_HEAD_REF}\n else\n branch_name=${{ github.ref_name }}\n fi \n\n modified_branch_name=`(echo $branch_name | tr '/' '_')`\n echo "thisBranchName=$modified_branch_name" >> $GITHUB_OUTPUT \n\n build:\n name: Build & Package\n runs-on: ubuntu-22.04\n needs: setup\n permissions:\n contents: read\n packages: write\n steps:\n - uses: actions/checkout@v4\n with:\n ref: ${{ github.event.pull_request.head.sha || github.event.after}}\n - name: Set up Java for publishing to GitHub Repository\n uses: actions/setup-java@v4\n with:\n java-version: '17'\n distribution: 'temurin'\n cache: 'maven'\n server-id: liquibase\n\n # Version artifact based off of branch and commit SHA.\n - name: Version Artifact\n run: |\n ./mvnw versions:set "-DnewVersion=${{ needs.setup.outputs.thisBranchName }}-SNAPSHOT"\n #./mvnw versions:set "-DnewVersion=${{ inputs.branchName }}-SNAPSHOT"\n\n # Publish to GitHub Packages\n - name: Publish package\n run: |\n ./mvnw -B clean deploy -pl '!liquibase-dist' -DskipTests=true "-Dbuild.repository.owner=liquibase" "-Dbuild.repository.name=liquibase" "-Dbuild.branch=${{ needs.setup.outputs.thisBranchName }}" "-Dbuild.number=${{ github.run_number }}" "-Dbuild.commit=${{ needs.setup.outputs.latestMergeSha }}" "-Dbuild.timestamp=${{ needs.setup.outputs.timeStamp }}"\n env:\n GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}\n
dataset_sample\yaml\liquibase_liquibase\.github\workflows\build-branch.yml
build-branch.yml
YAML
3,533
0.8
0.038462
0.067416
node-utils
903
2023-12-12T22:37:07.470723
Apache-2.0
false
bfe50348ee20f219f0b140d154dac295
name: Build & Deploy extensions\n\non:\n workflow_call:\n inputs:\n liquibase-version:\n description: "liquibase version"\n required: true\n type: string\n dependencies:\n description: "Comma separated list of needed dependencies to release the extensions list"\n required: false\n type: string\n extensions:\n description: "Comma separated list of extensions to release to GPM"\n required: true\n type: string\n branch:\n description: "branch to check out"\n required: true\n type: string\n workflow_dispatch:\n inputs:\n liquibase-version:\n description: "liquibase version"\n required: true\n type: string\n dependencies:\n description: "Comma separated list of needed dependencies to release the extensions list"\n required: false\n type: string\n extensions:\n description: "Comma separated list of extensions to release to GPM"\n required: true\n type: string\n branch:\n description: "branch to check out"\n required: true\n type: string\n\nenv:\n MAVEN_VERSION: "3.9.2"\n\njobs:\n setup_matrix:\n runs-on: ubuntu-latest\n outputs:\n dependencies_matrix: ${{ steps.setup_dependencies_matrix.outputs.matrix_output }}\n extensions_matrix: ${{ steps.setup_extensions_matrix.outputs.matrix_output }}\n steps:\n - id: setup_dependencies_matrix\n uses: cschadewitz/dynamic-matrix-input@v1.0.0\n with:\n matrix_input: ${{ inputs.dependencies }}\n - id: setup_extensions_matrix\n uses: cschadewitz/dynamic-matrix-input@v1.0.0\n with:\n matrix_input: ${{ inputs.extensions }}\n\n delete-dependency-packages:\n needs: setup_matrix\n runs-on: ubuntu-22.04\n continue-on-error: true\n strategy:\n matrix:\n dependencies: ${{ fromJson(needs.setup_matrix.outputs.dependencies_matrix) }}\n steps:\n - uses: actions/delete-package-versions@v5\n with:\n package-name: org.liquibase.ext.${{ matrix.dependencies }}\n package-type: "maven"\n token: ${{ secrets.BOT_TOKEN }}\n ignore-versions: "^((?!${{ inputs.liquibase-version }}$).)*$"\n\n delete-extension-packages:\n needs: setup_matrix\n runs-on: ubuntu-22.04\n continue-on-error: true\n strategy:\n matrix:\n extensions: ${{ fromJson(needs.setup_matrix.outputs.extensions_matrix) }}\n steps:\n - uses: actions/delete-package-versions@v5\n with:\n package-name: org.liquibase.ext.${{ matrix.extensions }}\n package-type: "maven"\n token: ${{ secrets.BOT_TOKEN }}\n ignore-versions: "^((?!${{ inputs.liquibase-version }}$).)*$"\n\n matching-branch-logic-extensions:\n runs-on: ubuntu-latest\n outputs:\n extensions_branch: ${{ steps.get-extensions-branch.outputs.extensions_branch }}\n steps:\n - id: get-extensions-branch\n name: Get Extensions branch to use\n run: |\n if [ "${{ inputs.branch }}" == "master" ]; then\n echo "extensions_branch=main" >> "$GITHUB_OUTPUT"\n else\n echo "extensions_branch=${{ inputs.branch }}" >> "$GITHUB_OUTPUT"\n fi\n\n matching-branch-logic-pro:\n runs-on: ubuntu-latest\n outputs:\n pro_branch: ${{ steps.get-pro-branch.outputs.pro_branch }}\n steps:\n - id: get-pro-branch\n name: Get Pro branch to use\n run: |\n if [ "${{ inputs.branch }}" == "master" ]; then\n echo "pro_branch=master" >> "$GITHUB_OUTPUT"\n else\n echo "pro_branch=${{ inputs.branch }}" >> "$GITHUB_OUTPUT"\n fi \n\n get-liquibase-checks-version:\n if: ${{ contains(inputs.extensions, 'liquibase-checks') }}\n runs-on: ubuntu-latest\n outputs:\n version: ${{ steps.extract-version.outputs.version }}\n steps:\n - uses: actions/checkout@v4\n with:\n repository: liquibase/liquibase-checks\n token: ${{ secrets.BOT_TOKEN }}\n\n - id: extract-version\n name: Extract version from pom.xml\n shell: bash\n run: |\n VERSION=$(grep '<version>' pom.xml | head -n 1 | sed 's/.*<version>\(.*\)-SNAPSHOT<\/version>.*/\1/')\n echo "version=$VERSION" >> $GITHUB_OUTPUT\n\n delete-checks-packages:\n needs: [ get-liquibase-checks-version ]\n runs-on: ubuntu-22.04\n continue-on-error: true\n steps:\n - uses: actions/delete-package-versions@v5\n with:\n package-name: org.liquibase.ext.liquibase-checks\n package-type: "maven"\n token: ${{ secrets.BOT_TOKEN }}\n ignore-versions: "^((?!${{ needs.get-liquibase-checks-version.outputs.version }}$).)*$"\n\n build-liquibase-checks:\n if: ${{ contains(inputs.extensions, 'liquibase-checks') }}\n needs: [ delete-extension-packages, get-liquibase-checks-version, delete-checks-packages, matching-branch-logic-extensions ]\n uses: liquibase/build-logic/.github/workflows/publish-for-liquibase.yml@main\n with:\n repository: liquibase/liquibase-checks\n version: ${{ needs.get-liquibase-checks-version.outputs.version }}\n branch: ${{ needs.matching-branch-logic-extensions.outputs.extensions_branch }}\n secrets: inherit\n\n build-and-deploy-extensions:\n needs: [delete-dependency-packages, delete-extension-packages, delete-checks-packages, matching-branch-logic-extensions, matching-branch-logic-pro]\n runs-on: ubuntu-22.04\n strategy:\n fail-fast: false\n env:\n GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}\n GIT_USERNAME: "liquibot"\n GIT_PASSWORD: ${{ secrets.LIQUIBOT_PAT_GPM_ACCESS }}\n steps:\n - uses: actions/checkout@v4\n name: Checkout liquibase\n\n - uses: actions/checkout@v4\n name: Checkout liquibase-pro\n with:\n repository: liquibase/liquibase-pro\n ref: ${{ needs.matching-branch-logic-pro.outputs.pro_branch }}\n path: liquibase-pro\n token: ${{ secrets.BOT_TOKEN }}\n\n - name: Checkout Dependencies\n run: |\n git config --global credential.helper store\n echo "https://$GIT_USERNAME:$GIT_PASSWORD@github.com" > ~/.git-credentials\n IFS=',' read -ra DEP_ARRAY <<< "${{ inputs.dependencies }}"\n for dep in "${DEP_ARRAY[@]}"; do\n dep=$(echo $dep | xargs) # Remove leading and trailing whitespaces\n echo "Checking out $dep"\n git clone https://github.com/liquibase/$dep.git $dep\n git checkout ${{ inputs.branch }} || echo "Branch ${{ inputs.branch }} not found, staying on default branch"\n done\n\n - name: Set up JDK 21\n uses: actions/setup-java@v4\n with:\n java-version: 21\n distribution: "temurin"\n cache: "maven"\n gpg-private-key: ${{ secrets.GPG_SECRET }}\n gpg-passphrase: GPG_PASSPHRASE\n env:\n GPG_PASSWORD: ${{ secrets.GPG_PASSPHRASE }}\n\n - name: Set up Maven\n uses: stCarolas/setup-maven@v5\n with:\n maven-version: ${{ env.MAVEN_VERSION }}\n\n - name: maven-settings-xml-action\n uses: whelk-io/maven-settings-xml-action@v22\n with:\n repositories: |\n [\n {\n "id": "liquibase",\n "url": "https://maven.pkg.github.com/liquibase/liquibase",\n "releases": {\n "enabled": "true"\n },\n "snapshots": {\n "enabled": "true",\n "updatePolicy": "always"\n }\n },\n {\n "id": "liquibase-pro",\n "url": "https://maven.pkg.github.com/liquibase/liquibase-pro",\n "releases": {\n "enabled": "true"\n },\n "snapshots": {\n "enabled": "true",\n "updatePolicy": "always"\n }\n }\n ]\n servers: |\n [\n {\n "id": "liquibase",\n "username": "liquibot",\n "password": "${{ secrets.LIQUIBOT_PAT_GPM_ACCESS }}"\n },\n {\n "id": "liquibase-pro",\n "username": "liquibot",\n "password": "${{ secrets.LIQUIBOT_PAT_GPM_ACCESS }}"\n }\n ]\n\n - name: Install liquibase 0-SNAPSHOT\n run: mvn clean install -DskipTests\n\n - name: Install liquibase-commercial 0-SNAPSHOT\n run: |\n cd liquibase-pro\n mvn clean install -DskipTests -P '!run-proguard'\n cd ..\n\n - name: Re-version and build Dependencies\n env:\n GPG_PASSWORD: ${{ secrets.GPG_PASSPHRASE }}\n if: ${{ inputs.dependencies != '' }}\n continue-on-error: true\n run: |\n scripts_branch=${{ inputs.branch }}\n IFS=',' read -ra DEP_ARRAY <<< "${{ inputs.dependencies }}"\n for dep in "${DEP_ARRAY[@]}"; do\n dep=$(echo $dep | xargs) # Remove leading and trailing whitespaces\n echo "Re-versioning $dep"\n cd $dep\n sed -i "s/<liquibase.version>.*<\/liquibase.version>/<liquibase.version>0-SNAPSHOT<\/liquibase.version>/" pom.xml\n mvn versions:set -DnewVersion=0-SNAPSHOT\n mvn clean install -DskipTests\n mkdir -p $PWD/.github/util/\n curl -o $PWD/.github/util/re-version.sh https://raw.githubusercontent.com/liquibase/liquibase/$scripts_branch/.github/util/re-version.sh\n curl -o $PWD/.github/util/sign-artifacts.sh https://raw.githubusercontent.com/liquibase/liquibase/$scripts_branch/.github/util/sign-artifacts.sh\n curl -o $PWD/.github/util/ManifestReversion.java https://raw.githubusercontent.com/liquibase/liquibase/$scripts_branch/.github/util/ManifestReversion.java\n chmod +x $PWD/.github/util/re-version.sh\n chmod +x $PWD/.github/util/sign-artifacts.sh\n chmod +x $PWD/.github/util/ManifestReversion.java\n $PWD/.github/util/re-version.sh ./target "${{ inputs.liquibase-version }}" "${{ inputs.branch }}" ${dep}\n $PWD/.github/util/sign-artifacts.sh ./target "${{ inputs.liquibase-version }}" "${{ inputs.branch }}"\n mvn versions:set -DnewVersion=${{ inputs.liquibase-version }}\n sed -i -e "s/<liquibase.version>0-SNAPSHOT<\/liquibase.version>/<liquibase.version>${{ inputs.liquibase-version }}<\/liquibase.version>/g" pom.xml\n mvn deploy:deploy-file \\n -Dfile=./re-version/out/${dep}-${{ inputs.liquibase-version }}.jar \\n -Dsources=./re-version/out/${dep}-${{ inputs.liquibase-version }}-sources.jar \\n -Djavadoc=./re-version/out/${dep}-${{ inputs.liquibase-version }}-javadoc.jar \\n -DrepositoryId=liquibase \\n -Durl=https://maven.pkg.github.com/liquibase/$dep \\n -DpomFile=pom.xml\n cd ..\n done\n\n - name: Checkout and build Extensions\n env:\n GPG_PASSWORD: ${{ secrets.GPG_PASSPHRASE }}\n continue-on-error: true\n run: |\n scripts_branch=${{ needs.matching-branch-logic-extensions.outputs.extensions_branch }}\n git config --global credential.helper store\n echo "https://$GIT_USERNAME:$GIT_PASSWORD@github.com" > ~/.git-credentials\n IFS=',' read -ra EXT_ARRAY <<< "${{ inputs.extensions }}"\n for ext in "${EXT_ARRAY[@]}"; do\n ext=$(echo $ext | xargs) # Remove leading and trailing whitespaces\n if [ "$ext" != "liquibase-checks" ]; then\n echo "Checking out and building $ext"\n git clone https://github.com/liquibase/$ext.git $ext\n git checkout $scripts_branch\n cd $ext\n sed -i "s/<liquibase.version>.*<\/liquibase.version>/<liquibase.version>0-SNAPSHOT<\/liquibase.version>/" pom.xml\n mvn versions:set -DnewVersion=0-SNAPSHOT\n\n for dep in "${DEP_ARRAY[@]}"; do\n dep=$(echo $dep | xargs)\n sed -i "/<artifactId>${dep//./\\.}<\/artifactId>/{N; s/<version>.*<\/version>/<version>${{ inputs.liquibase-version }}<\/version>/}" pom.xml || true\n done\n\n mvn clean install -DskipTests\n\n mkdir -p $PWD/.github/util/\n curl -o $PWD/.github/util/re-version.sh https://raw.githubusercontent.com/liquibase/liquibase/$scripts_branch/.github/util/re-version.sh\n curl -o $PWD/.github/util/sign-artifacts.sh https://raw.githubusercontent.com/liquibase/liquibase/$scripts_branch/.github/util/sign-artifacts.sh\n curl -o $PWD/.github/util/ManifestReversion.java https://raw.githubusercontent.com/liquibase/liquibase/$scripts_branch/.github/util/ManifestReversion.java\n chmod +x $PWD/.github/util/re-version.sh\n chmod +x $PWD/.github/util/sign-artifacts.sh\n chmod +x $PWD/.github/util/ManifestReversion.java\n $PWD/.github/util/re-version.sh ./target "${{ inputs.liquibase-version }}" "${{ inputs.branch }}" ${ext}\n $PWD/.github/util/sign-artifacts.sh ./target "${{ inputs.liquibase-version }}" "${{ inputs.branch }}"\n mvn versions:set -DnewVersion=${{ inputs.liquibase-version }}\n sed -i -e "s/<liquibase.version>0-SNAPSHOT<\/liquibase.version>/<liquibase.version>${{ inputs.liquibase-version }}<\/liquibase.version>/g" pom.xml\n mvn deploy:deploy-file \\n -Dfile=./re-version/out/${ext}-${{ inputs.liquibase-version }}.jar \\n -Dsources=./re-version/out/${ext}-${{ inputs.liquibase-version }}-sources.jar \\n -Djavadoc=./re-version/out/${ext}-${{ inputs.liquibase-version }}-javadoc.jar \\n -DrepositoryId=liquibase \\n -Durl=https://maven.pkg.github.com/liquibase/$ext \\n -DpomFile=pom.xml\n cd ..\n fi\n done\n
dataset_sample\yaml\liquibase_liquibase\.github\workflows\build-extension-jars.yml
build-extension-jars.yml
YAML
14,052
0.95
0.032353
0
react-lib
352
2025-06-12T17:23:31.062018
GPL-3.0
false
b356d001aeb859cd8a213279891924dd
name: Build Main/Master SNAPSHOT\n\non:\n workflow_dispatch:\n schedule:\n - cron: "5 6 * * *"\n push:\n branches:\n - 'main'\n - 'master'\n\njobs:\n setup:\n name: Setup\n runs-on: ubuntu-22.04\n outputs:\n timeStamp: ${{ steps.get-timestamp.outputs.timeStamp }}\n latestMergeSha: ${{ steps.get-sha.outputs.latestMergeSha }}\n setupSuccessful: "true"\n\n steps:\n - uses: actions/checkout@v4\n - name: Get Latest Merge Commit SHA\n id: get-sha\n run: |\n latest_merge_sha=`(git rev-parse --short HEAD)`\n echo "latestMergeSha=${latest_merge_sha}" >> $GITHUB_OUTPUT\n\n - name: Get Timestamp\n id: get-timestamp\n run: |\n timestamp=`(date +'%Y-%m-%d %H:%M:%S %Z')`\n echo "timeStamp=${timestamp}" >> $GITHUB_OUTPUT\n\n\n build-and-publish:\n name: Build & Publish\n runs-on: ubuntu-22.04\n needs: setup\n permissions:\n contents: read\n packages: write\n env:\n GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}\n steps:\n - uses: actions/checkout@v4\n - name: Set up Java for publishing to GitHub Repository\n uses: actions/setup-java@v4\n with:\n java-version: '17'\n distribution: 'temurin'\n cache: 'maven'\n server-id: liquibase\n\n # Publish master-SNAPSHOT version to GPM\n - name: Version Artifact as master-SNAPSHOT\n run: ./mvnw versions:set "-DnewVersion=master-SNAPSHOT"\n\n - name: Publish master-SNAPSHOT\n run: ./mvnw -B clean deploy -pl '!liquibase-dist' -DskipTests=true "-Dbuild.repository.owner=liquibase" "-Dbuild.repository.name=liquibase" "-Dbuild.branch=master" "-Dbuild.number=${{ github.run_number }}" "-Dbuild.commit=${{ needs.setup.outputs.latestMergeSha }}" "-Dbuild.timestamp=${{ needs.setup.outputs.timeStamp }}"\n\n # Publish <commitsha>-SNAPSHOT version to GPM\n - name: Version Artifact as commit-SNAPSHOT\n run: ./mvnw versions:set -DnewVersion="${{ needs.setup.outputs.latestMergeSha }}-SNAPSHOT"\n\n - name: Publish commit-SNAPSHOT\n run: ./mvnw -B clean deploy -pl '!liquibase-dist' -DskipTests=true "-Dbuild.repository.owner=liquibase" "-Dbuild.repository.name=liquibase" "-Dbuild.branch=master" "-Dbuild.number=${{ github.run_number }}" "-Dbuild.commit=${{ needs.setup.outputs.latestMergeSha }}" "-Dbuild.timestamp=${{ needs.setup.outputs.timeStamp }}"\n
dataset_sample\yaml\liquibase_liquibase\.github\workflows\build-main.yml
build-main.yml
YAML
2,399
0.8
0.014925
0.035088
react-lib
877
2025-04-03T01:07:30.901695
GPL-3.0
false
1ad70b72ac87818489ee9dadb1cb94e4
name: Build and Publish\n\n# The concurrency check will cancel in-progress runs on non-master branches and pull requests on forked repositories regardless of the target branch.\nconcurrency:\n group: ${{ github.workflow }}-${{ github.head_ref == 'master' && 'master' || github.run_id }}\n cancel-in-progress: true\n\non:\n workflow_call:\n workflow_dispatch:\n\njobs:\n\n setup:\n name: Setup\n runs-on: ubuntu-22.04\n permissions:\n contents: write # Ensures the workflow can read/write branches\n env:\n GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }} # authenticate GitHub Actions to the GitHub API.\n github-token: ${{ secrets.LIQUIBOT_PAT_GPM_ACCESS }} # accessing a private repository\n outputs:\n branchExistsInBoth: ${{ steps.check-branch.outputs.branchExistsInBoth }}\n thisBranchName: ${{ steps.get-branch-name.outputs.thisBranchName }}\n latestMergeSha: ${{ steps.get-sha.outputs.latestMergeSha }}\n timeStamp: ${{ steps.get-timestamp.outputs.timestamp }}\n proBranchNamePrefix: ${{ steps.check-branch.outputs.proBranchNamePrefix }}\n unmodifiedBranchName: ${{ steps.get-branch-name.outputs.unmodifiedBranchName }}\n\n steps:\n\n - uses: actions/checkout@v4\n with:\n ref: ${{ github.event.pull_request.head.sha || github.event.after}}\n - name: Get Latest Merge Commit SHA\n id: get-sha\n run: |\n latest_merge_sha=`(git rev-parse HEAD)`\n echo "latestMergeSha=${latest_merge_sha}" >> $GITHUB_OUTPUT\n\n - name: Get Timestamp\n id: get-timestamp\n run: |\n timeStamp=$(date +'%Y-%m-%d %H:%M:%S %Z')\n echo "timestamp=${timeStamp}" >> $GITHUB_OUTPUT\n\n - name: Get Current BranchName\n id: get-branch-name\n run: |\n # this logic checks if the branch is from a forked repository PR or not. Where -n is the inverse of -z (not empty)\n if [ -n "${GITHUB_HEAD_REF}" ];\n then\n branch_name=${GITHUB_HEAD_REF}\n else\n branch_name=${{ github.ref_name }}\n fi \n modified_branch_name=`(echo $branch_name | tr '/' '_')`\n echo "thisBranchName=$modified_branch_name" >> $GITHUB_OUTPUT \n\n #this is to check if the same branch name exists in PRO, we need it in check-branch step\n echo "unmodifiedBranchName=$branch_name" >> $GITHUB_OUTPUT \n echo "Modified Branch Name:" $modified_branch_name\n echo "Original Branch Name thisBranchName:" $branch_name\n\n - name: Checkout OSS Repo\n uses: actions/checkout@v4\n with:\n repository: liquibase/liquibase\n\n #RUN THIS LOGIC WHEN THERE IS A SAME BRANCH IN OSS AND PRO\n\n #checking if the branch specified by $currentBranch exists in the PRO repository, If the getBranch call throws an error, it means the branch doesn't exist in the PRO repo. Also, remove refs/heads/\n # if the value of branchExistsInBoth = false then set liquibaseProVersionPrefix = master else $currentBranch\n # if the value of branchExistsInBoth = true then set liquibaseProVersionPrefix = thisBranchName else $currentBranch as PRO also has replacement logic for / with _\n - name: Check branch existence\n id: check-branch\n uses: actions/github-script@v7\n with:\n github-token: ${{ secrets.LIQUIBOT_PAT_GPM_ACCESS }}\n script: |\n const currentBranch = "${{ steps.get-branch-name.outputs.unmodifiedBranchName }}" \n console.log(currentBranch)\n let branchExistsInBoth;\n try {\n await github.rest.repos.getBranch({\n owner: "liquibase",\n repo: "liquibase-pro",\n branch: currentBranch\n });\n branchExistsInBoth = true\n } catch (error){\n branchExistsInBoth = false\n }\n let liquibaseProVersionPrefix= branchExistsInBoth ? currentBranch : 'master'\n liquibaseProVersionPrefix=liquibaseProVersionPrefix.replaceAll("/","_")\n core.setOutput("branchExistsInBoth", branchExistsInBoth)\n core.setOutput("proBranchNamePrefix", liquibaseProVersionPrefix)\n console.log(liquibaseProVersionPrefix)\n\n\n ### MATCHING BRANCH RUN THIS LOGIC\n build_publish_branch:\n name: Artifacts - Build & Package\n runs-on: macos-latest\n needs: [ setup ]\n permissions:\n contents: read\n packages: write\n steps:\n - uses: actions/checkout@v4\n with:\n ref: ${{ github.event.pull_request.head.sha || github.event.after}}\n # this includes all the tar files included in the previous runs. So in the next step we deploy what was previously build\n - name: Download Artifacts for build.yml\n uses: actions/download-artifact@v4\n with:\n name: temp-artifact\n\n ##Cache based on install4j file, since changes to JVM is seen in there. If install4j version changes without changing the file, change the prefix letter before hashFiles to force a new cache\n - name: Install4j Cache\n uses: actions/cache@v4.2.0\n with:\n key: install4j-A${{ hashFiles('liquibase-dist/src/main/install4j/liquibase.install4j') }}\n path: ~/.install4j8/**\n\n - name: Set up JDK 11\n uses: actions/setup-java@v4\n with:\n java-version: '11'\n distribution: 'temurin'\n gpg-private-key: ${{ secrets.GPG_SECRET }}\n gpg-passphrase: GPG_PASSPHRASE\n cache: 'maven'\n overwrite-settings: false\n env:\n GPG_PASSWORD: ${{ secrets.GPG_PASSPHRASE }}\n\n #look for dependencies in maven\n - name: maven-settings-xml-action\n uses: whelk-io/maven-settings-xml-action@v22\n with:\n repositories: |\n [\n {\n "id": "liquibase",\n "url": "https://maven.pkg.github.com/liquibase/liquibase",\n "releases": {\n "enabled": "false"\n },\n "snapshots": {\n "enabled": "true",\n "updatePolicy": "always"\n }\n },\n {\n "id": "liquibase-pro",\n "url": "https://maven.pkg.github.com/liquibase/liquibase-pro",\n "releases": {\n "enabled": "false"\n },\n "snapshots": {\n "enabled": "true",\n "updatePolicy": "always"\n }\n }\n ]\n servers: |\n [\n {\n "id": "liquibase-pro",\n "username": "liquibot",\n "password": "${{ secrets.LIQUIBOT_PAT }}"\n },\n {\n "id": "liquibase",\n "username": "liquibot",\n "password": "${{ secrets.LIQUIBOT_PAT }}"\n }\n ]\n\n # Version artifact based off of branch\n - name: Version Artifact\n run: |\n version=${{ needs.setup.outputs.thisBranchName }}-SNAPSHOT\n ./mvnw versions:set -DnewVersion="$version"\n\n - name: Artifacts - Build & Sign\n env:\n INSTALL4J_LICENSE: ${{ secrets.INSTALL4J_LICENSE }}\n INSTALL4J_APPLE_KEY: ${{ secrets.INSTALL4J_APPLE_KEY }}\n INSTALL4J_APPLE_KEY_PASSWORD: ${{ secrets.INSTALL4J_APPLE_KEY_PASSWORD }}\n INSTALL4J_APPLE_ID: ${{ secrets.INSTALL4J_APPLE_ID }}\n INSTALL4J_APPLE_ID_PASSWORD: ${{ secrets.INSTALL4J_APPLE_ID_PASSWORD }}\n INSTALL4J_WINDOWS_KEY: ${{ secrets.INSTALL4J_WINDOWS_KEY }}\n INSTALL4J_WINDOWS_KEY_PASSWORD: ${{ secrets.INSTALL4J_WINDOWS_KEY_PASSWORD }}\n GPG_PASSWORD: ${{ secrets.GPG_PASSPHRASE }}\n GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}\n run: |\n ## save install4j code signing keys\n mkdir -p liquibase-dist/target/keys\n echo "Saving apple key"\n echo "$INSTALL4J_APPLE_KEY" | base64 -d > liquibase-dist/target/keys/datical_apple.p12\n\n echo "Saving windows key"\n echo "$INSTALL4J_WINDOWS_KEY" | base64 -d > liquibase-dist/target/keys/datical_windows.pfx\n # we are packaging and deploying liquibase tar which includes liquibase-commercial\n # Build liquibase-commercial version\n ./mvnw -B -pl liquibase-dist -P liquibase-commercial source:jar clean package failsafe:integration-test failsafe:verify \\n "-Dliquibase.checks=true" \\n "-Dliquibase-pro.version=${{ needs.setup.outputs.proBranchNamePrefix }}-SNAPSHOT" \\n "-Dbuild.repository.owner=liquibase" \\n "-Dbuild.repository.name=liquibase" \\n "-Dbuild.branch=${{ needs.setup.outputs.thisBranchName }}" \\n "-Dbuild.number=${{ github.run_number }}" \\n "-Dbuild.commit=${{ needs.setup.outputs.latestMergeSha }}-SNAPSHOT" \\n "-Dbuild.timestamp=${{ needs.setup.outputs.timeStamp }}"\n\n # backup commercial tarball to prevent it from being overwritten\n mkdir -p liquibase-dist/backup\n mv liquibase-dist/target/liquibase-${{ needs.setup.outputs.thisBranchName }}-SNAPSHOT.tar.gz liquibase-dist/backup/liquibase-${{ needs.setup.outputs.thisBranchName }}-SNAPSHOT.tar.gz\n mv liquibase-dist/target/liquibase-${{ needs.setup.outputs.thisBranchName }}-SNAPSHOT.zip liquibase-dist/backup/liquibase-${{ needs.setup.outputs.thisBranchName }}-SNAPSHOT.zip\n\n # Build liquibase-minimal version\n ./mvnw -B -pl liquibase-dist -P liquibase-minimal source:jar clean package failsafe:integration-test failsafe:verify \\n "-Dliquibase.checks=true" \\n "-Dliquibase-pro.version=${{ needs.setup.outputs.proBranchNamePrefix }}-SNAPSHOT" \\n "-Dbuild.repository.owner=liquibase" \\n "-Dbuild.repository.name=liquibase" \\n "-Dbuild.branch=${{ needs.setup.outputs.thisBranchName }}" \\n "-Dbuild.number=${{ github.run_number }}" \\n "-Dbuild.commit=${{ needs.setup.outputs.latestMergeSha }}-SNAPSHOT" \\n "-Dbuild.timestamp=${{ needs.setup.outputs.timeStamp }}"\n\n # Rename liquibase-minimal tarball to avoid overwriting the commercial one\n mv liquibase-dist/target/liquibase-${{ needs.setup.outputs.thisBranchName }}-SNAPSHOT.tar.gz liquibase-dist/target/liquibase-minimal-${{ needs.setup.outputs.thisBranchName }}-SNAPSHOT.tar.gz\n mv liquibase-dist/target/liquibase-${{ needs.setup.outputs.thisBranchName }}-SNAPSHOT.zip liquibase-dist/target/liquibase-minimal-${{ needs.setup.outputs.thisBranchName }}-SNAPSHOT.zip\n\n # Restore commercial tarballs from backup into target dir\n mv liquibase-dist/backup/liquibase-${{ needs.setup.outputs.thisBranchName }}-SNAPSHOT.tar.gz liquibase-dist/target/liquibase-${{ needs.setup.outputs.thisBranchName }}-SNAPSHOT.tar.gz\n mv liquibase-dist/backup/liquibase-${{ needs.setup.outputs.thisBranchName }}-SNAPSHOT.zip liquibase-dist/target/liquibase-${{ needs.setup.outputs.thisBranchName }}-SNAPSHOT.zip\n\n # Extract the tarball to allow us to upload the liquibase installation files to our\n # github actions workflow so that we don't double zip the archive\n mkdir -p branch-artifacts\n tar -xzf liquibase-dist/target/liquibase-${{ needs.setup.outputs.thisBranchName }}-SNAPSHOT.tar.gz -C branch-artifacts\n\n ##create installer - disabled here but run as nightly job and as part of release workflow\n # (cd liquibase-dist && ${{ github.workspace }}/.github/util/package-install4j.sh 0-SNAPSHOT)\n\n # Remove original JARs\n find . -name original-*.jar -exec rm {} \;\n\n # Copy artifacts\n mkdir -p artifacts\n\n cp liquibase-dist/target/liquibase-${{ needs.setup.outputs.thisBranchName }}-SNAPSHOT.tar.gz artifacts\n cp liquibase-dist/target/liquibase-${{ needs.setup.outputs.thisBranchName }}-SNAPSHOT.zip artifacts\n cp liquibase-dist/target/liquibase-minimal-${{ needs.setup.outputs.thisBranchName }}-SNAPSHOT.tar.gz artifacts\n cp liquibase-dist/target/liquibase-minimal-${{ needs.setup.outputs.thisBranchName }}-SNAPSHOT.zip artifacts\n cp liquibase-core/target/liquibase-core-0-SNAPSHOT.jar artifacts/liquibase-core-0-SNAPSHOT.jar\n #cp liquibase-dist/target/liquibase-${{ needs.setup.outputs.thisBranchName }}-SNAPSHOT/internal/lib/liquibase-core.jar artifacts/liquibase-core-0-SNAPSHOT.jar\n\n cp liquibase-core/target/liquibase-core-0-SNAPSHOT-sources.jar artifacts/liquibase-core-0-SNAPSHOT-sources.jar\n cp target/liquibase-0-SNAPSHOT-javadoc.jar artifacts/liquibase-core-0-SNAPSHOT-javadoc.jar\n\n ##create installer - disabled here but run as nightly job and as part of release workflow\n #cp liquibase-dist/target/liquibase-*-installer-* artifacts\n\n for i in 'liquibase-maven-plugin' 'liquibase-cli' 'liquibase-cdi' 'liquibase-cdi-jakarta'; do\n cp $i/target/$i-0-SNAPSHOT.jar artifacts\n cp $i/target/$i-0-SNAPSHOT-sources.jar artifacts\n cp $i/target/$i-0-SNAPSHOT-javadoc.jar artifacts\n done\n\n echo "Source code not available for liquibase-commercial" > /tmp/readme.source.txt\n (cd /tmp && jar cf liquibase-commercial-${{ needs.setup.outputs.proBranchNamePrefix }}-SNAPSHOT-sources.jar readme.source.txt)\n\n echo "Javadocs not available for liquibase-commercial" > /tmp/readme.javadocs.txt\n (cd /tmp && jar cf liquibase-commercial-${{ needs.setup.outputs.proBranchNamePrefix }}-SNAPSHOT-javadoc.jar readme.javadocs.txt)\n\n ls -R /Users/runner/.m2/repository/org/liquibase/liquibase-commercial\n echo "/Users/runner/.m2/repository/org/liquibase/liquibase-commercial "\n\n commercialJarM2Location=/Users/runner/.m2/repository/org/liquibase/liquibase-commercial/${{ needs.setup.outputs.proBranchNamePrefix }}-SNAPSHOT/liquibase-commercial-${{ needs.setup.outputs.proBranchNamePrefix }}-SNAPSHOT.jar\n cp $commercialJarM2Location artifacts\n cp /tmp/liquibase-commercial-${{ needs.setup.outputs.proBranchNamePrefix }}-SNAPSHOT-sources.jar artifacts\n cp /tmp/liquibase-commercial-${{ needs.setup.outputs.proBranchNamePrefix }}-SNAPSHOT-javadoc.jar artifacts\n\n .github/util/sign-artifacts.sh artifacts\n\n ##prepare branch-named convenience artifacts directories\n mkdir artifacts-named\n\n cp liquibase-dist/target/liquibase-${{ needs.setup.outputs.thisBranchName }}-SNAPSHOT.tar.gz artifacts-named/liquibase-${{ needs.setup.outputs.thisBranchName }}.tar.gz\n cp liquibase-dist/target/liquibase-minimal-${{ needs.setup.outputs.thisBranchName }}-SNAPSHOT.tar.gz artifacts-named/liquibase-${{ needs.setup.outputs.thisBranchName }}-minimal.tar.gz\n cp liquibase-extension-testing/target/liquibase-extension-testing-0-SNAPSHOT-deps.jar artifacts-named/liquibase-extension-testing-${{ needs.setup.outputs.thisBranchName }}-deps.jar\n\n for i in 'liquibase-core' 'liquibase-maven-plugin' 'liquibase-cli' 'liquibase-cdi' 'liquibase-cdi-jakarta' 'liquibase-extension-testing'; do\n cp $i/target/$i-0-SNAPSHOT.jar artifacts-named/$i-${{ needs.setup.outputs.thisBranchName }}.jar\n done\n\n #check why only this step has branchName when there is a standalone PR\n cp $commercialJarM2Location artifacts-named/liquibase-commercial-${{ needs.setup.outputs.thisBranchName }}.jar\n\n - name: Archive Packages\n uses: actions/upload-artifact@v4\n with:\n name: liquibase-artifacts\n path: artifacts/*\n\n - name: Archive Convenience Zip\n uses: actions/upload-artifact@v4\n with:\n name: liquibase-zip-${{ needs.setup.outputs.thisBranchName }}\n path: branch-artifacts/*\n\n - name: Archive Convenience Artifacts\n uses: actions/upload-artifact@v4\n with:\n name: liquibase-artifacts-${{ needs.setup.outputs.thisBranchName }}\n path: artifacts-named/*\n\n # Publish <commitsha>-SNAPSHOT version to GPM\n - name: Version Artifact Publish <commitsha>-SNAPSHOT version to GPM\n run: |\n ./mvnw versions:set "-DnewVersion=${{ needs.setup.outputs.latestMergeSha }}-SNAPSHOT"\n - name: Publish <commitsha>-SNAPSHOT package\n run: |\n ./mvnw -B clean deploy -pl '!liquibase-dist' -DskipTests=true "-Dbuild.repository.owner=liquibase" "-Dbuild.repository.name=liquibase" "-Dbuild.branch=${{ needs.setup.outputs.thisBranchName }}" "-Dbuild.number=${{ github.run_number }}" "-Dbuild.commit=${{ needs.setup.outputs.latestMergeSha }}" "-Dbuild.timestamp=${{ needs.setup.outputs.timeStamp }}"\n env:\n GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}\n\n #Publish to GitHub Packages. We are building a tar file which includes liquibase-commercial hence we need "-Dliquibase-pro.version"\n - name: Publish tar.gz package to GPM\n run: |\n ./mvnw -B -pl liquibase-dist -P liquibase-commercial clean deploy -DskipTests=true "-Dliquibase-pro.version=${{ needs.setup.outputs.proBranchNamePrefix }}-SNAPSHOT" "-Dbuild.timestamp=${{ needs.setup.outputs.timestamp }}" "-Dbuild.repository.owner=liquibase" "-Dbuild.repository.name=liquibase" "-Dbuild.branch=${{ needs.setup.outputs.thisBranchName }}" "-Dbuild.number=${{ github.run_number }}" "-Dbuild.commit=${{ needs.setup.outputs.thisSha }}"\n env:\n GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}\n\n run-functional-tests:\n needs: [ setup, build_publish_branch ]\n uses: liquibase/liquibase/.github/workflows/run-functional-tests.yml@master\n secrets: inherit\n with:\n latestMergeSha: ${{ needs.setup.outputs.latestMergeSha }}\n thisBranchName: ${{ needs.setup.outputs.unmodifiedBranchName }}\n\n run-tests-harness:\n needs: [ setup,build_publish_branch ]\n uses: liquibase/liquibase/.github/workflows/run-test-harness.yml@master\n secrets: inherit\n with:\n latestMergeSha: ${{ needs.setup.outputs.latestMergeSha }}\n thisBranchName: ${{ needs.setup.outputs.unmodifiedBranchName }}\n\n run-pro-build-and-test:\n needs: [ setup, build_publish_branch ]\n uses: liquibase/liquibase/.github/workflows/run-pro-build-and-test.yml@master\n secrets: inherit\n with:\n latestMergeSha: ${{ needs.setup.outputs.latestMergeSha }}\n thisBranchName: ${{ needs.setup.outputs.unmodifiedBranchName }}\n
dataset_sample\yaml\liquibase_liquibase\.github\workflows\build.yml
build.yml
YAML
18,462
0.95
0.041436
0.102236
python-kit
702
2023-10-26T02:00:12.902065
MIT
false
391dd0bf3bf134c2e345137d265d8f52
name: Cleanup on Branch Delete\n\non:\n workflow_dispatch:\n delete:\n branches:\n - '**DAT-**'\n - 'github-action-**'\n - 'dependabot-**'\n - 'feat/**'\n\njobs:\n\n extract-branch-info:\n name: Extract Branch Info\n runs-on: ubuntu-22.04\n outputs:\n branch_name: ${{ steps.get-branch-name.outputs.branch_name }}\n steps:\n \n - name: Extract branch name\n id: get-branch-name\n run: |\n branch_name=${{ github.event.ref }}\n branch_name=${branch_name#refs/heads/}\n echo "Branch name is $branch_name"\n echo "branch_name=$branch_name" >> $GITHUB_OUTPUT\n\n delete-dat-package:\n name: Delete DAT-SNAPSHOT Github Package for Branch\n needs: extract-branch-info\n runs-on: ubuntu-22.04\n\n steps:\n\n # Get version id(s) based on version name\n - uses: castlabs/get-package-version-id-action@v2.2\n id: version\n with:\n version: "${{ needs.extract-branch-info.outputs.branch_name }}-SNAPSHOT"\n\n # show versions to delete\n - run: |\n echo "Version to delete ${{ steps.version.outputs.ids }} for ${{ needs.extract-branch-info.outputs.branch_name }}-SNAPSHOT"\n \n - uses: actions/delete-package-versions@v5\n if: ${{ steps.version.outputs.ids != '' }}\n with:\n # Delete all deletable versions\n package-type: maven\n \n # on branch id's deletion we only want to delete that particular branch version\n package-version-ids: "${{ steps.version.outputs.ids }}"\n \n delete-sha-package:\n name: Delete SHA-SNAPSHOT Github Package for Branch\n needs: [ delete-dat-package, extract-branch-info ]\n runs-on: ubuntu-22.04\n\n strategy:\n matrix:\n packages_to_delete: [org.liquibase.liquibase-core, org.liquibase.liquibase-extension-testing, org.liquibase.liquibase-integration-tests, org.liquibase.liquibase-maven-plugin, org.liquibase.liquibase-cdi-jakarta, org.liquibase.liquibase-cdi, org.liquibase.liquibase]\n\n steps:\n\n - name: Fetch commits between branch and master via GitHub API\n id: fetch-commits\n env:\n GH_TOKEN: ${{ secrets.GITHUB_TOKEN }}\n run: |\n BRANCH_NAME=${{ needs.extract-branch-info.outputs.branch_name }}\n echo "Fetching commits for branch: $BRANCH_NAME"\n\n # Fetch commits between branch and master\n MASTER_COMMITS_RESPONSE=$(gh api -H "Accept: application/vnd.github.v3+json" /repos/${{ github.repository }}/compare/master...$BRANCH_NAME || echo "error")\n if [[ "$MASTER_COMMITS_RESPONSE" == "error" || "$MASTER_COMMITS_RESPONSE" == "null" ]]; then\n echo "No differences found between branch $BRANCH_NAME and master."\n MASTER_COMMITS=""\n else\n MASTER_COMMITS=$(echo "$MASTER_COMMITS_RESPONSE" | jq -r '.commits[].sha' | tr '\n' ',' | sed 's/,$//')\n fi\n\n # Fetch commits between branch and release\n RELEASE_COMMITS_RESPONSE=$(gh api -H "Accept: application/vnd.github.v3+json" /repos/${{ github.repository }}/compare/release...$BRANCH_NAME || echo "error")\n if [[ "$RELEASE_COMMITS_RESPONSE" == "error" || "$RELEASE_COMMITS_RESPONSE" == "null" ]]; then\n echo "No differences found between branch $BRANCH_NAME and release."\n RELEASE_COMMITS=""\n else\n RELEASE_COMMITS=$(echo "$RELEASE_COMMITS_RESPONSE" | jq -r '.commits[].sha' | tr '\n' ',' | sed 's/,$//')\n fi\n\n echo "MASTER_COMMITS=$MASTER_COMMITS" >> $GITHUB_ENV\n echo "RELEASE_COMMITS=$RELEASE_COMMITS" >> $GITHUB_ENV\n\n - name: Prepare list of package versions to delete\n id: prepare-packages\n run: |\n MASTER_COMMITS="${{ env.MASTER_COMMITS }}"\n RELEASE_COMMITS="${{ env.env.RELEASE_COMMITS }}"\n PACKAGE_VERSIONS=""\n\n for COMMIT in $(echo "$MASTER_COMMITS,$RELEASE_COMMITS" | tr ',' '\n'); do\n if [ -n "$COMMIT" ]; then\n PACKAGE_VERSIONS="${PACKAGE_VERSIONS}${COMMIT}-SNAPSHOT,"\n fi\n done\n\n # Remove trailing comma\n PACKAGE_VERSIONS="${PACKAGE_VERSIONS%,}"\n echo "PACKAGE_VERSIONS=$PACKAGE_VERSIONS" >> $GITHUB_ENV\n echo "Package versions to query: $PACKAGE_VERSIONS"\n\n - name: Get package version IDs\n env:\n GH_TOKEN: ${{ secrets.GITHUB_TOKEN }}\n run: |\n PACKAGE_IDS=""\n ORG="liquibase"\n PACKAGE_TYPE="maven"\n\n for VERSION in $(echo "${{ env.PACKAGE_VERSIONS }}" | tr ',' '\n'); do\n echo "Processing version: $VERSION"\n\n ID=$(gh api "orgs/$ORG/packages/$PACKAGE_TYPE/${{ matrix.packages_to_delete }}/versions" | jq -r --arg VERSION "$VERSION" '.[] | select(.name == $VERSION) | .id')\n\n if [ -n "$ID" ]; then\n echo "Found ID: $ID for version $VERSION"\n PACKAGE_IDS="${PACKAGE_IDS}${ID},"\n else\n echo "No ID found for version $VERSION"\n fi\n done\n echo "PACKAGE_IDS=$PACKAGE_IDS" >> $GITHUB_ENV\n echo "Package ID versions to delete: $PACKAGE_IDS"\n \n - name: Delete package versions\n if: env.PACKAGE_VERSIONS != ''\n env:\n GH_TOKEN: ${{ secrets.GITHUB_TOKEN }}\n run: |\n for ID in $(echo "${{ env.PACKAGE_IDS }}" | tr ',' '\n'); do\n echo "Deleting package version with ID: $ID"\n gh api -X DELETE "orgs/liquibase/packages/maven/${{ matrix.packages_to_delete }}/versions/$ID"\n done\n
dataset_sample\yaml\liquibase_liquibase\.github\workflows\cleanup-branch-builds.yml
cleanup-branch-builds.yml
YAML
5,422
0.8
0.104167
0.058824
node-utils
668
2025-02-22T17:34:24.361529
GPL-3.0
false
b0b86ecb524c423d89d3710149fe1aa4
name: Cleanup Master Builds on Release Published\n\non:\n workflow_dispatch:\n release:\n types: [ published ]\n\njobs:\n delete-package:\n name: Delete Github Packages for Master/Main\n runs-on: ubuntu-22.04\n\n strategy:\n matrix:\n packages_to_delete: [org.liquibase.liquibase-core, org.liquibase.liquibase-extension-testing, org.liquibase.liquibase-integration-tests, org.liquibase.liquibase-maven-plugin, org.liquibase.liquibase-cdi-jakarta, org.liquibase.liquibase-cdi, org.liquibase.liquibase]\n steps:\n - uses: actions/delete-package-versions@v5\n with:\n # Name of the package.\n # Defaults to an empty string.\n # Required if `package-version-ids` input is not given.\n package-name: ${{ matrix.packages_to_delete }}\n package-type: maven\n\n # The number of latest versions to keep.\n # This cannot be specified with `num-old-versions-to-delete`. By default, `num-old-versions-to-delete` takes precedence over `min-versions-to-keep`.\n # When set to 0, all deletable versions will be deleted.\n # When set greater than 0, all deletable package versions except the specified number will be deleted.\n min-versions-to-keep: 0\n\n # The package versions to exclude from deletion.\n # Takes regex for the version name as input.\n # By default nothing is ignored.\n ignore-versions: '^DAT-.*-SNAPSHOT$'\n
dataset_sample\yaml\liquibase_liquibase\.github\workflows\cleanup-master-builds.yml
cleanup-master-builds.yml
YAML
1,413
0.8
0.088235
0.344828
vue-tools
581
2025-05-02T20:22:56.326122
MIT
false
ce049b8eb7d97194c48cb233fe52155f
# For most projects, this workflow file will not need changing; you simply need\n# to commit it to your repository.\n#\n# You may wish to alter this file to override the set of languages analyzed,\n# or to provide custom queries or build logic.\n#\n# ******** NOTE ********\n# We have attempted to detect the languages in your repository. Please check\n# the `language` matrix defined below to confirm you have the correct set of\n# supported CodeQL languages.\n#\nname: "CodeQL"\n\non:\n push:\n branches: [ "master" ]\n pull_request:\n # The branches below must be a subset of the branches above\n branches: [ "master" ]\n schedule:\n - cron: '16 14 * * 4'\n\njobs:\n analyze:\n name: Analyze\n runs-on: ${{ (matrix.language == 'swift' && 'macos-latest') || 'ubuntu-latest' }}\n timeout-minutes: ${{ (matrix.language == 'swift' && 120) || 360 }}\n permissions:\n actions: read\n contents: read\n security-events: write\n\n strategy:\n fail-fast: false\n matrix:\n language: [ 'java' ]\n # CodeQL supports [ 'cpp', 'csharp', 'go', 'java', 'javascript', 'python', 'ruby', 'swift' ]\n # Use only 'java' to analyze code written in Java, Kotlin or both\n # Use only 'javascript' to analyze code written in JavaScript, TypeScript or both\n # Learn more about CodeQL language support at https://aka.ms/codeql-docs/language-support\n\n steps:\n - name: Checkout repository\n uses: actions/checkout@v4\n\n # Initializes the CodeQL tools for scanning.\n - name: Initialize CodeQL\n uses: github/codeql-action/init@v3\n with:\n languages: ${{ matrix.language }}\n # If you wish to specify custom queries, you can do so here or in a config file.\n # By default, queries listed here will override any specified in a config file.\n # Prefix the list here with "+" to use these queries and those in the config file.\n\n # For more details on CodeQL's query packs, refer to: https://docs.github.com/en/code-security/code-scanning/automatically-scanning-your-code-for-vulnerabilities-and-errors/configuring-code-scanning#using-queries-in-ql-packs\n queries: security-extended,security-and-quality\n\n\n # Autobuild attempts to build any compiled languages (C/C++, C#, Go, or Java).\n # If this step fails, then you should remove it and run the build manually (see below)\n - name: Autobuild\n uses: github/codeql-action/autobuild@v3\n\n # ℹ️ Command-line programs to run using the OS shell.\n # 📚 See https://docs.github.com/en/actions/using-workflows/workflow-syntax-for-github-actions#jobsjob_idstepsrun\n\n # If the Autobuild fails above, remove it and uncomment the following three lines.\n # modify them (or add more) to build your code if your project, please refer to the EXAMPLE below for guidance.\n\n # - run: |\n # echo "Run, Build Application using script"\n # ./location_of_script_within_repo/buildscript.sh\n\n - name: Perform CodeQL Analysis\n uses: github/codeql-action/analyze@v3\n with:\n category: "/language:${{matrix.language}}"\n\n - name: Generate Security Report\n uses: rsdmike/github-security-report-action@v3.0.4\n with:\n token: ${{ secrets.GITHUB_TOKEN }}\n outputDir: ./reports/\n\n - name: Upload Security Report\n uses: actions/upload-artifact@v4\n with:\n name: security-report\n path: ./reports/summary.pdf\n
dataset_sample\yaml\liquibase_liquibase\.github\workflows\codeql.yml
codeql.yml
YAML
3,427
0.8
0.05618
0.4
vue-tools
601
2024-05-18T12:16:26.041302
Apache-2.0
false
902200bf9f20861e74d905722992e860
name: Attach Artifacts To Draft Release\non:\n workflow_dispatch:\n inputs:\n version:\n description: "Version (example: 4.8.2)"\n required: true\n branch:\n description: "Branch to release (Defaults to master)"\n required: false\n default: "master"\n runId:\n description: "RunId of liquibase/liquibase artifacts to attach"\n required: true\n standalone_zip:\n description: "Flag to indicate if the workflow is triggered to create a standalone zip"\n required: false\n type: boolean\n default: false\n dry_run:\n description: "Flag to indicate if the workflow is triggered to create a dry-run release"\n required: false\n type: boolean\n default: false\n workflow_call:\n inputs:\n version:\n description: "Version (example: 4.8.2)"\n required: true\n type: string\n branch:\n description: "Branch to release (Defaults to master)"\n required: false\n default: "master"\n type: string\n runId:\n description: "RunId of liquibase/liquibase artifacts to attach"\n required: true\n type: string\n standalone_zip:\n description: "Flag to indicate if the workflow is triggered to create a standalone zip"\n required: true\n type: boolean\n default: false\n dry_run:\n description: "Flag to indicate if the workflow is triggered to create a dry-run release"\n required: true\n type: boolean\n default: false\n outputs:\n dry_run_zip_url:\n description: "The URL of the created zip file"\n value: ${{ jobs.build-installers.outputs.dry_run_zip_url }}\n dry_run_tar_gz_url:\n description: "The URL of the created tar.gz file"\n value: ${{ jobs.build-installers.outputs.dry_run_tar_gz_url }}\n\nenv:\n DEPENDENCIES: "liquibase-bigquery" # Comma separated list of dependencies to release the extensions list\n EXTENSIONS: "liquibase-commercial-bigquery,liquibase-checks" # Comma separated list of extensions to release to GPM\n\njobs:\n setup:\n name: Setup\n runs-on: ubuntu-22.04\n outputs:\n version: ${{ inputs.version }}\n branch: ${{ inputs.branch }}\n runId: ${{ inputs.runId }}\n uber_jar_runId: ${{ steps.get_run_id.outputs.run_id }}\n dependencies: ${{ env.DEPENDENCIES }}\n extensions: ${{ env.EXTENSIONS }}\n steps:\n - run: |\n echo "Creating version ${{ inputs.version }} from ${{ inputs.branch }} with artifacts from build ${{ inputs.runId }} "\n\n owasp-scanner:\n needs: [setup]\n uses: liquibase/build-logic/.github/workflows/owasp-scanner.yml@main\n with:\n branch: ${{ needs.setup.outputs.branch }}\n secrets: inherit\n\n build-azure-uber-jar:\n needs: [setup, owasp-scanner]\n uses: ./.github/workflows/build-azure-uber-jar.yml\n with:\n branch: ${{ needs.setup.outputs.branch }}\n liquibase-version: ${{ needs.setup.outputs.version }}\n secrets: inherit\n\n build-extension-jars:\n needs: [setup, owasp-scanner]\n uses: ./.github/workflows/build-extension-jars.yml\n with:\n liquibase-version: ${{ needs.setup.outputs.version }}\n dependencies: ${{ needs.setup.outputs.dependencies }}\n extensions: ${{ needs.setup.outputs.extensions }}\n branch: ${{ needs.setup.outputs.branch }}\n secrets: inherit\n\n get-liquibase-checks-version:\n needs: [setup]\n if: ${{ contains(needs.setup.outputs.extensions, 'liquibase-checks') }}\n runs-on: ubuntu-latest\n outputs:\n version: ${{ steps.extract-version.outputs.version }}\n steps:\n - uses: actions/checkout@v4\n with:\n repository: liquibase/liquibase-checks\n token: ${{ secrets.BOT_TOKEN }}\n\n - id: extract-version\n name: Extract version from pom.xml\n shell: bash\n run: |\n VERSION=$(grep '<version>' pom.xml | head -n 1 | sed 's/.*<version>\(.*\)-SNAPSHOT<\/version>.*/\1/')\n echo "version=$VERSION" >> $GITHUB_OUTPUT\n\n reversion:\n needs: [setup,build-azure-uber-jar,build-extension-jars,get-liquibase-checks-version]\n name: Re-version artifacts ${{ needs.setup.outputs.version }}\n runs-on: ubuntu-22.04\n steps:\n - uses: actions/checkout@v4\n\n - uses: actions/checkout@v4\n name: Checkout liquibase-pro\n with:\n repository: liquibase/liquibase-pro\n ref: "${{ needs.setup.outputs.branch }}"\n path: download/repo/liquibase-pro\n token: ${{ secrets.BOT_TOKEN }}\n\n - name: Download liquibase-artifacts\n uses: dawidd6/action-download-artifact@v8\n with:\n workflow: run-tests.yml\n run_id: ${{ needs.setup.outputs.runId }}\n name: liquibase-artifacts\n path: download/liquibase-artifacts\n\n - name: Get Current Run ID\n id: get_run_id\n run: |\n run_id=${{ github.run_id }}\n echo "uber_jar_runId=${run_id}" >> $GITHUB_OUTPUT\n\n - name: Download liquibase-pro-azure-artifacts\n uses: actions/download-artifact@v4\n with:\n name: liquibase-pro-azure-artifacts\n path: liquibase-pro/liquibase-azure-deps\n\n - name: maven-settings-xml-action\n uses: whelk-io/maven-settings-xml-action@v22\n with:\n repositories: |\n [\n {\n "id": "liquibase",\n "url": "https://maven.pkg.github.com/liquibase/liquibase",\n "releases": {\n "enabled": "true"\n },\n "snapshots": {\n "enabled": "true",\n "updatePolicy": "always"\n }\n },\n {\n "id": "liquibase-pro",\n "url": "https://maven.pkg.github.com/liquibase/liquibase-pro",\n "releases": {\n "enabled": "true"\n },\n "snapshots": {\n "enabled": "true",\n "updatePolicy": "always"\n }\n }\n ]\n servers: |\n [\n {\n "id": "liquibase",\n "username": "liquibot",\n "password": "${{ secrets.LIQUIBOT_PAT_GPM_ACCESS }}"\n },\n {\n "id": "liquibase-pro",\n "username": "liquibot",\n "password": "${{ secrets.LIQUIBOT_PAT_GPM_ACCESS }}"\n }\n ]\n\n - name: Get extensions artifacts\n run: |\n IFS=',' read -ra ADDR <<< "${{ needs.setup.outputs.extensions }}"\n for extension in "${ADDR[@]}"; do\n if [ "$extension" != "liquibase-checks" ]; then\n mvn dependency:get -DgroupId=org.liquibase.ext -DartifactId=$extension -Dversion=${{ needs.setup.outputs.version }} -Dtransitive=false || echo "Failed to download $extension artifact"\n else\n mvn dependency:get -DgroupId=org.liquibase.ext -DartifactId=$extension -Dversion=${{ needs.get-liquibase-checks-version.outputs.version }} -Dtransitive=false || echo "Failed to download $extension artifact"\n fi\n done\n\n - name: Set up JDK\n uses: actions/setup-java@v4\n with:\n java-version: "8"\n distribution: "adopt"\n gpg-private-key: ${{ secrets.GPG_SECRET }}\n gpg-passphrase: GPG_PASSPHRASE\n env:\n GPG_PASSWORD: ${{ secrets.GPG_PASSPHRASE }}\n\n - name: Re-version Artifacts\n env:\n GPG_PASSWORD: ${{ secrets.GPG_PASSPHRASE }}\n run: |\n scripts_branch=${{ github.ref }}\n mkdir -p $PWD/.github/util/\n # Download a script (re-version.sh) from a URL and save it to the specified directory\n curl -o $PWD/.github/util/re-version.sh https://raw.githubusercontent.com/liquibase/liquibase/$scripts_branch/.github/util/re-version.sh\n\n # Download another script (sign-artifacts.sh) from a URL and save it to the specified directory\n curl -o $PWD/.github/util/sign-artifacts.sh https://raw.githubusercontent.com/liquibase/liquibase/$scripts_branch/.github/util/sign-artifacts.sh\n curl -o $PWD/.github/util/ManifestReversion.java https://raw.githubusercontent.com/liquibase/liquibase/$scripts_branch/.github/util/ManifestReversion.java\n chmod +x $PWD/.github/util/re-version.sh\n chmod +x $PWD/.github/util/ManifestReversion.java\n chmod +x $PWD/.github/util/sign-artifacts.sh\n $PWD/.github/util/re-version.sh download/liquibase-artifacts "${{ needs.setup.outputs.version }}" "${{ needs.setup.outputs.branch }}"\n\n # Execute the sign-artifacts.sh script with specific arguments\n $PWD/.github/util/sign-artifacts.sh download/liquibase-artifacts "${{ needs.setup.outputs.version }}" "${{ needs.setup.outputs.branch }}"\n\n ## Sign Files\n ## liquibase-azure-deps and liquibase extensions are already on its correct version. Check reusable workflow: build-azure-uber-jar.yml and build-extension-jars.yml\n mv liquibase-pro/liquibase-azure-deps/* re-version/out\n\n # Modify the zip file\n unzip re-version/out/liquibase-${{ needs.setup.outputs.version }}.zip -d re-version/out/liquibase-${{ needs.setup.outputs.version }}\n mkdir -p re-version/out/liquibase-${{ needs.setup.outputs.version }}/internal/extensions\n rm -rf re-version/out/liquibase-${{ needs.setup.outputs.version }}.zip\n IFS=',' read -ra EXT <<< "${{ needs.setup.outputs.extensions }}"\n for i in "${EXT[@]}"; do\n if [ "$i" != "liquibase-checks" ]; then\n cp ~/.m2/repository/org/liquibase/ext/$i/${{ needs.setup.outputs.version }}/$i-${{ needs.setup.outputs.version }}.jar re-version/out/liquibase-${{ needs.setup.outputs.version }}/internal/extensions/$i.jar || echo "Failed to move $i artifact"\n else\n cp ~/.m2/repository/org/liquibase/ext/$i/${{ needs.get-liquibase-checks-version.outputs.version }}/$i-${{ needs.get-liquibase-checks-version.outputs.version }}.jar re-version/out/liquibase-${{ needs.setup.outputs.version }}/internal/extensions/$i.jar || echo "Failed to move $i artifact"\n fi\n done\n (cd re-version/out/liquibase-${{ needs.setup.outputs.version }} && zip -r ../liquibase-${{ needs.setup.outputs.version }}.zip . && cd .. && rm -rf liquibase-${{ needs.setup.outputs.version }})\n\n # Modify the tar.gz file\n mkdir -p re-version/out/liquibase-${{ needs.setup.outputs.version }}\n tar -xzvf re-version/out/liquibase-${{ needs.setup.outputs.version }}.tar.gz -C re-version/out/liquibase-${{ needs.setup.outputs.version }}\n rm -rf re-version/out/liquibase-${{ needs.setup.outputs.version }}.tar.gz\n mkdir -p re-version/out/liquibase-${{ needs.setup.outputs.version }}/internal/extensions\n for I in "${EXT[@]}"; do\n if [ "$I" != "liquibase-checks" ]; then\n cp ~/.m2/repository/org/liquibase/ext/$I/${{ needs.setup.outputs.version }}/$I-${{ needs.setup.outputs.version }}.jar re-version/out/liquibase-${{ needs.setup.outputs.version }}/internal/extensions/$I.jar || echo "Failed to move $I artifact"\n else\n cp ~/.m2/repository/org/liquibase/ext/$I/${{ needs.get-liquibase-checks-version.outputs.version }}/$I-${{ needs.get-liquibase-checks-version.outputs.version }}.jar re-version/out/liquibase-${{ needs.setup.outputs.version }}/internal/extensions/$I.jar || echo "Failed to move $I artifact"\n fi\n done\n (cd re-version/out/liquibase-${{ needs.setup.outputs.version }} && tar -czvf ../liquibase-${{ needs.setup.outputs.version }}.tar.gz * && cd .. && rm -rf liquibase-${{ needs.setup.outputs.version }})\n\n $PWD/.github/util/sign-artifacts.sh re-version/out\n\n # Move files to a specific directory\n mkdir re-version/final re-version/s3-minimal\n mv re-version/out/liquibase-core-${{ needs.setup.outputs.version }}.jar re-version/final\n mv re-version/out/liquibase-commercial-${{ needs.setup.outputs.version }}.jar re-version/final\n mv re-version/out/liquibase-${{ needs.setup.outputs.version }}.tar.gz re-version/final\n mv re-version/out/liquibase-${{ needs.setup.outputs.version }}.zip re-version/final\n mv re-version/out/liquibase-minimal-${{ needs.setup.outputs.version }}.zip re-version/s3-minimal\n mv re-version/out/liquibase-azure-deps-${{ needs.setup.outputs.version }}.jar re-version/final/liquibase-azure-deps-${{ needs.setup.outputs.version }}.jar\n (cd re-version/out/ && zip liquibase-additional-${{ needs.setup.outputs.version }}.zip *)\n mv re-version/out/liquibase-additional-${{ needs.setup.outputs.version }}.zip re-version/final\n\n - name: Cache Completed Artifacts\n uses: actions/cache@v4.2.0\n with:\n key: completed-artifacts-${{ github.run_number }}-${{ github.run_attempt }}\n path: re-version/final\n\n - name: Cache Minimal Completed Artifacts\n uses: actions/cache@v4.2.0\n with:\n key: minimal-artifacts-${{ github.run_number }}-${{ github.run_attempt }}\n path: re-version/s3-minimal\n\n - name: Set repository tags\n if: ${{ inputs.standalone_zip == false && inputs.dry_run == false }}\n run: |\n git tag -f v${{ needs.setup.outputs.version }}\n git push -f origin v${{ needs.setup.outputs.version }}\n (cd download/repo/liquibase-pro && git tag -f v${{ needs.setup.outputs.version }})\n (cd download/repo/liquibase-pro && git push -f origin v${{ needs.setup.outputs.version }})\n\n build-installers:\n permissions:\n contents: write # for softprops/action-gh-release to create GitHub release\n needs: [setup, reversion]\n name: Build Installers\n runs-on: ubuntu-22.04\n env:\n GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}\n steps:\n - uses: actions/checkout@v4\n\n - name: Restore Completed Artifacts\n uses: actions/cache@v4.2.0\n with:\n key: completed-artifacts-${{ github.run_number }}-${{ github.run_attempt }}\n path: re-version/final\n\n - name: Set up JDK for GPG\n uses: actions/setup-java@v4\n with:\n java-version: "8"\n distribution: "adopt"\n gpg-private-key: ${{ secrets.GPG_SECRET }}\n gpg-passphrase: GPG_PASSPHRASE\n env:\n GPG_PASSWORD: ${{ secrets.GPG_PASSPHRASE }}\n\n - name: Decode Client Authentication Certificate\n run: |\n echo "${{ secrets.SM_CLIENT_CERT_FILE_B64 }}" | base64 --decode > /tmp/client-auth.p12\n\n - name: Set Environment Variables for Signing\n run: |\n echo "SM_HOST=${{ secrets.SM_HOST }}" >> "$GITHUB_ENV"\n echo "SM_API_KEY=${{ secrets.SM_API_KEY }}" >> "$GITHUB_ENV"\n echo "SM_CLIENT_CERT_FILE=/tmp/client-auth.p12" >> "$GITHUB_ENV"\n echo "SM_CLIENT_CERT_PASSWORD=${{ secrets.SM_CLIENT_CERT_PASSWORD }}" >> "$GITHUB_ENV"\n echo "SM_CODE_SIGNING_CERT_SHA1_HASH=${{ secrets.SM_CODE_SIGNING_CERT_SHA1_HASH }}" >> "$GITHUB_ENV"\n echo "SM_KEY_PAIR_ALIAS=${{ secrets.SM_KEY_PAIR_ALIAS }}" >> "$GITHUB_ENV"\n\n - name: Install jsign and osslsigncode\n run: |\n curl -fSslL https://github.com/ebourg/jsign/releases/download/3.1/jsign_3.1_all.deb -o jsign_3.1_all.deb\n sudo dpkg --install jsign_3.1_all.deb\n sudo apt-get install osslsigncode\n \n - name: Create pkcs11properties.cfg\n run: |\n cat << EOF > pkcs11properties.cfg\n name=signingmanager\n library="$(pwd)/.github/util/smpkcs11.so"\n slotListIndex=0\n EOF\n mv pkcs11properties.cfg $(pwd)/.github/util/\n\n - name: Install DigiCert KeyLocker Client Tools\n uses: digicert/ssm-code-signing@v1.0.0\n\n - name: Build Unsigned Windows Installer\n env:\n INSTALL4J_LICENSE_KEY: ${{ secrets.INSTALL4J_10_LICENSE }}\n run: |\n version="${{ needs.setup.outputs.version }}"\n tarFile=$(pwd)/re-version/final/liquibase-$version.tar.gz\n scriptDir=$(pwd)/.github/util/\n\n mkdir -p liquibase-dist/target/liquibase-$version\n (cd liquibase-dist/target/liquibase-$version && tar xfz $tarFile)\n (cd liquibase-dist && $scriptDir/package-install4j.sh $version)\n mv liquibase-dist/target/liquibase-*-installer-* re-version/final\n\n - name: Find Unsigned Windows Installer\n id: find-installer\n run: |\n INSTALLER_PATH=$(find re-version/final -name "liquibase-*-installer-*" | head -n 1)\n if [ -z "$INSTALLER_PATH" ]; then\n echo "Error: No installer file found!"\n exit 1\n fi\n echo "Found installer: $INSTALLER_PATH"\n echo "INSTALLER_PATH=$INSTALLER_PATH" >> $GITHUB_ENV\n\n - name: Sign Windows Installer using KeyLocker\n run: |\n smctl sign -v --fingerprint "$SM_CODE_SIGNING_CERT_SHA1_HASH" \\n --keypair-alias $SM_KEY_PAIR_ALIAS \\n --certificate "$SM_CLIENT_CERT_FILE" \\n --config-file "$(pwd)/.github/util/pkcs11properties.cfg" \\n --input "$INSTALLER_PATH" --tool jsign\n\n - name: Verify Windows Installer Signature\n run: osslsigncode verify -CAfile /etc/ssl/certs/ca-certificates.crt -in "$INSTALLER_PATH"\n \n - name: Re-version Installers\n env:\n GPG_PASSWORD: ${{ secrets.GPG_PASSPHRASE }}\n run: |\n version="${{ needs.setup.outputs.version }}"\n\n ##Sign Files\n $PWD/.github/util/sign-artifacts.sh re-version/final\n\n (cd re-version/final && zip liquibase-additional-$version.zip *.asc *.md5 *.sha1)\n rm re-version/final/*.asc\n rm re-version/final/*.md5\n rm re-version/final/*.sha1\n\n # Check if the tag already exists in OSS repository, if it does, exit with error. This is to prevent the artifacts from being attached to the already released tag.\n - name: Check OSS release tag existence, if tag exists, exit with error\n run: |\n tag="v${{ needs.setup.outputs.version }}"\n if git show-ref --tags --quiet --verify -- "refs/tags/$tag"\n then\n echo "Tag $tag already exists"\n exit 1\n fi\n\n - name: Attach Files to Draft Release\n if: ${{ inputs.standalone_zip == false && inputs.dry_run == false }}\n uses: softprops/action-gh-release@v2\n with:\n tag_name: v${{ needs.setup.outputs.version }}\n fail_on_unmatched_files: true\n body: Liquibase ${{ needs.setup.outputs.version }}\n generate_release_notes: true\n draft: true\n files: re-version/final/*\n env:\n GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}\n\n - name: Attach Files to Dry-Run Draft Release\n id: attach-files-dry-run\n if: ${{ inputs.standalone_zip == false && inputs.dry_run == true }}\n uses: softprops/action-gh-release@v2\n with:\n tag_name: v${{ needs.setup.outputs.version }}\n fail_on_unmatched_files: true\n body: Liquibase ${{ needs.setup.outputs.version }} (Dry-Run)\n generate_release_notes: true\n draft: true\n files: re-version/final/*\n env:\n GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}\n\n - name: Attach standalone zip to Build\n if: ${{ inputs.standalone_zip == true && inputs.dry_run == false }}\n uses: actions/upload-artifact@v4\n with:\n name: liquibase-installers-${{ needs.setup.outputs.version }}\n path: re-version/final/*\n\n deploy_minimal_artifact_s3:\n if: ${{ inputs.dry_run == false }}\n name: Deploy minimal artifacts to lbio s3 bucket\n needs: [setup, reversion]\n permissions:\n id-token: write\n contents: read\n runs-on: ubuntu-latest\n steps:\n - name: Restore Minimal Completed Artifacts\n uses: actions/cache@v4.2.0\n with:\n key: minimal-artifacts-${{ github.run_number }}-${{ github.run_attempt }}\n path: re-version/s3-minimal\n\n - name: Configure AWS credentials\n uses: aws-actions/configure-aws-credentials@v4\n with:\n role-to-assume: ${{ secrets.AWS_IO_DEV_GITHUB_OIDC_ROLE_ARN_LIQUIBASE }}\n aws-region: us-east-2\n\n - name: Publish minimal assets to s3 bucket\n run: |\n aws s3 sync "re-version/s3-minimal" s3://${{ secrets.LBIO_DOWNLOADS_S3_BUCKET }}/downloads/liquibase-minimal/v${{ needs.setup.outputs.version }}/ --only-show-errors\n
dataset_sample\yaml\liquibase_liquibase\.github\workflows\create-release.yml
create-release.yml
YAML
20,690
0.95
0.050209
0.02331
vue-tools
848
2023-12-18T14:23:17.218212
Apache-2.0
false
5d8f9d9ae04c7634cf5eed2344fbaae1
name: Dry run release\n\non:\n workflow_dispatch: # Trigger on demand\n schedule: # Trigger weekly all Wednesdays at midnight UTC\n # Trigger weekly on Wednesday at midnight Austin time (Standard Time)\n - cron: "0 6 * * 3"\n\njobs:\n setup:\n runs-on: ubuntu-latest\n outputs:\n dry_run_id: ${{ steps.get_run_id.outputs.dry_run_id }}\n dry_run_branch_name: ${{ steps.get_run_id.outputs.dry_run_branch_name }}\n steps:\n - name: Get run-tests.yml runId\n id: get_run_id\n run: |\n # Fetch the list of workflow runs\n response=$(curl -s \\n -H "Authorization: token ${{ secrets.BOT_TOKEN }}" \\n -H "Accept: application/vnd.github.v3+json" \\n "https://api.github.com/repos/liquibase/liquibase/actions/workflows/run-tests.yml/runs?branch=master&status=success&per_page=1")\n # Extract the last successful run ID\n run_id=$(echo "$response" | jq -r '.workflow_runs[0].id')\n echo "dry_run_id=$run_id" >> $GITHUB_OUTPUT\n echo "dry_run_branch_name=${{ github.ref_name }}" >> $GITHUB_OUTPUT\n\n dry-run-create-release:\n needs: [setup]\n runs-on: ubuntu-latest\n outputs:\n dry_run_zip_url: ${{ steps.wait_and_get_outputs.outputs.dry_run_zip_url }}\n dry_run_tar_gz_url: ${{ steps.wait_and_get_outputs.outputs.dry_run_tar_gz_url }}\n dry_run_release_id: ${{ steps.wait_and_get_outputs.outputs.dry_run_release_id }}\n steps:\n - name: Invoke Create Release\n id: trigger-step\n uses: the-actions-org/workflow-dispatch@v4\n with:\n workflow: create-release.yml\n token: ${{ secrets.BOT_TOKEN }}\n inputs: '{"version": "dry-run-${{ github.run_id }}", "runId": "${{ needs.setup.outputs.dry_run_id }}", "standalone_zip": "false", "dry_run": "true"}'\n ref: master\n wait-for-completion: true\n\n - name: Get Release Outputs\n id: wait_and_get_outputs\n run: |\n # Get release info based on draft name\n release_name="vdry-run-${{ github.run_id }}"\n response=$(curl -s -H "Authorization: token ${{ secrets.BOT_TOKEN }}" \\n -H "Accept: application/vnd.github.v3+json" \\n "https://api.github.com/repos/liquibase/liquibase/releases")\n\n # Find the draft release and get its assets\n draft_release=$(echo "$response" | jq -r --arg name "$release_name" '.[] | select(.name == $name and .draft == true)')\n assets_url=$(echo "$draft_release" | jq -r '.assets_url')\n dry_run_release_id=$(echo "$draft_release" | jq -r '.id')\n\n # Get the assets\n assets_response=$(curl -s -H "Authorization: token ${{ secrets.BOT_TOKEN }}" \\n -H "Accept: application/vnd.github.v3+json" \\n "$assets_url")\n\n # Get URLs for zip and tar.gz files - getting browser_download_url instead of url\n dry_run_zip_url=$(echo "$assets_response" | jq -r '.[] | select(.name | test("^liquibase-dry-run-.*\\.zip$")) | .browser_download_url')\n dry_run_tar_gz_url=$(echo "$assets_response" | jq -r '.[] | select(.name | test("^liquibase-dry-run-.*\\.tar\\.gz$")) | .browser_download_url')\n\n # Remove any newlines or spaces from URLs before setting outputs\n dry_run_zip_url=$(echo "$dry_run_zip_url" | tr -d '[:space:]')\n dry_run_tar_gz_url=$(echo "$dry_run_tar_gz_url" | tr -d '[:space:]')\n\n echo "dry_run_release_id Id is $dry_run_release_id"\n echo "dry_run_tar_gz_url is $dry_run_tar_gz_url"\n echo "dry_run_zip_url is $dry_run_zip_url"\n\n echo "dry_run_tar_gz_url=$dry_run_tar_gz_url" >> $GITHUB_OUTPUT\n echo "dry_run_zip_url=$dry_run_zip_url" >> $GITHUB_OUTPUT\n echo "dry_run_release_id=$dry_run_release_id" >> $GITHUB_OUTPUT\n\n dry-run-release-published:\n needs: [setup, dry-run-create-release]\n runs-on: ubuntu-latest\n steps:\n - name: Trigger release-published workflow\n uses: the-actions-org/workflow-dispatch@v4\n with:\n workflow: release-published.yml\n token: ${{ secrets.BOT_TOKEN }}\n inputs: '{"tag": "vdry-run-${{ github.run_id }}", "dry_run_release_id": "${{ needs.dry-run-create-release.outputs.dry_run_release_id }}", "dry_run_zip_url": "${{ needs.dry-run-create-release.outputs.dry_run_zip_url }}", "dry_run_tar_gz_url": "${{ needs.dry-run-create-release.outputs.dry_run_tar_gz_url }}", "dry_run": "true", "dry_run_branch_name": "${{ needs.setup.outputs.dry_run_branch_name }}"}'\n ref: master\n wait-for-completion: true\n workflow-logs: json-output\n\n cleanup:\n runs-on: ubuntu-latest\n if: always()\n needs: [setup, dry-run-create-release, dry-run-release-published]\n steps:\n - name: Checkout liquibase\n uses: actions/checkout@v4\n\n - name: Set up Git\n run: |\n git config user.name "liquibot"\n git config user.email "liquibot@liquibase.org"\n\n - name: Delete liquibase dry-run tag\n if: always()\n run: |\n git push origin --delete refs/tags/vdry-run-${{ github.run_id }}\n echo "Remote tag vdry-run-${{ github.run_id }} deleted"\n\n - name: Delete the dry-run draft release\n if: always()\n run: |\n curl -X DELETE -H "Authorization: token ${{ secrets.BOT_TOKEN }}" \\n -H "Accept: application/vnd.github.v3+json" \\n "https://api.github.com/repos/${{ github.repository }}/releases/${{ needs.dry-run-create-release.outputs.dry_run_release_id}}"\n\n notify:\n if: failure()\n runs-on: ubuntu-latest\n needs:\n [\n setup,\n dry-run-create-release,\n dry-run-release-published,\n cleanup,\n ]\n steps:\n - name: Notify Slack on Build Failure\n uses: rtCamp/action-slack-notify@v2\n env:\n SLACK_COLOR: failure\n SLACK_MESSAGE: "View details on GitHub Actions: ${{ github.server_url }}/${{ github.repository }}/actions/runs/${{ github.run_id }} <@U040C8J8143> <@U04P39MS2SW> <@UHHJ6UAEQ> <@U042HRTL4DT>" # Jandro, Sailee, Jake, Filipe\n SLACK_TITLE: "❌ ${{ github.repository }} ❌ Build failed on branch ${{ github.ref }} for commit ${{ github.sha }} in repository ${{github.repository}}"\n SLACK_USERNAME: liquibot\n SLACK_WEBHOOK: ${{ secrets.DRY_RUN_RELEASE_SLACK_WEBHOOK }}\n SLACK_ICON_EMOJI: ":robot_face:"\n SLACK_FOOTER: "${{ github.repository }}"\n SLACK_LINK_NAMES: true\n
dataset_sample\yaml\liquibase_liquibase\.github\workflows\dry-run-release.yml
dry-run-release.yml
YAML
6,568
0.8
0.055944
0.062992
python-kit
265
2024-04-19T07:12:54.402839
BSD-3-Clause
false
50ffe3a06d608f4d33499db7b2f9f407
# Name of the GitHub Action workflow\nname: FOSSA License Compliance and Security Check\n\n# Event triggers for the workflow\non:\n workflow_dispatch: # Run this action manually from the GitHub Actions tab\n release:\n types: [published] # Run this action when a release is published\n \n# Define the jobs in the workflow\njobs:\n fossa-scan:\n # Specifies the type of runner to use\n runs-on: ubuntu-latest\n\n # Sequence of steps that make up a single job\n steps:\n # Checkout the code to the GitHub runner\n - name: Checkout Code\n uses: actions/checkout@v4\n \n - name: Set up JDK\n uses: actions/setup-java@v4\n with:\n distribution: 'temurin'\n java-version: '17'\n \n #look for dependencies in maven\n - name: maven-settings-xml-action\n uses: whelk-io/maven-settings-xml-action@v22\n with:\n repositories: |\n [\n {\n "id": "liquibase",\n "url": "https://maven.pkg.github.com/liquibase/liquibase",\n "releases": {\n "enabled": "false"\n },\n "snapshots": {\n "enabled": "true",\n "updatePolicy": "always"\n }\n },\n {\n "id": "liquibase-pro",\n "url": "https://maven.pkg.github.com/liquibase/liquibase-pro",\n "releases": {\n "enabled": "false"\n },\n "snapshots": {\n "enabled": "true",\n "updatePolicy": "always"\n }\n }\n ]\n servers: |\n [\n {\n "id": "liquibase-pro",\n "username": "liquibot",\n "password": "${{ secrets.LIQUIBOT_PAT }}"\n },\n {\n "id": "liquibase",\n "username": "liquibot",\n "password": "${{ secrets.LIQUIBOT_PAT }}"\n }\n ]\n\n - name: Install Maven dependencies (excluding test scope)\n run: mvn clean install -DskipTests\n \n - name: run FOSSA CLI\n uses: fossas/fossa-action@main\n with:\n api-key: ${{ secrets.FOSSA_API_KEY }}\n branch: ${{ github.ref }}\n\n - name: Get the commit sha\n id: get_commit_sha_oss\n run: |\n commit_sha_lb=`(git rev-parse HEAD)`\n echo "commit_sha_lb=${commit_sha_lb}" >> $GITHUB_OUTPUT\n\n # from the below curl commands we output the fossa report with the headers we need. Following this documentation : https://docs.fossa.com/docs/download-fossa-project-attribution-reports\n - name: Set the dependency metadata information\n run: |\n mkdir /home/runner/fossa_reports\n curl --location --request PUT 'https://app.fossa.com/api/projects/git+github.com/liquibase/liquibase' \\n --header 'Authorization: Bearer ${{ secrets.FOSSA_API_KEY }}' \\n --header 'Content-Type: application/x-www-form-urlencoded' \\n --data-urlencode 'bom_column_settings%5B%5D=Title' \\n --data-urlencode 'bom_column_settings%5B%5D=Version' \\n --data-urlencode 'bom_column_settings%5B%5D=Authors' \\n --data-urlencode 'bom_column_settings%5B%5D=DeclaredLicense' \\n --data-urlencode 'bom_column_settings%5B%5D=PackageHomepage' \\n \n\n curl --location 'https://app.fossa.com/api/revisions/git%2Bgithub.com%2Fliquibase%2Fliquibase%24${{ steps.get_commit_sha_oss.outputs.commit_sha_lb }}/attribution/download?format=CSV&includeDeepDependencies=true&download=true' \\n --header 'Authorization: Bearer ${{ secrets.FOSSA_API_KEY }}' -o /home/runner/fossa_reports/liquibase.csv\n\n # Upload to build page for the team to check the csv file contents.\n - name: Upload to build page\n uses: actions/upload-artifact@v4\n with:\n name: fossa-reports\n path: /home/runner/fossa_reports/liquibase.csv\n\n # Upload report to S3\n - name: Upload report to S3\n if: always()\n run: aws s3 cp /home/runner/fossa_reports/liquibase.csv s3://liquibaseorg-origin/fossa_reports_liquibase-pro/\n env:\n AWS_ACCESS_KEY_ID: ${{ secrets.LIQUIBASEORIGIN_ACCESS_KEY_ID }}\n AWS_SECRET_ACCESS_KEY: ${{ secrets.LIQUIBASEORIGIN_SECRET_ACCESS_KEY }}\n AWS_DEFAULT_REGION: us-east-1\n\n # trigger a job in liquibase-pro for the fossa-scan report generation and merging of OSS and PRO reports \n run-pro-fossa-report:\n runs-on: ubuntu-latest\n needs: [ fossa-scan ]\n steps:\n - name: Dispatch event to PRO repository\n uses: peter-evans/repository-dispatch@v3\n with:\n token: ${{ secrets.LIQUIBOT_PAT_GPM_ACCESS }}\n repository: liquibase/liquibase-pro\n event-type: FOSSA_Scan_PRO\n client-payload: '{"repository": "${{ github.repository }}", "branch": "${{ github.ref }}", "sha": "${{ github.sha }}", "server_url": "${{ github.server_url }}", "run_id": "${{ github.run_id }}"}'\n
dataset_sample\yaml\liquibase_liquibase\.github\workflows\fossa.yml
fossa.yml
YAML
4,931
0.8
0.038462
0.094828
awesome-app
62
2024-07-11T03:03:59.676585
BSD-3-Clause
false
5ff92b810b1b19802b6547c1c606bfa3
name: FOSSA License Compliance and Security Check\n\non:\n repository_dispatch:\n types: [ oss-released-version ]\n\njobs:\n fossa:\n uses: liquibase/build-logic/.github/workflows/fossa_ai.yml@main\n secrets: inherit\n with:\n latest_version: ${{ github.event.client_payload.latest_version }}\n repo_name: "liquibase/liquibase"\n
dataset_sample\yaml\liquibase_liquibase\.github\workflows\fossa_ai.yml
fossa_ai.yml
YAML
340
0.7
0
0
node-utils
261
2024-02-05T15:33:06.233920
Apache-2.0
false
a0cdc5f2fe7015f436605fea69d86143
name: Build Test Installers\n\non:\n workflow_dispatch:\n push:\n branches:\n - gha-scheduled-*\n\njobs:\n build-installers:\n name: Build Installers\n runs-on: ubuntu-22.04\n steps:\n - uses: actions/checkout@v4\n\n - name: Set up JDK for GPG\n uses: actions/setup-java@v4\n with:\n java-version: "17"\n distribution: "temurin"\n gpg-private-key: ${{ secrets.GPG_SECRET }}\n gpg-passphrase: GPG_PASSPHRASE\n overwrite-settings: false\n env:\n GPG_PASSWORD: ${{ secrets.GPG_PASSPHRASE }}\n\n # TODO: Add install4j and DigiCert KeyLocker Client Tools caching if possible\n # - name: Cache Install4j\n # id: cache-install4j\n # uses: actions/cache@v4\n # with:\n # path: /usr/local/bin/install4jc\n # key: install4j-${{ runner.os }}-10_0_9\n\n # - name: Install Install4j (if not cached)\n # if: steps.cache-install4j.outputs.cache-hit != 'true'\n # run: |\n # wget -nv -O install4j.deb "https://download.ej-technologies.com/install4j/install4j_linux-x64_10_0_9.deb"\n # sudo apt install -y ./install4j.deb\n # rm install4j.deb\n\n - name: Decode Client Authentication Certificate\n run: |\n echo "${{ secrets.SM_CLIENT_CERT_FILE_B64 }}" | base64 --decode > /tmp/client-auth.p12\n\n - name: Set Environment Variables for Signing\n run: |\n echo "SM_HOST=${{ secrets.SM_HOST }}" >> "$GITHUB_ENV"\n echo "SM_API_KEY=${{ secrets.SM_API_KEY }}" >> "$GITHUB_ENV"\n echo "SM_CLIENT_CERT_FILE=/tmp/client-auth.p12" >> "$GITHUB_ENV"\n echo "SM_CLIENT_CERT_PASSWORD=${{ secrets.SM_CLIENT_CERT_PASSWORD }}" >> "$GITHUB_ENV"\n echo "SM_CODE_SIGNING_CERT_SHA1_HASH=${{ secrets.SM_CODE_SIGNING_CERT_SHA1_HASH }}" >> "$GITHUB_ENV"\n echo "SM_KEY_PAIR_ALIAS=${{ secrets.SM_KEY_PAIR_ALIAS }}" >> "$GITHUB_ENV"\n\n - name: Install jsign and osslsigncode\n run: |\n curl -fSslL https://github.com/ebourg/jsign/releases/download/3.1/jsign_3.1_all.deb -o jsign_3.1_all.deb\n sudo dpkg --install jsign_3.1_all.deb\n sudo apt-get install osslsigncode\n\n - name: Create pkcs11properties.cfg\n run: |\n cat << EOF > pkcs11properties.cfg\n name=signingmanager\n library="$(pwd)/.github/util/smpkcs11.so"\n slotListIndex=0\n EOF\n mv pkcs11properties.cfg $(pwd)/.github/util/\n\n - name: Install DigiCert KeyLocker Client Tools\n uses: digicert/ssm-code-signing@v1.0.0\n\n - name: Set branch variable\n run: echo "BRANCH=${GITHUB_REF#refs/heads/}" >> $GITHUB_ENV\n\n - name: Download liquibase-artifacts\n uses: dawidd6/action-download-artifact@v8\n with:\n workflow: run-tests.yml\n workflow_conclusion: success\n branch: ${{ env.BRANCH }}\n name: liquibase-artifacts\n path: re-version/final\n\n - name: Build Unsigned Windows Installer\n env:\n INSTALL4J_LICENSE_KEY: ${{ secrets.INSTALL4J_10_LICENSE }}\n run: |\n tarFile=$(pwd)/re-version/final/liquibase-${BRANCH}-SNAPSHOT.tar.gz\n scriptDir=$(pwd)/.github/util/\n\n mkdir -p liquibase-dist/target/liquibase-0-SNAPSHOT\n (cd liquibase-dist/target/liquibase-0-SNAPSHOT && tar xfz $tarFile)\n (cd liquibase-dist && $scriptDir/package-install4j.sh 0-SNAPSHOT)\n mv liquibase-dist/target/liquibase-*-installer-* re-version/final\n\n - name: Find Unsigned Windows Installer\n id: find-installer\n run: |\n INSTALLER_PATH=$(find re-version/final -name "liquibase-*-installer-*" | head -n 1)\n if [ -z "$INSTALLER_PATH" ]; then\n echo "Error: No installer file found!"\n exit 1\n fi\n echo "Found installer: $INSTALLER_PATH"\n echo "INSTALLER_PATH=$INSTALLER_PATH" >> $GITHUB_ENV\n\n - name: Sign Windows Installer using KeyLocker\n run: |\n smctl sign -v --fingerprint "$SM_CODE_SIGNING_CERT_SHA1_HASH" \\n --keypair-alias $SM_KEY_PAIR_ALIAS \\n --certificate "$SM_CLIENT_CERT_FILE" \\n --config-file "$(pwd)/.github/util/pkcs11properties.cfg" \\n --input "$INSTALLER_PATH" --tool jsign\n \n - name: Verify Windows Installer Signature\n run: osslsigncode verify -CAfile /etc/ssl/certs/ca-certificates.crt -in "$INSTALLER_PATH"\n\n - name: GPG Sign Artifacts (After Code Signing the Windows Installer)\n env:\n GPG_PASSWORD: ${{ secrets.GPG_PASSPHRASE }}\n run: |\n .github/util/sign-artifacts.sh re-version/final\n\n - name: Re-version Installers\n run: |\n (cd re-version/final && zip liquibase-additional-0-SNAPSHOT.zip *.asc *.md5 *.sha1)\n rm re-version/final/*.asc\n rm re-version/final/*.md5\n rm re-version/final/*.sha1\n\n - name: Archive Packages\n uses: actions/upload-artifact@v4\n with:\n name: liquibase-test-packages\n path: re-version/final/*\n
dataset_sample\yaml\liquibase_liquibase\.github\workflows\installer-build-check.yml
installer-build-check.yml
YAML
5,094
0.8
0.044118
0.112069
react-lib
524
2024-07-27T16:24:17.099408
GPL-3.0
false
1ac65287e5bd38ff4cd918dea34c0415
name: Pull Request Labels\non:\n pull_request:\n types: [opened, labeled, unlabeled, synchronize, reopened]\njobs:\n label:\n runs-on: ubuntu-latest\n permissions:\n issues: write\n pull-requests: write\n steps:\n - name: Validate PR Labels\n uses: mheap/github-action-required-labels@v5\n with:\n mode: minimum\n count: 1\n labels: "APIBreakingChanges, breakingChanges, newContributors, notableChanges, sdou, skipReleaseNotes, TypeBug, TypeEnhancement, TypeTest"\n add_comment: true\n message: "Label error: This PR is being prevented from merging because you have not added one of the labels: {{ provided }}. You'll need to add it before this PR can be merged."
dataset_sample\yaml\liquibase_liquibase\.github\workflows\label-pr.yml
label-pr.yml
YAML
733
0.85
0
0
python-kit
640
2024-01-02T06:31:27.493401
Apache-2.0
false
6309de7536431c0b9ef276db9cda3d40
name: Release Extensions\non:\n workflow_dispatch:\n inputs:\n version:\n description: 'Liquibase Version'\n required: true\n\njobs:\n automated-os-extensions-release:\n uses: liquibase/build-logic/.github/workflows/os-extension-automated-release.yml@main\n secrets: inherit\n with:\n version: ${{ inputs.version }}\n\n
dataset_sample\yaml\liquibase_liquibase\.github\workflows\release-extensions.yml
release-extensions.yml
YAML
342
0.85
0
0
react-lib
997
2025-05-27T18:14:35.070521
BSD-3-Clause
false
155807b6eef98a9d707a3294ed76eea6
name: Release Published\non:\n release:\n types: [published]\n workflow_dispatch:\n inputs:\n tag:\n description: "Release Tag"\n required: true\n dry_run:\n description: "Flag to indicate if the workflow is triggered to create a dry-run release"\n required: false\n type: boolean\n default: false\n dry_run_release_id:\n description: "The release id of the dry-run release"\n required: false\n type: string\n dry_run_zip_url:\n description: "URL of the dry-run zip release"\n required: false\n type: string\n dry_run_tar_gz_url:\n description: "URL of the dry-run tar.gz release"\n required: false\n type: string\n dry_run_branch_name:\n description: "The branch name of the dry-run"\n required: false\n type: string\n workflow_call:\n inputs:\n tag:\n description: "Release Tag"\n required: true\n type: string\n dry_run:\n description: "Flag to indicate if the workflow is triggered to create a dry-run release"\n required: true\n type: boolean\n default: false\n dry_run_release_id:\n description: "The release id of the dry-run release"\n required: false\n type: string\n dry_run_zip_url:\n description: "URL of the dry-run zip release"\n required: false\n type: string\n dry_run_tar_gz_url:\n description: "URL of the dry-run tar.gz release"\n required: false\n type: string\n dry_run_branch_name:\n description: "The branch name of the dry-run"\n required: false\n type: string\nenv:\n MAVEN_VERSION: "3.9.2"\njobs:\n setup:\n name: Setup\n runs-on: ubuntu-22.04\n outputs:\n tag: ${{ steps.collect-data.outputs.tag }}\n version: ${{ steps.collect-data.outputs.version }}\n ref_branch: ${{ steps.ref-branch.outputs.branch }}\n dry_run_branch_name: ${{ inputs.dry_run_branch_name }}\n steps:\n - name: Collect Data\n id: collect-data\n uses: actions/github-script@v7\n with:\n script: |\n let tag;\n if (context.payload.inputs) {\n tag = context.payload.inputs.tag;\n } else if ("${{ inputs.tag }}" != "") {\n tag = "${{ inputs.tag }}";\n } else {\n tag = context.payload.release.tag_name;\n }\n let version = tag.replace(/^v/, "");\n\n core.setOutput("tag", tag);\n core.setOutput("version", version);\n\n - run: |\n echo "Publishing version ${{ steps.collect-data.outputs.version }} from ${{ steps.collect-data.outputs.tag }}"\n\n # Use the 'release' branch for hotfix changes;\n # otherwise, use the 'master' branch for regular updates.\n - name: Ref branch to checkout\n id: ref-branch\n run: |\n if [ "${{ github.ref_name }}" == "release" ]; then\n echo "branch=release" >> $GITHUB_OUTPUT\n elif [ "${{ github.ref_name }}" == "master" ]; then\n echo "branch=master" >> $GITHUB_OUTPUT\n else\n echo "branch=${{ inputs.dry_run_branch_name }}" >> $GITHUB_OUTPUT\n fi\n\n manual_trigger_deployment:\n if: ${{ inputs.dry_run == false }}\n name: Manually trigger deployment\n needs: [setup]\n runs-on: ubuntu-22.04\n steps:\n - name: Get Version to deploy\n uses: trstringer/manual-approval@v1\n with:\n secret: ${{ secrets.GITHUB_TOKEN }}\n approvers: suryaaki2,rberezen,jnewton03,kristyldatical,sayaliM0412\n minimum-approvals: 2\n issue-title: "Deploying ${{ needs.setup.outputs.version }} to sonatype"\n issue-body: "Please approve or deny the deployment of version ${{ needs.setup.outputs.version }}"\n additional-approved-words: "lgtm,✅,👍,proceed,shipit,:shipit:"\n additional-denied-words: "stop,error,failed,fail,broken,:x:,👎"\n\n deploy_maven:\n if: ${{ inputs.dry_run == false }}\n name: Deploy to Maven\n needs:\n [\n setup,\n manual_trigger_deployment,\n deploy_javadocs,\n publish_to_github_packages,\n deploy_xsd,\n release-docker,\n ]\n runs-on: ubuntu-22.04\n steps:\n - name: Download release assets\n uses: robinraju/release-downloader@v1.11\n with:\n repository: "liquibase/liquibase"\n tag: "${{ needs.setup.outputs.tag }}"\n fileName: "*"\n out-file-path: "."\n\n - name: Set up Java for publishing to Maven Central Repository\n uses: actions/setup-java@v4\n with:\n java-version: "11"\n distribution: "adopt"\n server-id: sonatype-nexus-staging\n server-username: MAVEN_USERNAME\n server-password: MAVEN_PASSWORD\n gpg-private-key: ${{ secrets.GPG_SECRET }}\n gpg-passphrase: GPG_PASSPHRASE\n env:\n GPG_PASSWORD: ${{ secrets.GPG_PASSPHRASE }}\n\n - name: Publish to Maven Central\n env:\n MAVEN_USERNAME: ${{ secrets.SONATYPE_USERNAME }}\n MAVEN_PASSWORD: ${{ secrets.SONATYPE_TOKEN }}\n GPG_PASSWORD: ${{ secrets.GPG_PASSPHRASE }}\n run: |\n version=${{ needs.setup.outputs.version }}\n\n unzip -j liquibase-additional-*.zip\n\n ##extracts and sign poms\n for i in 'liquibase-core' 'liquibase-maven-plugin' 'liquibase-cli' 'liquibase-cdi' 'liquibase-cdi-jakarta' 'liquibase-commercial'; do\n unzip -j $i-${version}.jar META-INF/maven/org.liquibase/$i/pom.xml\n sed -i -e "s/<version>\(release\|master\)-SNAPSHOT<\/version>/<version>${version}<\/version>/" pom.xml\n mv pom.xml $i-${version}.pom\n if test 'liquibase-commercial' == $i; then\n sed -i -e "s/<\/licenses>/<\/licenses><scm><connection>private<\/connection><developerConnection>private<\/developerConnection><url>private<\/url><\/scm>/" $i-${version}.pom ## scm info not in the pom\n fi\n\n gpg --batch --pinentry-mode=loopback --passphrase "$GPG_PASSWORD" -ab $i-${version}.pom\n\n mvn -B org.apache.maven.plugins:maven-deploy-plugin:3.0.0-M1:deploy-file \\n -Durl=https://oss.sonatype.org/service/local/staging/deploy/maven2/ \\n -DrepositoryId=sonatype-nexus-staging \\n -DpomFile=$i-${version}.pom \\n -DgeneratePom=false \\n -Dfile=$i-${version}.jar \\n -Dsources=$i-${version}-sources.jar \\n -Djavadoc=$i-${version}-javadoc.jar \\n -Dfiles=$i-${version}.jar.asc,$i-${version}-sources.jar.asc,$i-${version}-javadoc.jar.asc,$i-${version}.pom.asc \\n -Dtypes=jar.asc,jar.asc,jar.asc,pom.asc \\n -Dclassifiers=,sources,javadoc,\n done\n\n ## Release repository\n ## Have to find the stagingRepositoryId that was auto-generated\n api_output=$(curl -s -u "${{ secrets.SONATYPE_USERNAME }}:${{ secrets.SONATYPE_TOKEN }}" "https://oss.sonatype.org/service/local/staging/profile_repositories")\n # Extract ids of repositories-item containing the string "liquibase"\n repositories=$(echo "$api_output" | grep -B 8 "liquibase" | grep "<repositoryId>" | awk -F"<|>" '{print $3}')\n echo "Repository IDs containing 'liquibase': $repositories"\n for repo_id in $repositories; do\n echo "Check if $repo_id repository is an extension"\n api_output=$(curl -s -u "${{ secrets.SONATYPE_USERNAME }}:${{ secrets.SONATYPE_TOKEN }}" "https://oss.sonatype.org/service/local/repositories/$repo_id/content/org/liquibase/")\n relative_path=$(echo "$api_output" | grep -oP '<relativePath>\K[^<]+' | awk 'NR==1')\n echo "Relative path: $relative_path"\n if [[ "$relative_path" == *"/org/liquibase/ext/"* ]]; then\n echo "Relative path contains '/org/liquibase/ext/'. It is an extension."\n else\n echo "Relative path does not contain '/org/liquibase/ext/'. It is not an extension."\n stagingRepositoryId=$repo_id\n break\n fi\n done\n\n echo "Staging Repository Id: $stagingRepositoryId"\n\n mvn org.sonatype.plugins:nexus-staging-maven-plugin:1.6.8:rc-close \\n -DnexusUrl=https://oss.sonatype.org/ \\n -DserverId=sonatype-nexus-staging \\n -DstagingRepositoryId=$stagingRepositoryId \\n -DstagingProgressTimeoutMinutes=10\n\n mvn org.sonatype.plugins:nexus-staging-maven-plugin:1.6.8:rc-release \\n -DautoDropAfterRelease=true \\n -DnexusUrl=https://oss.sonatype.org/ \\n -DserverId=sonatype-nexus-staging \\n -DstagingRepositoryId=$stagingRepositoryId \\n -DstagingProgressTimeoutMinutes=10\n\n deploy_javadocs:\n if: ${{ inputs.dry_run == false }}\n name: Upload Javadocs\n needs: [setup, manual_trigger_deployment]\n runs-on: ubuntu-22.04\n # These permissions are needed to interact with GitHub's OIDC Token endpoint.\n permissions:\n id-token: write\n contents: read\n steps:\n - name: Download release javadocs\n uses: robinraju/release-downloader@v1.11\n with:\n repository: "liquibase/liquibase"\n tag: "${{ needs.setup.outputs.tag }}"\n fileName: "liquibase-additional*.zip"\n out-file-path: "."\n\n - name: Unpack javadoc files and upload to s3\n env:\n AWS_ACCESS_KEY_ID: ${{ secrets.JAVADOCS_ACCESS_KEY_ID }}\n AWS_SECRET_ACCESS_KEY: ${{ secrets.JAVADOCS_SECRET_ACCESS_KEY }}\n AWS_DEFAULT_REGION: us-east-1\n # unzip archives, create directory for each jar, and extract jar contents to directory. Then delete the jar and zip files. Then upload to s3.\n run: |\n unzip -j '*.zip' '*javadoc*.jar'\n\n for jar in *liquibase*.jar; do\n dir_name=$(basename "$jar" .jar)\n dir_name=$(echo "$dir_name" | sed -E 's/(-[0-9]+(\.[0-9]+)*(-javadoc)?)//')\n mkdir -p "$dir_name"\n unzip -o "$jar" -d "$dir_name"\n done\n\n rm -rf *.jar *.zip\n aws s3 sync . s3://javadocsliquibasecom-origin --only-show-errors\n\n publish_to_github_packages:\n if: ${{ inputs.dry_run == false }}\n name: Publish artifacts to Github Packages\n runs-on: ubuntu-22.04\n needs: [setup, manual_trigger_deployment]\n permissions:\n contents: read\n packages: write\n\n steps:\n - uses: actions/checkout@v4\n - name: Set up Java for publishing to GitHub Repository\n uses: actions/setup-java@v4\n with:\n java-version: "17"\n distribution: "temurin"\n cache: "maven"\n server-id: liquibase\n - name: Set up Maven\n uses: stCarolas/setup-maven@v5\n with:\n maven-version: ${{ env.MAVEN_VERSION }}\n\n - name: Version artifact\n run: mvn versions:set -DnewVersion="${{ needs.setup.outputs.version }}"\n\n # Publish to GitHub Packages\n - name: Publish package to Github\n run: mvn -B clean deploy -DskipTests=true\n env:\n GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}\n TARGET_ARTIFACT_REPOSITORY: liquibase\n\n deploy_xsd:\n if: ${{ inputs.dry_run == false }}\n name: Upload xsds\n needs: [setup, manual_trigger_deployment]\n runs-on: ubuntu-22.04\n outputs:\n tag: ${{ steps.collect-data.outputs.tag }}\n version: ${{ needs.setup.outputs.version }}\n steps:\n - name: Download liquibase xsd\n uses: actions/checkout@v4\n with:\n # Relative path under $GITHUB_WORKSPACE to place the repository\n path: liquibase-core-repo\n repository: "liquibase/liquibase"\n\n - name: Download liquibase-pro xsd\n uses: actions/checkout@v4\n with:\n ref: "${{ needs.setup.outputs.ref_branch }}"\n token: ${{ secrets.BOT_TOKEN }}\n # Relative path under $GITHUB_WORKSPACE to place the repository\n path: liquibase-pro-repo\n repository: "liquibase/liquibase-pro"\n\n - name: Upload to s3\n env:\n AWS_ACCESS_KEY_ID: ${{ secrets.LIQUIBASEORIGIN_ACCESS_KEY_ID }}\n AWS_SECRET_ACCESS_KEY: ${{ secrets.LIQUIBASEORIGIN_SECRET_ACCESS_KEY }}\n AWS_DEFAULT_REGION: us-east-1\n # aws s3 sync syncs directories and S3 prefixes.\n run: |\n aws s3 sync liquibase-pro-repo/pro/src/main/resources/www.liquibase.org/xml/ns/pro/ s3://liquibaseorg-origin/xml/ns/pro/ --content-type application/octet-stream --only-show-errors\n aws s3 sync liquibase-core-repo/liquibase-standard/src/main/resources/www.liquibase.org/xml/ns/dbchangelog/ s3://liquibaseorg-origin/xml/ns/dbchangelog/ --content-type application/octet-stream --only-show-errors\n aws s3 sync liquibase-pro-repo/pro/src/main/resources/schemas/ s3://liquibaseorg-origin/json/schema/ --content-type application/octet-stream --only-show-errors\n\n - name: Index.htm file upload\n env:\n AWS_ACCESS_KEY_ID: ${{ secrets.LIQUIBASEORIGIN_ACCESS_KEY_ID }}\n AWS_SECRET_ACCESS_KEY: ${{ secrets.LIQUIBASEORIGIN_SECRET_ACCESS_KEY }}\n AWS_DEFAULT_REGION: us-east-1\n # get the major.minor xsd version. grab the index.htm from s3, add the new verison of xsd and sync with the s3 again\n run: |\n version=${{ needs.setup.outputs.version }}\n arr=(${version//./ })\n xsd_version=${arr[0]}"."${arr[1]}\n mkdir index-file\n aws s3 cp s3://liquibaseorg-origin/xml/ns/dbchangelog/index.htm index-file\n if ! grep -q ${xsd_version} index-file/index.htm ; then\n sed -ie "s/<\/ul>/ <li><a href=\"\/xml\/ns\/dbchangelog\/dbchangelog-${xsd_version}.xsd\">dbchangelog-${xsd_version}.xsd<\/a><\/li>\n<\/ul>/" index-file/index.htm\n aws s3 sync index-file s3://liquibaseorg-origin/xml/ns/dbchangelog/ --only-show-errors\n fi\n\n - name: Liquibase xsds SFTP upload\n uses: wangyucode/sftp-upload-action@v2.0.4\n with:\n host: ${{ secrets.WPENGINE_SFTP_HOST }}\n port: ${{ secrets.WPENGINE_SFTP_PORT }}\n username: ${{ secrets.WPENGINE_SFTP_USER }}\n password: ${{ secrets.WPENGINE_SFTP_PASSWORD }}\n compress: true\n forceUpload: true\n localDir: "liquibase-core-repo/liquibase-standard/src/main/resources/www.liquibase.org/xml/ns/dbchangelog/"\n remoteDir: "/xml/ns/dbchangelog/"\n\n - name: Liquibase PRO xsds SFTP upload\n uses: wangyucode/sftp-upload-action@v2.0.4\n with:\n host: ${{ secrets.WPENGINE_SFTP_HOST }}\n port: ${{ secrets.WPENGINE_SFTP_PORT }}\n username: ${{ secrets.WPENGINE_SFTP_USER }}\n password: ${{ secrets.WPENGINE_SFTP_PASSWORD }}\n compress: false\n forceUpload: true\n localDir: "liquibase-pro-repo/pro/src/main/resources/www.liquibase.org/xml/ns/pro/"\n remoteDir: "/xml/ns/pro/"\n\n - name: Liquibase flow-file schema SFTP upload\n uses: wangyucode/sftp-upload-action@v2.0.4\n with:\n host: ${{ secrets.WPENGINE_SFTP_HOST }}\n port: ${{ secrets.WPENGINE_SFTP_PORT }}\n username: ${{ secrets.WPENGINE_SFTP_USER }}\n password: ${{ secrets.WPENGINE_SFTP_PASSWORD }}\n compress: false\n forceUpload: true\n localDir: "liquibase-pro-repo/pro/src/main/resources/schemas/"\n remoteDir: "/json/schema/"\n\n - name: Liquibase index.htm SFTP upload\n uses: wangyucode/sftp-upload-action@v2.0.4\n with:\n host: ${{ secrets.WPENGINE_SFTP_HOST }}\n port: ${{ secrets.WPENGINE_SFTP_PORT }}\n username: ${{ secrets.WPENGINE_SFTP_USER }}\n password: ${{ secrets.WPENGINE_SFTP_PASSWORD }}\n compress: false\n forceUpload: true\n localDir: "index-file/"\n remoteDir: "/xml/ns/dbchangelog/"\n\n release-docker:\n if: always()\n name: Release docker images\n needs: [setup, manual_trigger_deployment]\n runs-on: ubuntu-22.04\n steps:\n - name: Release liquibase/docker v${{ needs.setup.outputs.version }}\n uses: the-actions-org/workflow-dispatch@v4\n id: docker_dispatch\n with:\n workflow: create-release.yml\n token: ${{ secrets.BOT_TOKEN }}\n inputs: '{ "liquibaseVersion": "${{ needs.setup.outputs.version }}", "dryRun": "${{ inputs.dry_run || false }}" }'\n ref: main\n repo: liquibase/docker\n wait-for-completion: true\n workflow-logs: json-output\n\n - name: Adding Docker run to job summary\n if: success()\n continue-on-error: true\n run: echo '### 🐳 Docker Release Job -> ${{steps.docker_dispatch.outputs.workflow-url}}' >> $GITHUB_STEP_SUMMARY\n\n release-minimal-docker:\n if: ${{ inputs.dry_run == false }}\n name: Release Minimal docker image\n needs: [setup, manual_trigger_deployment]\n runs-on: ubuntu-22.04\n steps:\n - name: Release liquibase/liquibase-infrastructure v${{ needs.setup.outputs.version }}\n uses: the-actions-org/workflow-dispatch@v4\n id: docker_dispatch\n with:\n workflow: build-liquibase-minimal.yml\n token: ${{ secrets.BOT_TOKEN }}\n inputs: '{ "liquibaseVersion": "${{ needs.setup.outputs.version }}" }'\n ref: master\n repo: liquibase/liquibase-infrastructure\n wait-for-completion: true\n workflow-logs: json-output\n\n generate-PRO-tag:\n if: ${{ inputs.dry_run == false }}\n name: Generate PRO tags based on OSS release\n needs: [setup]\n runs-on: ubuntu-22.04\n steps:\n - name: Repository Dispatch\n uses: peter-evans/repository-dispatch@v3\n with:\n token: ${{ secrets.LIQUIBOT_PAT_GPM_ACCESS }}\n repository: liquibase/liquibase-pro\n event-type: oss-released-tag\n\n package:\n uses: liquibase/build-logic/.github/workflows/package.yml@main\n needs: [setup]\n secrets: inherit\n with:\n groupId: "org.liquibase"\n artifactId: "liquibase"\n version: ${{ needs.setup.outputs.version }}\n dry_run: ${{ inputs.dry_run || false}}\n dry_run_zip_url: ${{ inputs.dry_run_zip_url || '' }}\n dry_run_tar_gz_url: ${{ inputs.dry_run_tar_gz_url || '' }}\n dry_run_release_id: ${{ inputs.dry_run_release_id || '' }}\n\n update-docs-oss-pro-version:\n if: ${{ inputs.dry_run == false }}\n name: Update OSS and PRO tags based on OSS release\n needs: [setup]\n runs-on: ubuntu-latest\n outputs:\n latest_version: ${{ steps.get_latest_oss_version.outputs.latest_version }}\n previous_version: ${{ steps.get_latest_oss_version.outputs.previous_version }}\n steps:\n - name: Checkout repository\n uses: actions/checkout@v4\n with:\n repository: liquibase/liquibase\n ref: "${{ needs.setup.outputs.ref_branch }}"\n fetch-depth: 0\n\n - name: Get the oss release version\n id: get_latest_oss_version\n run: |\n # Fetch all tags from the remote\n git fetch --tags\n # Get the latest tag\n echo "latest_version=$(git describe --tags $(git rev-list --tags --max-count=1) | sed 's/^v//')" >> $GITHUB_OUTPUT\n # Get the previous released version tag\n echo "previous_version=$(git for-each-ref --sort=creatordate --format '%(refname:short)' refs/tags |sed 's/^v//')" >> $GITHUB_OUTPUT\n\n - name: Repository Dispatch\n uses: peter-evans/repository-dispatch@v3\n with:\n token: ${{ secrets.LIQUIBOT_PAT_GPM_ACCESS }}\n repository: liquibase/liquibase-docs\n event-type: oss-released-version\n client-payload: '{"latest_version": "${{ steps.get_latest_oss_version.outputs.latest_version }}", "previous_version": "${{ steps.get_latest_oss_version.outputs.previous_version }}"}'\n\n # dispatch an event to `liquibase-aws-license-service` repository to update pom.xml with latest OSS Release\n - name: Repository Dispatch\n uses: peter-evans/repository-dispatch@v3\n with:\n token: ${{ secrets.LIQUIBOT_PAT_GPM_ACCESS }}\n repository: liquibase/liquibase-aws-license-service\n event-type: oss-released-version\n\n trigger-sbom-generation-lb-lbpo:\n if: ${{ inputs.dry_run == false }}\n name: Generate SBOMS for liquibase and liquibase-pro\n runs-on: ubuntu-latest\n needs: [setup, update-docs-oss-pro-version]\n strategy:\n matrix:\n repo: ["liquibase/liquibase", "liquibase/liquibase-pro"]\n permissions:\n contents: read\n packages: write\n\n steps:\n - name: Repository Dispatch\n uses: peter-evans/repository-dispatch@v3\n with:\n token: ${{ secrets.LIQUIBOT_PAT_GPM_ACCESS }}\n repository: liquibase/build-logic\n event-type: oss-released-version\n client-payload: '{"latest_version": "${{ needs.update-docs-oss-pro-version.outputs.latest_version }}", "repo_name": "${{ matrix.repo }}", "branch_name": "${{ github.ref }}" }'\n\n dry_run_deploy_maven:\n if: ${{ inputs.dry_run == true }}\n name: Deploy to repo.liquibase.net\n needs: [setup]\n runs-on: ubuntu-22.04\n steps:\n - name: Download dry-run release assets\n uses: robinraju/release-downloader@v1.11\n with:\n repository: "liquibase/liquibase"\n releaseId: "${{ inputs.dry_run_release_id }}"\n fileName: "*"\n out-file-path: "."\n\n - name: Set up Java for publishing to Maven Central Repository\n uses: actions/setup-java@v4\n with:\n java-version: "11"\n distribution: "adopt"\n server-id: dry-run-sonatype-nexus-staging\n server-username: MAVEN_USERNAME\n server-password: MAVEN_PASSWORD\n gpg-private-key: ${{ secrets.GPG_SECRET }}\n gpg-passphrase: GPG_PASSPHRASE\n env:\n GPG_PASSWORD: ${{ secrets.GPG_PASSPHRASE }}\n\n - name: Publish to repo.liquibase.net\n env:\n MAVEN_USERNAME: ${{ secrets.REPO_LIQUIBASE_NET_USER }}\n MAVEN_PASSWORD: ${{ secrets.REPO_LIQUIBASE_NET_PASSWORD }}\n GPG_PASSWORD: ${{ secrets.GPG_PASSPHRASE }}\n run: |\n version=${{ needs.setup.outputs.version }}\n\n unzip -j liquibase-additional-*.zip\n\n ##extracts and sign poms\n for i in 'liquibase-core' 'liquibase-maven-plugin' 'liquibase-cli' 'liquibase-cdi' 'liquibase-cdi-jakarta' 'liquibase-commercial'; do\n unzip -j $i-${version}.jar META-INF/maven/org.liquibase/$i/pom.xml\n sed -i -e "s/<version>\(release\|master\)-SNAPSHOT<\/version>/<version>${version}<\/version>/" pom.xml\n mv pom.xml $i-${version}.pom\n if test 'liquibase-commercial' == $i; then\n sed -i -e "s/<\/licenses>/<\/licenses><scm><connection>private<\/connection><developerConnection>private<\/developerConnection><url>private<\/url><\/scm>/" $i-${version}.pom ## scm info not in the pom\n fi\n\n gpg --batch --pinentry-mode=loopback --passphrase "$GPG_PASSWORD" -ab $i-${version}.pom\n\n mvn -B org.apache.maven.plugins:maven-deploy-plugin:3.0.0-M1:deploy-file \\n -Durl=https://repo.liquibase.net/repository/dry-run-sonatype-nexus-staging/ \\n -DrepositoryId=dry-run-sonatype-nexus-staging \\n -DpomFile=$i-${version}.pom \\n -DgeneratePom=false \\n -Dfile=$i-${version}.jar \\n -Dsources=$i-${version}-sources.jar \\n -Djavadoc=$i-${version}-javadoc.jar \\n -Dfiles=$i-${version}.jar.asc,$i-${version}-sources.jar.asc,$i-${version}-javadoc.jar.asc,$i-${version}.pom.asc \\n -Dtypes=jar.asc,jar.asc,jar.asc,pom.asc \\n -Dclassifiers=,sources,javadoc,\n\n done\n\n publish_assets_to_s3_bucket:\n if: ${{ inputs.dry_run == false }}\n name: Publish OSS released assets to s3 bucket liquibaseorg-origin\n needs: [setup]\n runs-on: ubuntu-22.04\n steps:\n - name: Download released assets\n uses: robinraju/release-downloader@v1.11\n with:\n repository: "liquibase/liquibase"\n latest: true\n fileName: "*"\n out-file-path: "./${{ needs.setup.outputs.version }}-released-assets"\n\n - name: Get current timestamp\n id: timestamp\n run: echo "timestamp=$(date +%Y%m%d-%H%M%S)" >> $GITHUB_ENV\n\n - name: Publish released assets to s3 bucket\n env:\n AWS_ACCESS_KEY_ID: ${{ secrets.LIQUIBASEORIGIN_ACCESS_KEY_ID }}\n AWS_SECRET_ACCESS_KEY: ${{ secrets.LIQUIBASEORIGIN_SECRET_ACCESS_KEY }}\n AWS_DEFAULT_REGION: us-east-1\n run: |\n aws s3 sync "./${{ needs.setup.outputs.version }}-released-assets" s3://liquibaseorg-origin/oss-released-assets/${{ needs.setup.outputs.version }}-released-assets-${{ env.timestamp }}/ --only-show-errors\n
dataset_sample\yaml\liquibase_liquibase\.github\workflows\release-published.yml
release-published.yml
YAML
24,989
0.95
0.060458
0.0322
awesome-app
814
2023-08-07T15:13:58.852840
Apache-2.0
false
5e61c52b250a4bf58e3123326875e580
\nname: Liquibase Url Prerequisites Check\n\n# Job will run nightly at 02:10 EDT / 01:10 CDT\n# Time below is UTC\non:\n schedule:\n - cron: "10 6 * * *"\n workflow_dispatch:\n\njobs:\n urlcheck:\n runs-on: ubuntu-22.04\n steps:\n - uses: actions/checkout@v4\n - name: Check URLs\n uses: urlstechie/urlchecker-action@master\n with:\n # A comma-separated list of file types to cover in the URL checks\n include_files: url-test.md\n # Choose whether to include file with no URLs in the printouts.\n print_all: false\n # Choose whether to print a more verbose end summary with files and broken URLs.\n verbose: true\n # The timeout seconds to provide to requests, defaults to 5 seconds\n timeout: 5\n # How many times to retry a failed request (each is logged, defaults to 1)\n retry_count: 2\n # choose if the force pass or not\n force_pass : false\n\n - name: Slack Notification\n if: ${{ failure() }}\n uses: rtCamp/action-slack-notify@v2\n env:\n SLACK_CHANNEL: team-liquibase-foundation\n SLACK_COLOR: ${{ job.status }} # or a specific color like 'good' or '#ff00ff'\n SLACK_MESSAGE: "${{ github.job }}: ${{ job.status }} @here"\n SLACK_USERNAME: "liquibot"\n SLACK_WEBHOOK: ${{ secrets.SNYK_LIQUIBASE_SLACK_WEBHOOK }}\n MSG_MINIMAL: actions url\n SLACK_ICON_EMOJI: ':liquibase:'\n SLACK_LINK_NAMES: true\n \n
dataset_sample\yaml\liquibase_liquibase\.github\workflows\url-check.yml
url-check.yml
YAML
1,518
0.8
0.045455
0.205128
python-kit
917
2023-11-12T17:56:00.983853
Apache-2.0
false
2e5f373590f5efe09932af2602dc5b34
databaseChangeLog:\n- changeLogId: 5641ba64-35b7-491d-8dc9-d53189f987cf\n- changeSet:\n id: 1604960286751-1\n author: wesley (generated)\n changes:\n - createTable:\n columns:\n - column:\n name: ACCT_NUM\n type: NUMBER(*, 0)\n - column:\n name: AMOUNT\n type: NUMBER(10, 2)\n tableName: ACCOUNT\n tablespace: USERS\n- changeSet:\n id: 1604960286751-2\n author: wesley (generated)\n changes:\n - createProcedure:\n path: objects/storedprocedure/PROC_HELLO_WORLD.sql\n procedureName: PROC_HELLO_WORLD\n relativeToChangelogFile: true\n- changeSet:\n id: 1604960286751-4\n author: wesley (generated)\n changes:\n - createView:\n fullDefinition: true\n selectQuery: CREATE OR REPLACE FORCE VIEW "VIEW1" ("AMOUNT") AS SELECT AMOUNT\n FROM "PROSCHEMA"."ACCOUNT"\n viewName: VIEW1\n- changeSet:\n id: 1604960286751-5\n author: wesley (generated)\n changes:\n - createTable:\n columns:\n - column:\n name: ACCT_NUM\n type: NUMBER(*, 0)\n - column:\n name: AMOUNT\n type: NUMBER(10, 2)\n tableName: ACCOUNT2A\n tablespace: USERS\n- changeSet:\n id: 1604960286751-6\n author: wesley (generated)\n changes:\n - createTable:\n columns:\n - column:\n name: ACCT_NUM\n type: NUMBER(*, 0)\n - column:\n name: AMOUNT\n type: NUMBER(10, 2)\n tableName: ACCOUNT2AA\n tablespace: USERS\n- changeSet:\n id: 1604960286751-7\n author: wesley (generated)\n changes:\n - createTable:\n columns:\n - column:\n name: ACCT_NUM\n type: NUMBER(*, 0)\n - column:\n name: AMOUNT\n type: NUMBER(10, 2)\n tableName: ACCOUNT2AAA\n tablespace: USERS\n- changeSet:\n id: 1604960286751-8\n author: wesley (generated)\n changes:\n - createTable:\n columns:\n - column:\n name: ACCT_NUM\n type: NUMBER(*, 0)\n - column:\n name: AMOUNT\n type: NUMBER(10, 2)\n tableName: ACCOUNT2AAAA\n tablespace: USERS\n- changeSet:\n id: 1604960286751-9\n author: wesley (generated)\n changes:\n - createTable:\n columns:\n - column:\n name: ACCT_NUM\n type: NUMBER(*, 0)\n - column:\n name: AMOUNT\n type: NUMBER(10, 2)\n tableName: ACCOUNT2AAAAA\n tablespace: USERS\n- changeSet:\n id: 1604960286751-10\n author: wesley (generated)\n changes:\n - createTable:\n columns:\n - column:\n constraints:\n nullable: false\n name: ID\n type: NUMBER(20, 0)\n - column:\n name: FIRST\n type: VARCHAR2(64 BYTE)\n - column:\n name: LAST\n type: VARCHAR2(64 BYTE)\n tableName: POPULATED\n tablespace: USERS\n- changeSet:\n id: 1604960286751-11\n author: wesley (generated)\n changes:\n - createTable:\n columns:\n - column:\n name: NULLABLECOL\n type: VARCHAR2(20 BYTE)\n - column:\n name: NOTNULLCOL\n type: VARCHAR2(20 BYTE)\n - column:\n name: PKCOL\n type: VARCHAR2(20 BYTE)\n - column:\n name: UQCOL\n type: NUMBER(*, 0)\n - column:\n name: INCCOL\n type: NUMBER(*, 0)\n - column:\n defaultValue: defaultVal\n name: DEFVALCOL\n type: VARCHAR2(20 BYTE)\n - column:\n defaultValueNumeric: !!float '22'\n name: NUMVALCOL\n type: NUMBER(*, 0)\n - column:\n defaultValueComputed: TO_DATE('2014-08-11', 'YYYY-MM-DD')\n name: DATEVALCOL\n type: date\n - column:\n defaultValueComputed: 1*22\n name: COMPVALCOL\n type: NUMBER\n - column:\n name: IDXUNCLUSTEREDCOL\n type: NUMBER\n - column:\n name: COMMENTCOL\n remarks: This is a remark.\n type: VARCHAR2(20 BYTE)\n - column:\n name: COLUMNTOMARKUNUSED\n remarks: This column is not used by any other refactorings in this changelog.\n type: VARCHAR2(20 BYTE)\n tableName: PRIMARY_TABLE\n tablespace: USERS\n- changeSet:\n id: 1604960286751-12\n author: wesley (generated)\n changes:\n - createTable:\n columns:\n - column:\n name: SUPPLIER_ID\n type: NUMBER(4, 0)\n - column:\n name: SUPPLIER_NAME\n type: VARCHAR2(50 BYTE)\n tableName: SUPPLIERS\n tablespace: USERS\n- changeSet:\n id: 1604960286751-13\n author: wesley (generated)\n changes:\n - createTable:\n columns:\n - column:\n name: SUPPLIER_ID\n type: NUMBER(4, 0)\n - column:\n name: SUPPLIER_NAME\n type: VARCHAR2(50 BYTE)\n tableName: SUPPLIERS1\n tablespace: USERS\n
dataset_sample\yaml\liquibase_liquibase\liquibase-standard\src\test\resources\liquibase\registered-changelog.yml
registered-changelog.yml
YAML
5,108
0.7
0
0
python-kit
373
2024-05-22T00:58:02.521858
GPL-3.0
true
c84031673cfda72658eb84d83ac60c69
databaseChangeLog: \n- preConditions:\n - onFail: WARN \n - sqlCheck:\n expectedResult: 0\n sql: select count(*) from DATABASECHANGELOGLOCK\n- changeSet:\n id: 1 \n author: Liquibase User\n labels: setup\n changes:\n - createTable:\n columns:\n - column:\n name: id\n type: int\n autoIncrement: true\n constraints:\n primaryKey: true\n nullable: false\n - column: \n name: firstname \n type: varchar(50) \n - column: \n name: surname \n type: varchar(50) \n constraints: \n nullable: false \n - column: \n name: state \n type: char(2)\n tableName: example_table\n
dataset_sample\yaml\liquibase_liquibase\liquibase-standard\src\test\resources\liquibase\simple.yml
simple.yml
YAML
795
0.7
0
0
python-kit
992
2024-06-26T15:01:55.055709
BSD-3-Clause
true
d402551e968ca53be14e75ceac11e9dd
databaseChangeLog:\n - changeSet:\n id: cs1\n author: wesley (generated)\n changes:\n - createTable:\n columns:\n - column:\n name: ACCT_NUM\n type: NUMBER(*, 0)\n - column:\n name: AMOUNT\n type: NUMBER(10, 2)\n tableName: ACCOUNT\n tablespace: USERS
dataset_sample\yaml\liquibase_liquibase\liquibase-standard\src\test\resources\liquibase\resource\changelog1.yml
changelog1.yml
YAML
387
0.7
0
0
react-lib
747
2024-04-12T15:42:54.274944
Apache-2.0
true
59da731b68e4d0d37ad80cfc09222153
databaseChangeLog:\n - include:\n file: subresources/child-changelog1.yml\n relativeToChangelogFile: true
dataset_sample\yaml\liquibase_liquibase\liquibase-standard\src\test\resources\liquibase\resource\changelog2.yml
changelog2.yml
YAML
113
0.7
0
0
awesome-app
991
2024-05-23T03:19:46.969036
GPL-3.0
true
260cf71a495f1c8db102be3f724db065
databaseChangeLog:\n - changeSet:\n id: cs2\n author: wesley (generated)\n changes:\n - createTable:\n columns:\n - column:\n name: ACCT_NUM\n type: NUMBER(*, 0)\n - column:\n name: AMOUNT\n type: NUMBER(10, 2)\n tableName: ACCOUNT\n tablespace: USERS
dataset_sample\yaml\liquibase_liquibase\liquibase-standard\src\test\resources\liquibase\resource\subresources\child-changelog.yml
child-changelog.yml
YAML
387
0.7
0
0
python-kit
370
2024-08-10T20:11:35.574185
MIT
true
a719bf8f3cfd3cc73b1534be0e230249
databaseChangeLog:\n - include:\n file: child-changelog2.yml\n relativeToChangelogFile: true
dataset_sample\yaml\liquibase_liquibase\liquibase-standard\src\test\resources\liquibase\resource\subresources\child-changelog1.yml
child-changelog1.yml
YAML
100
0.7
0
0
awesome-app
339
2024-08-28T07:25:09.797247
Apache-2.0
true
d05065aa0a3c3fd89f6c541a18f51b23
databaseChangeLog:\n - changeSet:\n id: cs2\n author: wesley (generated)\n changes:\n - createTable:\n columns:\n - column:\n name: ACCT_NUM\n type: NUMBER(*, 0)\n - column:\n name: AMOUNT\n type: NUMBER(10, 2)\n tableName: ACCOUNT\n tablespace: USERS
dataset_sample\yaml\liquibase_liquibase\liquibase-standard\src\test\resources\liquibase\resource\subresources\child-changelog2.yml
child-changelog2.yml
YAML
387
0.7
0
0
awesome-app
310
2025-02-12T12:19:48.635878
GPL-3.0
true
a719bf8f3cfd3cc73b1534be0e230249
version: '3.8'\n\nservices:\n app:\n build:\n context: .\n dockerfile: Dockerfile\n env_file:\n - .env\n\n volumes:\n - .:/workspaces:cached\n\n # Overrides default command so things don't shut down after the process ends.\n command: sleep infinity\n\n app2:\n build:\n context: .\n dockerfile: devcontainer.Dockerfile2\n env_file:\n - .env\n\n volumes:\n - .:/workspaces:cached\n\n # Overrides default command so things don't shut down after the process ends.\n command: sleep infinity\n\n db:\n image: postgres:latest\n restart: unless-stopped\n volumes:\n - postgres-data:/var/lib/postgresql/data\n env_file:\n - .env\nvolumes:\n postgres-data:\n
dataset_sample\yaml\loft-sh_devpod\examples\compose\docker-compose.devcontainer.yml
docker-compose.devcontainer.yml
YAML
707
0.8
0
0.064516
python-kit
624
2024-11-05T22:34:09.829695
Apache-2.0
false
34c96c958d8000bd0effef5b653edc34
# Setting "sudo" to false forces Travis-CI to use its\n# container-based build infrastructure, which has shorter\n# queue times.\nsudo: false\n\n# Use the newer Travis-CI build templates based on the\n# Debian Linux distribution "Trusty" release.\ndist: trusty\n\n# Select Go as the language used to run the buid.\nlanguage: go\ngo:\n - 1.8.x\n - 1.9.x\n - 1.10.x\ngo_import_path: github.com/akutz/memconn\n\ninstall: true\nscript:\n - make test\n - make benchmark\n
dataset_sample\yaml\loft-sh_devpod\vendor\github.com\akutz\memconn\.travis.yml
.travis.yml
YAML
450
0.95
0
0.333333
awesome-app
459
2024-04-27T16:44:00.339352
Apache-2.0
false
66e582ddc3661ba02d282cbe6a67db64
# run:\n# # timeout for analysis, e.g. 30s, 5m, default is 1m\n# timeout: 5m\n\nlinters:\n disable-all: true\n enable:\n - bodyclose\n - deadcode\n - depguard\n - dogsled\n - goconst\n - gocritic\n - gofmt\n - goimports\n - golint\n - gosec\n - gosimple\n - govet\n - ineffassign\n - interfacer\n - maligned\n - misspell\n - prealloc\n - scopelint\n - staticcheck\n - structcheck\n - stylecheck\n - typecheck\n - unconvert\n - unparam\n - unused\n - misspell\n - wsl\n\nissues:\n exclude-rules:\n - text: "Use of weak random number generator"\n linters:\n - gosec\n - text: "comment on exported var"\n linters:\n - golint\n - text: "don't use an underscore in package name"\n linters:\n - golint\n - text: "ST1003:"\n linters:\n - stylecheck\n # FIXME: Disabled until golangci-lint updates stylecheck with this fix:\n # https://github.com/dominikh/go-tools/issues/389\n - text: "ST1016:"\n linters:\n - stylecheck\n\nlinters-settings:\n dogsled:\n max-blank-identifiers: 3\n maligned:\n # print struct with more effective memory layout or not, false by default\n suggest-new: true\n\nrun:\n tests: false\n
dataset_sample\yaml\loft-sh_devpod\vendor\github.com\alessio\shellescape\.golangci.yml
.golangci.yml
YAML
1,222
0.95
0.015625
0.1
python-kit
158
2023-12-24T03:14:00.413122
MIT
false
a49660cd429903e123c0400504c85d6b
# This is an example goreleaser.yaml file with some sane defaults.\n# Make sure to check the documentation at http://goreleaser.com\nbefore:\n hooks:\n # You may remove this if you don't use go modules.\n - go mod download\n # you may remove this if you don't need go generate\n - go generate ./...\nbuilds:\n - env:\n - CGO_ENABLED=0\n main: ./cmd/escargs\n goos:\n - linux\n - windows\n - darwin\narchives:\n - replacements:\n darwin: Darwin\n linux: Linux\n windows: Windows\n 386: i386\n amd64: x86_64\nchecksum:\n name_template: 'checksums.txt'\nsnapshot:\n name_template: "{{ .Tag }}-next"\nchangelog:\n sort: asc\n filters:\n exclude:\n - '^docs:'\n - '^test:'\n
dataset_sample\yaml\loft-sh_devpod\vendor\github.com\alessio\shellescape\.goreleaser.yml
.goreleaser.yml
YAML
719
0.8
0.060606
0.121212
vue-tools
861
2024-11-13T20:31:47.182131
GPL-3.0
false
5fb2175e99e1aa27b49da356a4cbda30
language: go\ndist: xenial\ngo:\n - '1.10'\n - '1.11'\n - '1.12'\n - '1.13'\n - 'tip'\n\nscript:\n - go test -coverpkg=./... -coverprofile=coverage.info -timeout=5s\n - bash <(curl -s https://codecov.io/bash)\n
dataset_sample\yaml\loft-sh_devpod\vendor\github.com\asaskevich\govalidator\.travis.yml
.travis.yml
YAML
211
0.8
0
0
react-lib
696
2025-06-19T02:44:24.432985
Apache-2.0
false
4fb366cf760148d33c44700fde620a9a
box: golang\nbuild:\n steps:\n - setup-go-workspace\n\n - script:\n name: go get\n code: |\n go version\n go get -t ./...\n\n - script:\n name: go test\n code: |\n go test -race -v ./...\n
dataset_sample\yaml\loft-sh_devpod\vendor\github.com\asaskevich\govalidator\wercker.yml
wercker.yml
YAML
238
0.7
0
0
react-lib
511
2023-08-30T12:08:32.734948
Apache-2.0
false
b0676e8689976b4ad0d5a9347664a743
language: go\n\nos:\n - linux\n - osx\n - windows\n\ngo:\n - go1.13.x\n - go1.x\n\nservices:\n - xvfb\n\nbefore_install:\n - export DISPLAY=:99.0\n\nscript:\n - if [ "$TRAVIS_OS_NAME" = "linux" ]; then sudo apt-get install xsel; fi\n - go test -v .\n - if [ "$TRAVIS_OS_NAME" = "linux" ]; then sudo apt-get install xclip; fi\n - go test -v .\n
dataset_sample\yaml\loft-sh_devpod\vendor\github.com\atotto\clipboard\.travis.yml
.travis.yml
YAML
321
0.7
0.090909
0
awesome-app
562
2024-10-23T10:35:12.949155
BSD-3-Clause
false
a07df666b82ed543e17afd7ed9f8806f
language: go\nsudo: true\ndist: bionic\n\nbranches:\n only:\n - main\n\nos:\n - linux\n - osx\n # Travis doesn't work with windows and Go tip\n #- windows\n\ngo:\n - tip\n\nmatrix:\n allow_failures:\n - go: tip\n\nbefore_install:\n - if [ "$TRAVIS_OS_NAME" = "windows" ]; then choco install make; fi\n - (cd /tmp/; go get golang.org/x/lint/golint)\n\nscript:\n - make go test -v ./...;\n\n
dataset_sample\yaml\loft-sh_devpod\vendor\github.com\aws\smithy-go\.travis.yml
.travis.yml
YAML
377
0.8
0.035714
0.095238
node-utils
836
2024-02-12T11:08:58.600479
MIT
false
b615be337e1992846d32c080fc02c3fe
variables:\n GOPATH: '$(system.defaultWorkingDirectory)/work'\n sdkPath: '$(GOPATH)/src/github.com/$(build.repository.name)'\n\njobs:\n - job: 'goautorest'\n displayName: 'Run go-autorest CI Checks'\n\n strategy:\n matrix:\n Linux_Go113:\n vm.image: 'ubuntu-18.04'\n go.version: '1.13'\n Linux_Go114:\n vm.image: 'ubuntu-18.04'\n go.version: '1.14'\n\n pool:\n vmImage: '$(vm.image)'\n\n steps:\n - task: GoTool@0\n inputs:\n version: '$(go.version)'\n displayName: "Select Go Version"\n\n - script: |\n set -e\n mkdir -p '$(GOPATH)/bin'\n mkdir -p '$(sdkPath)'\n shopt -s extglob\n mv !(work) '$(sdkPath)'\n echo '##vso[task.prependpath]$(GOPATH)/bin'\n displayName: 'Create Go Workspace'\n\n - script: |\n set -e\n curl -sSL https://raw.githubusercontent.com/golang/dep/master/install.sh | sh\n dep ensure -v\n go install ./vendor/golang.org/x/lint/golint\n go get github.com/jstemmer/go-junit-report\n go get github.com/axw/gocov/gocov\n go get github.com/AlekSi/gocov-xml\n go get -u github.com/matm/gocov-html\n workingDirectory: '$(sdkPath)'\n displayName: 'Install Dependencies'\n\n - script: |\n go vet ./autorest/...\n go vet ./logger/...\n go vet ./tracing/...\n workingDirectory: '$(sdkPath)'\n displayName: 'Vet'\n\n - script: |\n go build -v ./autorest/...\n go build -v ./logger/...\n go build -v ./tracing/...\n workingDirectory: '$(sdkPath)'\n displayName: 'Build'\n\n - script: |\n set -e\n go test -race -v -coverprofile=coverage.txt -covermode atomic ./autorest/... ./logger/... ./tracing/... 2>&1 | go-junit-report > report.xml\n gocov convert coverage.txt > coverage.json\n gocov-xml < coverage.json > coverage.xml\n gocov-html < coverage.json > coverage.html\n workingDirectory: '$(sdkPath)'\n displayName: 'Run Tests'\n\n - script: grep -L -r --include *.go --exclude-dir vendor -P "Copyright (\d{4}|\(c\)) Microsoft" ./ | tee >&2\n workingDirectory: '$(sdkPath)'\n displayName: 'Copyright Header Check'\n failOnStderr: true\n condition: succeededOrFailed()\n\n - script: |\n gofmt -s -l -w ./autorest/. >&2\n gofmt -s -l -w ./logger/. >&2\n gofmt -s -l -w ./tracing/. >&2\n workingDirectory: '$(sdkPath)'\n displayName: 'Format Check'\n failOnStderr: true\n condition: succeededOrFailed()\n\n - script: |\n golint ./autorest/... >&2\n golint ./logger/... >&2\n golint ./tracing/... >&2\n workingDirectory: '$(sdkPath)'\n displayName: 'Linter Check'\n failOnStderr: true\n condition: succeededOrFailed()\n\n - task: PublishTestResults@2\n inputs:\n testRunner: JUnit\n testResultsFiles: $(sdkPath)/report.xml\n failTaskOnFailedTests: true\n\n - task: PublishCodeCoverageResults@1\n inputs:\n codeCoverageTool: Cobertura \n summaryFileLocation: $(sdkPath)/coverage.xml\n additionalCodeCoverageFiles: $(sdkPath)/coverage.html\n
dataset_sample\yaml\loft-sh_devpod\vendor\github.com\Azure\go-autorest\azure-pipelines.yml
azure-pipelines.yml
YAML
3,290
0.8
0
0
react-lib
591
2025-06-27T10:28:56.356972
MIT
false
022982803f7dfe5197a49605b582c4de
language: go\n\nsudo: false\n\nbranches:\n except:\n - release\n\nbranches:\n only:\n - master\n - travis\n\ngo:\n - "1.11.x"\n - tip\n\nmatrix:\n allow_failures:\n - go: tip\n\nbefore_install:\n - if [ -n "$GH_USER" ]; then git config --global github.user ${GH_USER}; fi;\n - if [ -n "$GH_TOKEN" ]; then git config --global github.token ${GH_TOKEN}; fi;\n - go get github.com/mattn/goveralls\n\nbefore_script:\n - make deps\n\nscript:\n - make qa\n\nafter_failure:\n - cat ./target/test/report.xml\n\nafter_success:\n - if [ "$TRAVIS_GO_VERSION" = "1.11.1" ]; then $HOME/gopath/bin/goveralls -covermode=count -coverprofile=target/report/coverage.out -service=travis-ci; fi;\n
dataset_sample\yaml\loft-sh_devpod\vendor\github.com\bits-and-blooms\bitset\.travis.yml
.travis.yml
YAML
663
0.7
0.081081
0
node-utils
852
2024-11-12T23:00:55.183132
GPL-3.0
false
9f88f79c62e770a313fa6cb5e1357f80
# Go\n# Build your Go project.\n# Add steps that test, save build artifacts, deploy, and more:\n# https://docs.microsoft.com/azure/devops/pipelines/languages/go\n\ntrigger:\n- master\n\npool:\n vmImage: 'Ubuntu-16.04'\n\nvariables:\n GOBIN: '$(GOPATH)/bin' # Go binaries path\n GOROOT: '/usr/local/go1.11' # Go installation path\n GOPATH: '$(system.defaultWorkingDirectory)/gopath' # Go workspace path\n modulePath: '$(GOPATH)/src/github.com/$(build.repository.name)' # Path to the module's code\n\nsteps:\n- script: |\n mkdir -p '$(GOBIN)'\n mkdir -p '$(GOPATH)/pkg'\n mkdir -p '$(modulePath)'\n shopt -s extglob\n shopt -s dotglob\n mv !(gopath) '$(modulePath)'\n echo '##vso[task.prependpath]$(GOBIN)'\n echo '##vso[task.prependpath]$(GOROOT)/bin'\n displayName: 'Set up the Go workspace'\n\n- script: |\n go version\n go get -v -t -d ./...\n if [ -f Gopkg.toml ]; then\n curl https://raw.githubusercontent.com/golang/dep/master/install.sh | sh\n dep ensure\n fi\n go build -v .\n workingDirectory: '$(modulePath)'\n displayName: 'Get dependencies, then build'\n
dataset_sample\yaml\loft-sh_devpod\vendor\github.com\bits-and-blooms\bitset\azure-pipelines.yml
azure-pipelines.yml
YAML
1,088
0.8
0.025641
0.117647
awesome-app
509
2024-02-24T18:01:22.731242
Apache-2.0
false
910449991284db2e2701c068eac66862
language: go\nmatrix:\n include:\n - go: 1.4.3\n - go: 1.5.4\n - go: 1.6.3\n - go: 1.7\n - go: tip\n allow_failures:\n - go: tip\ninstall:\n- go get golang.org/x/tools/cmd/cover\n- go get github.com/mattn/goveralls\nscript:\n- echo "Test and track coverage" ; $HOME/gopath/bin/goveralls -package "." -service=travis-ci\n -repotoken $COVERALLS_TOKEN\n- echo "Build examples" ; cd examples && go build\n- echo "Check if gofmt'd" ; diff -u <(echo -n) <(gofmt -d -s .)\nenv:\n global:\n secure: HroGEAUQpVq9zX1b1VIkraLiywhGbzvNnTZq2TMxgK7JHP8xqNplAeF1izrR2i4QLL9nsY+9WtYss4QuPvEtZcVHUobw6XnL6radF7jS1LgfYZ9Y7oF+zogZ2I5QUMRLGA7rcxQ05s7mKq3XZQfeqaNts4bms/eZRefWuaFZbkw=\n
dataset_sample\yaml\loft-sh_devpod\vendor\github.com\blang\semver\.travis.yml
.travis.yml
YAML
657
0.7
0.047619
0
node-utils
137
2025-01-08T07:14:14.846131
BSD-3-Clause
false
e65c6b9f1c960962fde1ef3a8ce75a45
coverage:\n status:\n project:\n default:\n threshold: 1%\n patch:\n default:\n target: 70%\nignore:\n - globoptions.go\n
dataset_sample\yaml\loft-sh_devpod\vendor\github.com\bmatcuk\doublestar\v4\.codecov.yml
.codecov.yml
YAML
143
0.7
0
0
node-utils
608
2024-07-02T10:12:19.372464
Apache-2.0
false
eef6bd47bfea08b4d2b429c74f8d8c45
run:\n tests: false\n\nissues:\n include:\n - EXC0001\n - EXC0005\n - EXC0011\n - EXC0012\n - EXC0013\n\n max-issues-per-linter: 0\n max-same-issues: 0\n\nlinters:\n enable:\n # - dupl\n - exhaustive\n # - exhaustivestruct\n - goconst\n - godot\n - godox\n - gomnd\n - gomoddirectives\n - goprintffuncname\n # - lll\n - misspell\n - nakedret\n - nestif\n - noctx\n - nolintlint\n - prealloc\n - wrapcheck\n\n # disable default linters, they are already enabled in .golangci.yml\n disable:\n - deadcode\n - errcheck\n - gosimple\n - govet\n - ineffassign\n - staticcheck\n - structcheck\n - typecheck\n - unused\n - varcheck\n
dataset_sample\yaml\loft-sh_devpod\vendor\github.com\charmbracelet\bubbletea\.golangci-soft.yml
.golangci-soft.yml
YAML
682
0.8
0
0.095238
python-kit
221
2025-06-23T11:48:28.222415
BSD-3-Clause
false
68c0bc78b8976054556bfeb97ca84419
run:\n tests: false\n\nissues:\n include:\n - EXC0001\n - EXC0005\n - EXC0011\n - EXC0012\n - EXC0013\n\n max-issues-per-linter: 0\n max-same-issues: 0\n\nlinters:\n enable:\n - bodyclose\n - exportloopref\n - gofumpt\n - goimports\n - gosec\n - nilerr\n - predeclared\n - revive\n - rowserrcheck\n - sqlclosecheck\n - tparallel\n - unconvert\n - unparam\n - whitespace\n
dataset_sample\yaml\loft-sh_devpod\vendor\github.com\charmbracelet\bubbletea\.golangci.yml
.golangci.yml
YAML
403
0.85
0
0
python-kit
880
2024-08-29T16:48:31.274566
MIT
false
a58c6fc6814d36d7242e0902bf770d46
includes:\n - from_url:\n url: charmbracelet/meta/main/goreleaser-lib.yaml\n\n# yaml-language-server: $schema=https://goreleaser.com/static/schema-pro.json\n\n
dataset_sample\yaml\loft-sh_devpod\vendor\github.com\charmbracelet\bubbletea\.goreleaser.yml
.goreleaser.yml
YAML
159
0.8
0
0.25
python-kit
631
2023-12-07T04:38:55.374194
GPL-3.0
false
7ff9ab22c2a905170aa831725a4f6c0d
run:\n tests: false\n\nissues:\n include:\n - EXC0001\n - EXC0005\n - EXC0011\n - EXC0012\n - EXC0013\n\n max-issues-per-linter: 0\n max-same-issues: 0\n\nlinters:\n enable:\n - exhaustive\n - goconst\n - godot\n - mnd\n - gomoddirectives\n - goprintffuncname\n - misspell\n - nakedret\n - nestif\n - noctx\n - nolintlint\n - prealloc\n - wrapcheck\n\n # disable default linters, they are already enabled in .golangci.yml\n disable:\n - errcheck\n - gosimple\n - govet\n - ineffassign\n - staticcheck\n - typecheck\n - unused\n
dataset_sample\yaml\loft-sh_devpod\vendor\github.com\charmbracelet\huh\.golangci-soft.yml
.golangci-soft.yml
YAML
570
0.8
0
0.028571
react-lib
8
2025-03-17T16:03:19.587097
MIT
false
180a0a617cabbcd0b44122b8c3608725
run:\n tests: false\n\nissues:\n include:\n - EXC0001\n - EXC0005\n - EXC0011\n - EXC0012\n - EXC0013\n\n max-issues-per-linter: 0\n max-same-issues: 0\n\nlinters:\n enable:\n - bodyclose\n - goimports\n - gosec\n - nilerr\n - predeclared\n - revive\n - rowserrcheck\n - sqlclosecheck\n - tparallel\n - unconvert\n - unparam\n - whitespace\n
dataset_sample\yaml\loft-sh_devpod\vendor\github.com\charmbracelet\huh\.golangci.yml
.golangci.yml
YAML
369
0.85
0
0
node-utils
464
2025-05-29T16:48:56.534584
GPL-3.0
false
8ce38dd5c31bc5a2ccb601d6bbeeb320
run:\n tests: false\n\nissues:\n include:\n - EXC0001\n - EXC0005\n - EXC0011\n - EXC0012\n - EXC0013\n\n max-issues-per-linter: 0\n max-same-issues: 0\n\nlinters:\n enable:\n # - dupl\n - exhaustive\n # - exhaustivestruct\n - goconst\n - godot\n - godox\n - gomnd\n - gomoddirectives\n - goprintffuncname\n # - lll\n - misspell\n - nakedret\n - nestif\n - noctx\n - nolintlint\n - prealloc\n - wrapcheck\n\n # disable default linters, they are already enabled in .golangci.yml\n disable:\n - deadcode\n - errcheck\n - gosimple\n - govet\n - ineffassign\n - staticcheck\n - structcheck\n - typecheck\n - unused\n - varcheck\n
dataset_sample\yaml\loft-sh_devpod\vendor\github.com\charmbracelet\lipgloss\.golangci-soft.yml
.golangci-soft.yml
YAML
682
0.8
0
0.095238
awesome-app
143
2024-03-18T10:20:01.709402
MIT
false
68c0bc78b8976054556bfeb97ca84419
run:\n tests: false\n\nissues:\n include:\n - EXC0001\n - EXC0005\n - EXC0011\n - EXC0012\n - EXC0013\n\n max-issues-per-linter: 0\n max-same-issues: 0\n\nlinters:\n enable:\n - bodyclose\n - exportloopref\n - gofumpt\n - goimports\n - gosec\n - nilerr\n - predeclared\n - revive\n - rowserrcheck\n - sqlclosecheck\n - tparallel\n - unconvert\n - unparam\n - whitespace\n
dataset_sample\yaml\loft-sh_devpod\vendor\github.com\charmbracelet\lipgloss\.golangci.yml
.golangci.yml
YAML
403
0.85
0
0
awesome-app
352
2024-05-05T03:37:42.320353
Apache-2.0
false
a58c6fc6814d36d7242e0902bf770d46
includes:\n - from_url:\n url: charmbracelet/meta/main/goreleaser-lib.yaml\n# yaml-language-server: $schema=https://goreleaser.com/static/schema-pro.json\n\n
dataset_sample\yaml\loft-sh_devpod\vendor\github.com\charmbracelet\lipgloss\.goreleaser.yml
.goreleaser.yml
YAML
158
0.8
0
0.25
vue-tools
748
2024-06-08T22:14:12.484087
MIT
false
1de1344d0961a2f92aeafcb3f4c0f8e7
linters:\n enable:\n - gofmt\n - goimports\n - ineffassign\n - misspell\n - revive\n - staticcheck\n - structcheck\n - unconvert\n - unused\n - varcheck\n - vet\n disable:\n - errcheck\n\nrun:\n timeout: 3m\n skip-dirs:\n - vendor\n
dataset_sample\yaml\loft-sh_devpod\vendor\github.com\containerd\console\.golangci.yml
.golangci.yml
YAML
255
0.85
0
0
node-utils
402
2025-03-06T16:33:32.362721
MIT
false
a11b4c6573b3396d5a5cf92605c2721a
linters:\n enable:\n - exportloopref # Checks for pointers to enclosing loop variables\n - gofmt\n - goimports\n - gosec\n - ineffassign\n - misspell\n - nolintlint\n - revive\n - staticcheck\n - tenv # Detects using os.Setenv instead of t.Setenv since Go 1.17\n - unconvert\n - unused\n - vet\n - dupword # Checks for duplicate words in the source code\n disable:\n - errcheck\n\nrun:\n timeout: 5m\n skip-dirs:\n - api\n - cluster\n - design\n - docs\n - docs/man\n - releases\n - reports\n - test # e2e scripts\n
dataset_sample\yaml\loft-sh_devpod\vendor\github.com\containerd\log\.golangci.yml
.golangci.yml
YAML
558
0.95
0.066667
0
react-lib
265
2025-05-01T22:28:33.641433
Apache-2.0
false
b8375607b0dbb7da9b703c9efacd207f
linters:\n enable:\n - copyloopvar\n - gofmt\n - goimports\n - gosec\n - ineffassign\n - misspell\n - nolintlint\n - revive\n - staticcheck\n - tenv # Detects using os.Setenv instead of t.Setenv since Go 1.17\n - unconvert\n - unused\n - govet\n - dupword # Checks for duplicate words in the source code\n disable:\n - errcheck\n\nrun:\n timeout: 5m\n\nissues:\n exclude-dirs:\n - api\n - cluster\n - design\n - docs\n - docs/man\n - releases\n - reports\n - test # e2e scripts\n
dataset_sample\yaml\loft-sh_devpod\vendor\github.com\containerd\platforms\.golangci.yml
.golangci.yml
YAML
520
0.95
0.03125
0
awesome-app
889
2024-10-07T20:52:27.026276
BSD-3-Clause
false
fae484d7bf1b9704341b44d339aaba46
linters:\n enable:\n - staticcheck\n - unconvert\n - gofmt\n - goimports\n - revive\n - ineffassign\n - vet\n - unused\n - misspell\n disable:\n - errcheck\n\nlinters-settings:\n revive:\n ignore-generated-headers: true\n rules:\n - name: blank-imports\n - name: context-as-argument\n - name: context-keys-type\n - name: dot-imports\n - name: error-return\n - name: error-strings\n - name: error-naming\n - name: exported\n - name: if-return\n - name: increment-decrement\n - name: var-naming\n arguments: [["UID", "GID"], []]\n - name: var-declaration\n - name: package-comments\n - name: range\n - name: receiver-naming\n - name: time-naming\n - name: unexported-return\n - name: indent-error-flow\n - name: errorf\n - name: empty-block\n - name: superfluous-else\n - name: unused-parameter\n - name: unreachable-code\n - name: redefines-builtin-id\n\nissues:\n include:\n - EXC0002\n\nrun:\n timeout: 8m\n skip-dirs:\n - example\n
dataset_sample\yaml\loft-sh_devpod\vendor\github.com\containerd\ttrpc\.golangci.yml
.golangci.yml
YAML
1,056
0.85
0.019231
0
react-lib
502
2024-08-18T08:53:25.312943
BSD-3-Clause
false
3fa4f43aeebbf8fa3bef90c586b55323
---\n# Reference: https://golangci-lint.run/usage/configuration/\nrun:\n timeout: 5m\n # modules-download-mode: vendor\n\n # Include test files.\n tests: true\n\n skip-dirs: []\n\n skip-files: []\n\noutput:\n # colored-line-number|line-number|json|tab|checkstyle|code-climate, default is "colored-line-number".\n format: colored-line-number\n print-issued-lines: true\n print-linter-name: true\n\n# Linter specific settings. See below in the `linter.enable` section for details on what each linter is doing.\nlinters-settings:\n dogsled:\n # Checks assignments with too many blank identifiers. Default is 2.\n max-blank-identifiers: 2\n\n dupl:\n # Tokens count to trigger issue.\n threshold: 150\n\n errcheck:\n # Report about not checking of errors in type assertions: `a := b.(MyStruct)`.\n # Enabled as this is often overlooked by developers.\n check-type-assertions: true\n # Report about assignment of errors to blank identifier: `num, _ := strconv.Atoi(numStr)`.\n # Disabled as we consider that if the developer did type `_`, it was on purpose.\n # Note that while this isn't enforced by the linter, each and every case of ignored error should\n # be accompanied with a comment explaining why that error is being discarded.\n check-blank: false\n\n exhaustive:\n # Indicates that switch statements are to be considered exhaustive if a\n # 'default' case is present, even if all enum members aren't listed in the\n # switch.\n default-signifies-exhaustive: false\n\n funlen:\n # funlen checks the number of lines/statements in a function.\n # While is is always best to keep functions short for readability, maintainability and testing,\n # the default are a bit too strict (60 lines / 40 statements), increase it to be more flexible.\n lines: 160\n statements: 70\n\n # NOTE: We don't set `gci` for import order as it supports only one prefix. Use `goimports.local-prefixes` instead.\n\n gocognit:\n # Minimal code complexity to report, defaults to 30 in gocognit, defaults 10 in golangci.\n # Use 15 as it allows for some flexibility while preventing too much complexity.\n # NOTE: Similar to gocyclo.\n min-complexity: 35\n\n nestif:\n # Minimal complexity of if statements to report.\n min-complexity: 8\n\n goconst:\n # Minimal length of string constant.\n min-len: 4\n # Minimal occurrences count to trigger.\n # Increase the default from 3 to 5 as small number of const usage can reduce readability instead of improving it.\n min-occurrences: 5\n\n gocritic:\n # Which checks should be disabled; can't be combined with 'enabled-checks'.\n # See https://go-critic.github.io/overview#checks-overview\n # To check which checks are enabled run `GL_DEBUG=gocritic golangci-lint run`\n disabled-checks:\n - hugeParam # Very strict check on the size of variables being copied. Too strict for most developer.\n # Enable multiple checks by tags, run `GL_DEBUG=gocritic golangci-lint run` to see all tags and checks.\n # Empty list by default. See https://github.com/go-critic/go-critic#usage -> section "Tags".\n enabled-tags:\n - diagnostic\n - style\n - opinionated\n - performance\n settings:\n rangeValCopy:\n sizeThreshold: 1024 # Increase the allowed copied bytes in range.\n\n cyclop:\n max-complexity: 35\n\n gocyclo:\n # Similar check as gocognit.\n # NOTE: We might be able to remove this linter as it is redundant with gocyclo. It is in golangci-lint, so we keep it for now.\n min-complexity: 35\n\n godot:\n # Check all top-level comments, not only declarations.\n check-all: true\n\n gofmt:\n # simplify code: gofmt with `-s` option.\n simplify: true\n\n # NOTE: the goheader settings are set per-project.\n\n goimports:\n # Put imports beginning with prefix after 3rd-party packages.\n # It's a comma-separated list of prefixes.\n local-prefixes: "github.com/creack/pty"\n\n golint:\n # Minimal confidence for issues, default is 0.8.\n min-confidence: 0.8\n\n gosimple:\n # Select the Go version to target. The default is '1.13'.\n go: "1.18"\n # https://staticcheck.io/docs/options#checks\n checks: ["all"]\n\n gosec:\n\n govet:\n # Enable all available checks from go vet.\n enable-all: false\n # Report about shadowed variables.\n check-shadowing: true\n\n # NOTE: depguard is disabled as it is very slow and made redundant by gomodguard.\n\n lll:\n # Make sure everyone is on the same level, fix the tab width to go's default.\n tab-width: 8\n # Increase the default max line length to give more flexibility. Forcing newlines can reduce readability instead of improving it.\n line-length: 180\n\n misspell:\n locale: US\n ignore-words:\n\n nakedret:\n # Make an issue if func has more lines of code than this setting and it has naked returns; default is 30.\n # NOTE: Consider setting this to 1 to prevent naked returns.\n max-func-lines: 30\n\n nolintlint:\n # Prevent ununsed directive to avoid stale comments.\n allow-unused: false\n # Require an explanation of nonzero length after each nolint directive.\n require-explanation: true\n # Exclude following linters from requiring an explanation.\n # NOTE: It is strongly discouraged to put anything in there.\n allow-no-explanation: []\n # Enable to require nolint directives to mention the specific linter being suppressed. This ensurce the developer understand the reason being the error.\n require-specific: true\n\n prealloc:\n # NOTE: For most programs usage of prealloc will be a premature optimization.\n # Keep thing simple, pre-alloc what is obvious and profile the program for more complex scenarios.\n #\n simple: true # Checkonly on simple loops that have no returns/breaks/continues/gotos in them.\n range-loops: true # Check range loops, true by default\n for-loops: false # Check suggestions on for loops, false by default\n\n rowserrcheck:\n packages: []\n\n staticcheck:\n # Select the Go version to target. The default is '1.13'.\n go: "1.18"\n # https://staticcheck.io/docs/options#checks\n checks: ["all"]\n\n stylecheck:\n # Select the Go version to target. The default is '1.13'.\n go: "1.18"\n # https://staticcheck.io/docs/options#checks\n checks: ["all"] # "-ST1000", "-ST1003", "-ST1016", "-ST1020", "-ST1021", "-ST1022"]\n\n tagliatelle:\n # Check the struck tag name case.\n case:\n # Use the struct field name to check the name of the struct tag.\n use-field-name: false\n rules:\n # Any struct tag type can be used.\n # support string case: `camel`, `pascal`, `kebab`, `snake`, `goCamel`, `goPascal`, `goKebab`, `goSnake`, `upper`, `lower`\n json: snake\n firestore: camel\n yaml: camel\n xml: camel\n bson: camel\n avro: snake\n mapstructure: kebab\n envconfig: upper\n\n unparam:\n # Don't create an error if an exported code have static params being used. It is often expected in libraries.\n # NOTE: It would be nice if this linter would differentiate between a main package and a lib.\n check-exported: true\n\n unused: {}\n\n whitespace:\n multi-if: false # Enforces newlines (or comments) after every multi-line if statement\n multi-func: false # Enforces newlines (or comments) after every multi-line function signature\n\n# Run `golangci-lint help linters` to get the full list of linter with their description.\nlinters:\n disable-all: true\n # NOTE: enable-all is deprecated because too many people don't pin versions...\n # We still require explicit documentation on why some linters are disabled.\n # disable:\n # - depguard # Go linter that checks if package imports are in a list of acceptable packages [fast: true, auto-fix: false]\n # - exhaustivestruct # Checks if all struct's fields are initialized [fast: true, auto-fix: false]\n # - forbidigo # Forbids identifiers [fast: true, auto-fix: false]\n # - gci # Gci control golang package import order and make it always deterministic. [fast: true, auto-fix: true]\n # - godox # Tool for detection of FIXME, TODO and other comment keywords [fast: true, auto-fix: false]\n # - goerr113 # Golang linter to check the errors handling expressions [fast: true, auto-fix: false]\n # - golint # Golint differs from gofmt. Gofmt reformats Go source code, whereas golint prints out style mistakes [fast: false, auto-fix: false]\n # - gomnd # An analyzer to detect magic numbers. [fast: true, auto-fix: false]\n # - gomoddirectives # Manage the use of 'replace', 'retract', and 'excludes' directives in go.mod. [fast: true, auto-fix: false]\n # - interfacer # Linter that suggests narrower interface types [fast: false, auto-fix: false]\n # - maligned # Tool to detect Go structs that would take less memory if their fields were sorted [fast: false, auto-fix: false]\n # - nlreturn # nlreturn checks for a new line before return and branch statements to increase code clarity [fast: true, auto-fix: false]\n # - scopelint # Scopelint checks for unpinned variables in go programs [fast: true, auto-fix: false]\n # - wrapcheck # Checks that errors returned from external packages are wrapped [fast: false, auto-fix: false]\n # - wsl # Whitespace Linter - Forces you to use empty lines! [fast: true, auto-fix: false]\n\n # disable-reasons:\n # - depguard # Checks whitelisted/blacklisted import path, but runs way too slow. Not that useful.\n # - exhaustivestruct # Good concept, but not mature enough (errors on not assignable fields like locks) and too noisy when using AWS SDK as most fields are unused.\n # - forbidigo # Great idea, but too strict out of the box. Probably will re-enable soon.\n # - gci # Conflicts with goimports/gofumpt.\n # - godox # Don't fail when finding TODO, FIXME, etc.\n # - goerr113 # Too many false positives.\n # - golint # Deprecated (since v1.41.0) due to: The repository of the linter has been archived by the owner. Replaced by revive.\n # - gomnd # Checks for magic numbers. Disabled due to too many false positives not configurable (03/01/2020 v1.23.7).\n # - gomoddirectives # Doesn't support //nolint to whitelist.\n # - interfacer # Deprecated (since v1.38.0) due to: The repository of the linter has been archived by the owner.\n # - maligned # Deprecated (since v1.38.0) due to: The repository of the linter has been archived by the owner. Replaced by govet 'fieldalignment'.\n # - nlreturn # Actually reduces readability in most cases.\n # - scopelint # Deprecated (since v1.39.0) due to: The repository of the linter has been deprecated by the owner. Replaced by exportloopref.\n # - wrapcheck # Good concept, but always warns for http coded errors. Need to re-enable and whitelist our error package.\n # - wsl # Forces to add newlines around blocks. Lots of false positives, not that useful.\n\n enable:\n - asciicheck # Simple linter to check that your code does not contain non-ASCII identifiers [fast: true, auto-fix: false]\n - bodyclose # checks whether HTTP response body is closed successfully [fast: false, auto-fix: false]\n - cyclop # checks function and package cyclomatic complexity [fast: false, auto-fix: false]\n - dogsled # Checks assignments with too many blank identifiers (e.g. x, _, _, _, := f()) [fast: true, auto-fix: false]\n - dupl # Tool for code clone detection [fast: true, auto-fix: false]\n - durationcheck # check for two durations multiplied together [fast: false, auto-fix: false]\n - errcheck # Errcheck is a program for checking for unchecked errors in go programs. These unchecked errors can be critical bugs in some cases [fast: false, auto-fix: false]\n - errname # Checks that sentinel errors are prefixed with the `Err` and error types are suffixed with the `Error`. [fast: false, auto-fix: false]\n - errorlint # go-errorlint is a source code linter for Go software that can be used to find code that will cause problems with the error wrapping scheme introduced in Go 1.13. [fast: false, auto-fix: false]\n - exhaustive # check exhaustiveness of enum switch statements [fast: false, auto-fix: false]\n - exportloopref # checks for pointers to enclosing loop variables [fast: false, auto-fix: false]\n - forcetypeassert # finds forced type assertions [fast: true, auto-fix: false]\n - funlen # Tool for detection of long functions [fast: true, auto-fix: false]\n - gochecknoglobals # check that no global variables exist [fast: true, auto-fix: false]\n - gochecknoinits # Checks that no init functions are present in Go code [fast: true, auto-fix: false]\n - gocognit # Computes and checks the cognitive complexity of functions [fast: true, auto-fix: false]\n - goconst # Finds repeated strings that could be replaced by a constant [fast: true, auto-fix: false]\n - gocritic # Provides many diagnostics that check for bugs, performance and style issues. [fast: false, auto-fix: false]\n - gocyclo # Computes and checks the cyclomatic complexity of functions [fast: true, auto-fix: false]\n - godot # Check if comments end in a period [fast: true, auto-fix: true]\n - gofmt # Gofmt checks whether code was gofmt-ed. By default this tool runs with -s option to check for code simplification [fast: true, auto-fix: true]\n - gofumpt # Gofumpt checks whether code was gofumpt-ed. [fast: true, auto-fix: true]\n - goheader # Checks is file header matches to pattern [fast: true, auto-fix: false]\n - goimports # Goimports does everything that gofmt does. Additionally it checks unused imports [fast: true, auto-fix: true]\n - gomodguard # Allow and block list linter for direct Go module dependencies. This is different from depguard where there are different block types for example version constraints and module recommendations. [fast: true, auto-fix: false]\n - goprintffuncname # Checks that printf-like functions are named with `f` at the end [fast: true, auto-fix: false]\n - gosec # (gas): Inspects source code for security problems [fast: false, auto-fix: false]\n - gosimple # (megacheck): Linter for Go source code that specializes in simplifying a code [fast: false, auto-fix: false]\n - govet # (vet, vetshadow): Vet examines Go source code and reports suspicious constructs, such as Printf calls whose arguments do not align with the format string [fast: false, auto-fix: false]\n - importas # Enforces consistent import aliases [fast: false, auto-fix: false]\n - ineffassign # Detects when assignments to existing variables are not used [fast: true, auto-fix: false]\n - lll # Reports long lines [fast: true, auto-fix: false]\n - makezero # Finds slice declarations with non-zero initial length [fast: false, auto-fix: false]\n - misspell # Finds commonly misspelled English words in comments [fast: true, auto-fix: true]\n - nakedret # Finds naked returns in functions greater than a specified function length [fast: true, auto-fix: false]\n - nestif # Reports deeply nested if statements [fast: true, auto-fix: false]\n - nilerr # Finds the code that returns nil even if it checks that the error is not nil. [fast: false, auto-fix: false]\n - noctx # noctx finds sending http request without context.Context [fast: false, auto-fix: false]\n - nolintlint # Reports ill-formed or insufficient nolint directives [fast: true, auto-fix: false]\n - paralleltest # paralleltest detects missing usage of t.Parallel() method in your Go test [fast: true, auto-fix: false]\n - prealloc # Finds slice declarations that could potentially be preallocated [fast: true, auto-fix: false]\n - predeclared # find code that shadows one of Go's predeclared identifiers [fast: true, auto-fix: false]\n - promlinter # Check Prometheus metrics naming via promlint [fast: true, auto-fix: false]\n - revive # Fast, configurable, extensible, flexible, and beautiful linter for Go. Drop-in replacement of golint. [fast: false, auto-fix: false]\n # Disabled due to generic. Work in progress upstream.\n # - rowserrcheck # checks whether Err of rows is checked successfully [fast: false, auto-fix: false]\n # Disabled due to generic. Work in progress upstream.\n # - sqlclosecheck # Checks that sql.Rows and sql.Stmt are closed. [fast: false, auto-fix: false]\n - staticcheck # (megacheck): Staticcheck is a go vet on steroids, applying a ton of static analysis checks [fast: false, auto-fix: false]\n - stylecheck # Stylecheck is a replacement for golint [fast: false, auto-fix: false]\n # Disabled due to generic. Work in progress upstream.\n # - tagliatelle # Checks the struct tags. [fast: true, auto-fix: false]\n # - testpackage # linter that makes you use a separate _test package [fast: true, auto-fix: false]\n - thelper # thelper detects golang test helpers without t.Helper() call and checks the consistency of test helpers [fast: false, auto-fix: false]\n - tparallel # tparallel detects inappropriate usage of t.Parallel() method in your Go test codes [fast: false, auto-fix: false]\n - typecheck # Like the front-end of a Go compiler, parses and type-checks Go code [fast: false, auto-fix: false]\n - unconvert # Remove unnecessary type conversions [fast: false, auto-fix: false]\n - unparam # Reports unused function parameters [fast: false, auto-fix: false]\n # Disabled due to way too many false positive in go1.20.\n # - unused # (megacheck): Checks Go code for unused constants, variables, functions and types [fast: false, auto-fix: false]\n # Disabled due to generic. Work in progress upstream.\n # - wastedassign # wastedassign finds wasted assignment statements. [fast: false, auto-fix: false]\n - whitespace # Tool for detection of leading and trailing whitespace [fast: true, auto-fix: true]\n\nissues:\n exclude:\n # Allow shadowing of 'err'.\n - 'shadow: declaration of "err" shadows declaration'\n # Allow shadowing of `ctx`.\n - 'shadow: declaration of "ctx" shadows declaration'\n # Maximum issues count per one linter. Set to 0 to disable. Default is 50.\n max-per-linter: 10\n # Disable default excludes. Always be explicit on what we exclude.\n exclude-use-default: false\n # Exclude some linters from running on tests files.\n exclude-rules: []\n
dataset_sample\yaml\loft-sh_devpod\vendor\github.com\creack\pty\.golangci.yml
.golangci.yml
YAML
18,922
0.95
0.175926
0.416961
react-lib
53
2025-06-01T21:48:32.352682
Apache-2.0
false
ef1ff7a7d39446d2637cdcb8e4c939c9
language: go\nsudo: false\n\ngo:\n - 1.10.x\n - 1.11.x\n - 1.12.x \n - 1.13.x\n - 1.14.x\n - 1.15.x\n\ncache:\n directories:\n - $HOME/.cache/go-build\n - $HOME/gopath/pkg/mod\n\nenv:\n global:\n - GO111MODULE=on\n\nbefore_install:\n - go get github.com/mattn/goveralls\n - go get golang.org/x/tools/cmd/cover\n - go get golang.org/x/tools/cmd/goimports\n - go get golang.org/x/lint/golint\nscript:\n - gofiles=$(find ./ -name '*.go') && [ -z "$gofiles" ] || unformatted=$(goimports -l $gofiles) && [ -z "$unformatted" ] || (echo >&2 "Go files must be formatted with gofmt. Following files has problem:\n $unformatted" && false)\n - golint ./... # This won't break the build, just show warnings\n - $HOME/gopath/bin/goveralls -service=travis-ci\n
dataset_sample\yaml\loft-sh_devpod\vendor\github.com\dimchansky\utfbom\.travis.yml
.travis.yml
YAML
742
0.95
0
0
vue-tools
433
2025-03-22T18:53:12.108575
GPL-3.0
false
3b8947c1a963469c966969cc983aa615
linters:\n enable:\n - bodyclose\n - dupword # Checks for duplicate words in the source code\n - gofmt\n - goimports\n - ineffassign\n - misspell\n - revive\n - staticcheck\n - unconvert\n - unused\n - vet\n disable:\n - errcheck\n\nrun:\n deadline: 2m\n
dataset_sample\yaml\loft-sh_devpod\vendor\github.com\distribution\reference\.golangci.yml
.golangci.yml
YAML
275
0.95
0.055556
0
vue-tools
22
2024-08-15T15:35:57.085803
BSD-3-Clause
false
bf3404c60cd0bd42f202dc7ddfd34a58
dependencies:\n post:\n # install golint\n - go get golang.org/x/lint/golint\n\ntest:\n pre:\n # run analysis before tests\n - go vet ./...\n - test -z "$(golint ./... | tee /dev/stderr)"\n - test -z "$(gofmt -s -l . | tee /dev/stderr)"\n
dataset_sample\yaml\loft-sh_devpod\vendor\github.com\docker\go-units\circle.yml
circle.yml
YAML
247
0.8
0
0.2
awesome-app
331
2024-08-27T18:11:44.159031
Apache-2.0
false
c943afaea172b047085823d6e7e87fd8
sudo: false\nlanguage: go\ngo_import_path: github.com/dustin/go-humanize\ngo:\n - 1.13.x\n - 1.14.x\n - 1.15.x\n - 1.16.x\n - stable\n - master\nmatrix:\n allow_failures:\n - go: master\n fast_finish: true\ninstall:\n - # Do nothing. This is needed to prevent default install action "go get -t -v ./..." from happening here (we want it to happen inside script step).\nscript:\n - diff -u <(echo -n) <(gofmt -d -s .)\n - go vet .\n - go install -v -race ./...\n - go test -v -race ./...\n
dataset_sample\yaml\loft-sh_devpod\vendor\github.com\dustin\go-humanize\.travis.yml
.travis.yml
YAML
482
0.95
0
0
python-kit
297
2025-01-26T10:00:39.469436
Apache-2.0
false
41f052dec02291a32a59e5bc9ae5cd6b
language: go\n\ngo:\n - 1.x\n\nbefore_install:\n - go test -v\n\nscript:\n - go test -race -coverprofile=coverage.txt -covermode=atomic\n\nafter_success:\n - bash <(curl -s https://codecov.io/bash)
dataset_sample\yaml\loft-sh_devpod\vendor\github.com\emicklei\go-restful\v3\.travis.yml
.travis.yml
YAML
189
0.8
0
0
vue-tools
344
2025-04-20T08:24:24.466080
Apache-2.0
false
20ab9ad977d4a5252204f66008d029a1
linters:\n enable-all: true\n disable:\n - golint\n - interfacer\n - scopelint\n - maligned\n - rowserrcheck\n - funlen\n - depguard\n - goerr113\n - exhaustivestruct\n - testpackage\n - gochecknoglobals\n - wrapcheck\n - forbidigo\n - ifshort\n - cyclop\n - gomoddirectives\nlinters-settings:\n exhaustive:\n default-signifies-exhaustive: true\nissues:\n exclude-use-default: false\n
dataset_sample\yaml\loft-sh_devpod\vendor\github.com\erikgeiser\coninput\.golangci.yml
.golangci.yml
YAML
415
0.7
0
0
react-lib
303
2024-05-19T00:57:34.908039
MIT
false
d67d01eae48ae3fa3717bf3902f0cd9b
freebsd_task:\n name: 'FreeBSD'\n freebsd_instance:\n image_family: freebsd-13-2\n install_script:\n - pkg update -f\n - pkg install -y go\n test_script:\n # run tests as user "cirrus" instead of root\n - pw useradd cirrus -m\n - chown -R cirrus:cirrus .\n - FSNOTIFY_BUFFER=4096 sudo --preserve-env=FSNOTIFY_BUFFER -u cirrus go test -parallel 1 -race ./...\n - sudo --preserve-env=FSNOTIFY_BUFFER -u cirrus go test -parallel 1 -race ./...\n
dataset_sample\yaml\loft-sh_devpod\vendor\github.com\fsnotify\fsnotify\.cirrus.yml
.cirrus.yml
YAML
479
0.8
0
0.076923
react-lib
172
2023-08-22T18:46:42.143738
BSD-3-Clause
false
dc3669dfa73049b7764268060a1b5b75