content stringlengths 1 103k ⌀ | path stringlengths 8 216 | filename stringlengths 2 179 | language stringclasses 15
values | size_bytes int64 2 189k | quality_score float64 0.5 0.95 | complexity float64 0 1 | documentation_ratio float64 0 1 | repository stringclasses 5
values | stars int64 0 1k | created_date stringdate 2023-07-10 19:21:08 2025-07-09 19:11:45 | license stringclasses 4
values | is_test bool 2
classes | file_hash stringlengths 32 32 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
# If we decide to run build-image.yml on every PR, we could deprecate this workflow.\nname: Docker Build\n\non:\n workflow_dispatch:\n pull_request:\n paths-ignore:\n - '**/*.md'\n push:\n branches:\n - 'master'\n\nconcurrency:\n group: ${{ github.workflow }}-${{ github.event_name }}-${{ github.event_name == 'push' && github.sha || github.ref }}\n cancel-in-progress: true\n\njobs:\n docker-build:\n if: github.repository == 'ipfs/kubo' || github.event_name == 'workflow_dispatch'\n runs-on: ubuntu-latest\n timeout-minutes: 10\n env:\n IMAGE_NAME: ipfs/kubo\n WIP_IMAGE_TAG: wip\n defaults:\n run:\n shell: bash\n steps:\n - uses: actions/setup-go@v5\n with:\n go-version: 1.24.x\n - uses: actions/checkout@v4\n - run: docker build -t $IMAGE_NAME:$WIP_IMAGE_TAG .\n - run: docker run --rm $IMAGE_NAME:$WIP_IMAGE_TAG --version\n | dataset_sample\yaml\ipfs_kubo\.github\workflows\docker-build.yml | docker-build.yml | YAML | 894 | 0.8 | 0.029412 | 0.032258 | react-lib | 813 | 2025-06-01T18:04:12.880245 | Apache-2.0 | false | 99a94c36c7f829c1a638efb883c4b34e |
name: Gateway Conformance\n\non:\n workflow_dispatch:\n push:\n branches:\n - master\n pull_request:\n paths-ignore:\n - '**/*.md'\n\nconcurrency:\n group: ${{ github.workflow }}-${{ github.event_name }}-${{ github.event_name == 'push' && github.sha || github.ref }}\n cancel-in-progress: true\n\ndefaults:\n run:\n shell: bash\n\nenv:\n # hostnames expected by https://github.com/ipfs/gateway-conformance\n GATEWAY_PUBLIC_GATEWAYS: |\n {\n "example.com": {\n "UseSubdomains": true,\n "InlineDNSLink": true,\n "Paths": ["/ipfs", "/ipns"]\n },\n "localhost": {\n "UseSubdomains": true,\n "InlineDNSLink": true,\n "Paths": ["/ipfs", "/ipns"]\n }\n }\n\njobs:\n # Testing all gateway features via TCP port specified in Addresses.Gateway\n gateway-conformance:\n runs-on: ubuntu-latest\n timeout-minutes: 10\n steps:\n # 1. Download the gateway-conformance fixtures\n - name: Download gateway-conformance fixtures\n uses: ipfs/gateway-conformance/.github/actions/extract-fixtures@v0.7\n with:\n output: fixtures\n\n # 2. Build the kubo-gateway\n - name: Setup Go\n uses: actions/setup-go@v5\n with:\n go-version: 1.24.x\n - uses: protocol/cache-go-action@v1\n with:\n name: ${{ github.job }}\n - name: Checkout kubo-gateway\n uses: actions/checkout@v4\n with:\n path: kubo-gateway\n - name: Build kubo-gateway\n run: make build\n working-directory: kubo-gateway\n\n # 3. Init the kubo-gateway\n - name: Init kubo-gateway\n run: |\n ./ipfs init -e\n ./ipfs config --json Gateway.PublicGateways "$GATEWAY_PUBLIC_GATEWAYS"\n working-directory: kubo-gateway/cmd/ipfs\n\n # 4. Populate the Kubo gateway with the gateway-conformance fixtures\n - name: Import fixtures\n run: |\n # Import car files\n find ./fixtures -name '*.car' -exec kubo-gateway/cmd/ipfs/ipfs dag import --pin-roots=false {} \;\n\n # Import ipns records\n records=$(find ./fixtures -name '*.ipns-record')\n for record in $records\n do\n key=$(basename -s .ipns-record "$record" | cut -d'_' -f1)\n kubo-gateway/cmd/ipfs/ipfs routing put --allow-offline "/ipns/$key" "$record"\n done\n\n # Import dnslink records\n # the IPFS_NS_MAP env will be used by the daemon\n echo "IPFS_NS_MAP=$(cat ./fixtures/dnslinks.IPFS_NS_MAP)" >> $GITHUB_ENV\n\n # 5. Start the kubo-gateway\n - name: Start kubo-gateway\n run: |\n ./ipfs daemon --offline &\n working-directory: kubo-gateway/cmd/ipfs\n\n # 6. Run the gateway-conformance tests\n - name: Run gateway-conformance tests\n uses: ipfs/gateway-conformance/.github/actions/test@v0.7\n with:\n gateway-url: http://127.0.0.1:8080\n subdomain-url: http://localhost:8080\n args: -skip 'TestGatewayCar/GET_response_for_application/vnd.ipld.car/Header_Content-Length'\n json: output.json\n xml: output.xml\n html: output.html\n markdown: output.md\n\n # 7. Upload the results\n - name: Upload MD summary\n if: failure() || success()\n run: cat output.md >> $GITHUB_STEP_SUMMARY\n - name: Upload HTML report\n if: failure() || success()\n uses: actions/upload-artifact@v4\n with:\n name: gateway-conformance.html\n path: output.html\n - name: Upload JSON report\n if: failure() || success()\n uses: actions/upload-artifact@v4\n with:\n name: gateway-conformance.json\n path: output.json\n\n # Testing trustless gateway feature subset exposed as libp2p protocol\n gateway-conformance-libp2p-experiment:\n runs-on: ubuntu-latest\n timeout-minutes: 10\n steps:\n # 1. Download the gateway-conformance fixtures\n - name: Download gateway-conformance fixtures\n uses: ipfs/gateway-conformance/.github/actions/extract-fixtures@v0.7\n with:\n output: fixtures\n\n # 2. Build the kubo-gateway\n - name: Setup Go\n uses: actions/setup-go@v5\n with:\n go-version: 1.24.x\n - uses: protocol/cache-go-action@v1\n with:\n name: ${{ github.job }}\n - name: Checkout kubo-gateway\n uses: actions/checkout@v4\n with:\n path: kubo-gateway\n - name: Build kubo-gateway\n run: make build\n working-directory: kubo-gateway\n\n # 3. Init the kubo-gateway\n - name: Init kubo-gateway\n run: |\n ./ipfs init --profile=test\n ./ipfs config --json Gateway.PublicGateways "$GATEWAY_PUBLIC_GATEWAYS"\n ./ipfs config --json Experimental.GatewayOverLibp2p true\n ./ipfs config Addresses.Gateway "/ip4/127.0.0.1/tcp/8080"\n ./ipfs config Addresses.API "/ip4/127.0.0.1/tcp/5001"\n working-directory: kubo-gateway/cmd/ipfs\n\n # 4. Populate the Kubo gateway with the gateway-conformance fixtures\n - name: Import fixtures\n run: |\n # Import car files\n find ./fixtures -name '*.car' -exec kubo-gateway/cmd/ipfs/ipfs dag import --pin-roots=false {} \;\n\n # 5. Start the kubo-gateway\n - name: Start kubo-gateway\n run: |\n ( ./ipfs daemon & ) | sed '/Daemon is ready/q'\n while [[ "$(./ipfs id | jq '.Addresses | length')" == '0' ]]; do sleep 1; done\n working-directory: kubo-gateway/cmd/ipfs\n\n # 6. Setup a kubo http-p2p-proxy to expose libp2p protocol as a regular HTTP port for gateway conformance tests\n - name: Init p2p-proxy kubo node\n env:\n IPFS_PATH: "~/.kubo-p2p-proxy"\n run: |\n ./ipfs init --profile=test -e\n ./ipfs config --json Experimental.Libp2pStreamMounting true\n ./ipfs config Addresses.Gateway "/ip4/127.0.0.1/tcp/8081"\n ./ipfs config Addresses.API "/ip4/127.0.0.1/tcp/5002"\n working-directory: kubo-gateway/cmd/ipfs\n\n # 7. Start the kubo http-p2p-proxy\n - name: Start kubo http-p2p-proxy\n env:\n IPFS_PATH: "~/.kubo-p2p-proxy"\n run: |\n ( ./ipfs daemon & ) | sed '/Daemon is ready/q'\n while [[ "$(./ipfs id | jq '.Addresses | length')" == '0' ]]; do sleep 1; done\n working-directory: kubo-gateway/cmd/ipfs\n\n # 8. Start forwarding data from the http-p2p-proxy to the node serving the Gateway API over libp2p\n - name: Start http-over-libp2p forwarding proxy\n run: |\n gatewayNodeId=$(./ipfs --api=/ip4/127.0.0.1/tcp/5001 id -f="<id>")\n ./ipfs --api=/ip4/127.0.0.1/tcp/5002 swarm connect $(./ipfs --api=/ip4/127.0.0.1/tcp/5001 swarm addrs local --id | head -n 1)\n ./ipfs --api=/ip4/127.0.0.1/tcp/5002 p2p forward --allow-custom-protocol /http/1.1 /ip4/127.0.0.1/tcp/8092 /p2p/$gatewayNodeId\n working-directory: kubo-gateway/cmd/ipfs\n\n # 9. Run the gateway-conformance tests over libp2p\n - name: Run gateway-conformance tests over libp2p\n uses: ipfs/gateway-conformance/.github/actions/test@v0.7\n with:\n gateway-url: http://127.0.0.1:8092\n args: --specs "trustless-gateway,-trustless-ipns-gateway" -skip 'TestGatewayCar/GET_response_for_application/vnd.ipld.car/Header_Content-Length'\n json: output.json\n xml: output.xml\n html: output.html\n markdown: output.md\n\n # 10. Upload the results\n - name: Upload MD summary\n if: failure() || success()\n run: cat output.md >> $GITHUB_STEP_SUMMARY\n - name: Upload HTML report\n if: failure() || success()\n uses: actions/upload-artifact@v4\n with:\n name: gateway-conformance-libp2p.html\n path: output.html\n - name: Upload JSON report\n if: failure() || success()\n uses: actions/upload-artifact@v4\n with:\n name: gateway-conformance-libp2p.json\n path: output.json\n | dataset_sample\yaml\ipfs_kubo\.github\workflows\gateway-conformance.yml | gateway-conformance.yml | YAML | 8,011 | 0.95 | 0.04386 | 0.121951 | awesome-app | 280 | 2025-01-24T01:53:01.128453 | GPL-3.0 | false | 751dbde3e3326555a121d986fd9e266f |
name: Go Build\n\non:\n workflow_dispatch:\n pull_request:\n paths-ignore:\n - '**/*.md'\n push:\n branches:\n - 'master'\n\nconcurrency:\n group: ${{ github.workflow }}-${{ github.event_name }}-${{ github.event_name == 'push' && github.sha || github.ref }}\n cancel-in-progress: true\n\njobs:\n go-build:\n if: github.repository == 'ipfs/kubo' || github.event_name == 'workflow_dispatch'\n runs-on: ${{ fromJSON(github.repository == 'ipfs/kubo' && '["self-hosted", "linux", "x64", "4xlarge"]' || '"ubuntu-latest"') }}\n timeout-minutes: 20\n env:\n TEST_DOCKER: 0\n TEST_VERBOSE: 1\n TRAVIS: 1\n GIT_PAGER: cat\n IPFS_CHECK_RCMGR_DEFAULTS: 1\n defaults:\n run:\n shell: bash\n steps:\n - uses: actions/setup-go@v5\n with:\n go-version: 1.24.x\n - uses: actions/checkout@v4\n - run: make cmd/ipfs-try-build\n env:\n TEST_FUSE: 1\n - run: make cmd/ipfs-try-build\n env:\n TEST_FUSE: 0\n | dataset_sample\yaml\ipfs_kubo\.github\workflows\gobuild.yml | gobuild.yml | YAML | 990 | 0.8 | 0.075 | 0 | vue-tools | 875 | 2023-10-28T00:24:22.058952 | GPL-3.0 | false | 031d43db1c1431c67c941a21a8738dd6 |
name: Go Check\n\non:\n workflow_dispatch:\n pull_request:\n paths-ignore:\n - '**/*.md'\n push:\n branches:\n - 'master'\n\npermissions:\n contents: read # to fetch code (actions/checkout)\n\nconcurrency:\n group: ${{ github.workflow }}-${{ github.event_name }}-${{ github.event_name == 'push' && github.sha || github.ref }}\n cancel-in-progress: true\n\njobs:\n go-check:\n if: github.repository == 'ipfs/kubo' || github.event_name == 'workflow_dispatch'\n runs-on: ubuntu-latest\n timeout-minutes: 10\n steps:\n - uses: actions/checkout@v4\n with:\n submodules: recursive\n - uses: actions/setup-go@v5\n with:\n go-version: "1.24.x"\n - name: Check that go.mod is tidy\n uses: protocol/multiple-go-modules@v1.4\n with:\n run: |\n go mod tidy\n if [[ -n $(git ls-files --other --exclude-standard --directory -- go.sum) ]]; then\n echo "go.sum was added by go mod tidy"\n exit 1\n fi\n git diff --exit-code -- go.sum go.mod\n - name: go fmt\n if: always() # run this step even if the previous one failed\n run: |\n out=$(go fmt ./...)\n if [[ -n "$out" ]]; then\n echo "Files are not go-fmt-ed:"\n echo "$out"\n exit 1\n fi\n - name: go vet\n if: always() # run this step even if the previous one failed\n uses: protocol/multiple-go-modules@v1.4\n with:\n run: go vet ./...\n | dataset_sample\yaml\ipfs_kubo\.github\workflows\golang-analysis.yml | golang-analysis.yml | YAML | 1,513 | 0.8 | 0.12963 | 0 | react-lib | 735 | 2023-07-20T11:39:18.848137 | BSD-3-Clause | false | d91402ae354ca6670a72efacbf3565a6 |
name: Go Lint\n\non:\n workflow_dispatch:\n pull_request:\n paths-ignore:\n - '**/*.md'\n push:\n branches:\n - 'master'\n\nconcurrency:\n group: ${{ github.workflow }}-${{ github.event_name }}-${{ github.event_name == 'push' && github.sha || github.ref }}\n cancel-in-progress: true\n\njobs:\n go-lint:\n if: github.repository == 'ipfs/kubo' || github.event_name == 'workflow_dispatch'\n runs-on: ubuntu-latest\n timeout-minutes: 10\n env:\n TEST_DOCKER: 0\n TEST_FUSE: 0\n TEST_VERBOSE: 1\n TRAVIS: 1\n GIT_PAGER: cat\n IPFS_CHECK_RCMGR_DEFAULTS: 1\n defaults:\n run:\n shell: bash\n steps:\n - uses: actions/setup-go@v5\n with:\n go-version: 1.24.x\n - uses: actions/checkout@v4\n - run: make -O test_go_lint\n | dataset_sample\yaml\ipfs_kubo\.github\workflows\golint.yml | golint.yml | YAML | 790 | 0.8 | 0.027778 | 0 | node-utils | 140 | 2025-03-23T12:32:48.321907 | MIT | false | b409ceb3aaf9d9bef319d05f69ebc3c2 |
name: Interop\n\non:\n workflow_dispatch:\n pull_request:\n paths-ignore:\n - '**/*.md'\n push:\n branches:\n - 'master'\n\nenv:\n GO_VERSION: 1.24.x\n\nconcurrency:\n group: ${{ github.workflow }}-${{ github.event_name }}-${{ github.event_name == 'push' && github.sha || github.ref }}\n cancel-in-progress: true\n\ndefaults:\n run:\n shell: bash\n\njobs:\n interop-prep:\n if: github.repository == 'ipfs/kubo' || github.event_name == 'workflow_dispatch'\n runs-on: ubuntu-latest\n timeout-minutes: 5\n env:\n TEST_DOCKER: 0\n TEST_FUSE: 0\n TEST_VERBOSE: 1\n TRAVIS: 1\n GIT_PAGER: cat\n IPFS_CHECK_RCMGR_DEFAULTS: 1\n defaults:\n run:\n shell: bash\n steps:\n - uses: actions/setup-go@v5\n with:\n go-version: ${{ env.GO_VERSION }}\n - uses: actions/checkout@v4\n - run: make build\n - uses: actions/upload-artifact@v4\n with:\n name: kubo\n path: cmd/ipfs/ipfs\n helia-interop:\n needs: [interop-prep]\n runs-on: ${{ fromJSON(github.repository == 'ipfs/kubo' && '["self-hosted", "linux", "x64", "2xlarge"]' || '"ubuntu-latest"') }}\n timeout-minutes: 20\n defaults:\n run:\n shell: bash\n steps:\n - uses: actions/setup-node@v4\n with:\n node-version: lts/*\n - uses: actions/download-artifact@v4\n with:\n name: kubo\n path: cmd/ipfs\n - run: chmod +x cmd/ipfs/ipfs\n - run: echo "dir=$(npm config get cache)" >> $GITHUB_OUTPUT\n id: npm-cache-dir\n - uses: actions/cache@v4\n with:\n path: ${{ steps.npm-cache-dir.outputs.dir }}\n key: ${{ runner.os }}-${{ github.job }}-helia-${{ hashFiles('**/package-lock.json') }}\n restore-keys: ${{ runner.os }}-${{ github.job }}-helia-\n - run: sudo apt update\n - run: sudo apt install -y libxkbcommon0 libxdamage1 libgbm1 libpango-1.0-0 libcairo2 # dependencies for playwright\n - run: npx --package @helia/interop helia-interop\n env:\n KUBO_BINARY: ${{ github.workspace }}/cmd/ipfs/ipfs\n ipfs-webui:\n needs: [interop-prep]\n runs-on: ${{ fromJSON(github.repository == 'ipfs/kubo' && '["self-hosted", "linux", "x64", "2xlarge"]' || '"ubuntu-latest"') }}\n timeout-minutes: 20\n env:\n NO_SANDBOX: true\n LIBP2P_TCP_REUSEPORT: false\n LIBP2P_ALLOW_WEAK_RSA_KEYS: 1\n E2E_IPFSD_TYPE: go\n TRAVIS: 1\n GIT_PAGER: cat\n IPFS_CHECK_RCMGR_DEFAULTS: 1\n defaults:\n run:\n shell: bash\n steps:\n - uses: actions/setup-node@v4\n with:\n node-version: 18.14.0\n - uses: actions/download-artifact@v4\n with:\n name: kubo\n path: cmd/ipfs\n - run: chmod +x cmd/ipfs/ipfs\n - uses: actions/checkout@v4\n with:\n repository: ipfs/ipfs-webui\n path: ipfs-webui\n - run: |\n echo "dir=$(npm config get cache)" >> $GITHUB_OUTPUT\n id: npm-cache-dir\n - uses: actions/cache@v4\n with:\n path: ${{ steps.npm-cache-dir.outputs.dir }}\n key: ${{ runner.os }}-${{ github.job }}-${{ hashFiles('**/package-lock.json') }}\n restore-keys: |\n ${{ runner.os }}-${{ github.job }}-\n - env:\n NPM_CACHE_DIR: ${{ steps.npm-cache-dir.outputs.dir }}\n run: |\n npm ci --prefer-offline --no-audit --progress=false --cache "$NPM_CACHE_DIR"\n npx playwright install --with-deps\n working-directory: ipfs-webui\n - id: ref\n run: echo "ref=$(git rev-parse --short HEAD)" | tee -a $GITHUB_OUTPUT\n working-directory: ipfs-webui\n - id: state\n env:\n GITHUB_TOKEN: ${{ github.token }}\n ENDPOINT: repos/ipfs/ipfs-webui/commits/${{ steps.ref.outputs.ref }}/status\n SELECTOR: .state\n KEY: state\n run: gh api "$ENDPOINT" --jq "$SELECTOR" | xargs -I{} echo "$KEY={}" | tee -a $GITHUB_OUTPUT\n - name: Build ipfs-webui@main (state=${{ steps.state.outputs.state }})\n run: npm run test:build\n working-directory: ipfs-webui\n - name: Test ipfs-webui@main (state=${{ steps.state.outputs.state }}) E2E against the locally built Kubo binary\n run: npm run test:e2e\n env:\n IPFS_GO_EXEC: ${{ github.workspace }}/cmd/ipfs/ipfs\n working-directory: ipfs-webui\n | dataset_sample\yaml\ipfs_kubo\.github\workflows\interop.yml | interop.yml | YAML | 4,342 | 0.8 | 0.014706 | 0 | node-utils | 244 | 2023-11-23T22:11:40.581662 | GPL-3.0 | false | 4bc61e5a1d5be3809cdee88781269441 |
name: Sharness\n\non:\n workflow_dispatch:\n pull_request:\n paths-ignore:\n - '**/*.md'\n push:\n branches:\n - 'master'\n\nconcurrency:\n group: ${{ github.workflow }}-${{ github.event_name }}-${{ github.event_name == 'push' && github.sha || github.ref }}\n cancel-in-progress: true\n\njobs:\n sharness-test:\n if: github.repository == 'ipfs/kubo' || github.event_name == 'workflow_dispatch'\n runs-on: ${{ fromJSON(github.repository == 'ipfs/kubo' && '["self-hosted", "linux", "x64", "4xlarge"]' || '"ubuntu-latest"') }}\n timeout-minutes: ${{ github.repository == 'ipfs/kubo' && 15 || 60 }}\n defaults:\n run:\n shell: bash\n steps:\n - name: Setup Go\n uses: actions/setup-go@v5\n with:\n go-version: 1.24.x\n - name: Checkout Kubo\n uses: actions/checkout@v4\n with:\n path: kubo\n - name: Install missing tools\n run: sudo apt update && sudo apt install -y socat net-tools fish libxml2-utils\n - uses: actions/cache@v4\n with:\n path: test/sharness/lib/dependencies\n key: ${{ runner.os }}-test-generate-junit-html-${{ hashFiles('test/sharness/lib/test-generate-junit-html.sh') }}\n - name: Run Sharness tests\n run: |\n make -O -j "$PARALLEL" \\n test_sharness \\n coverage/sharness_tests.coverprofile \\n test/sharness/test-results/sharness.xml\n working-directory: kubo\n env:\n TEST_DOCKER: 1\n TEST_PLUGIN: 0\n TEST_FUSE: 0\n TEST_VERBOSE: 1\n TEST_JUNIT: 1\n TEST_EXPENSIVE: 1\n IPFS_CHECK_RCMGR_DEFAULTS: 1\n CONTINUE_ON_S_FAILURE: 1\n # increasing parallelism beyond 10 doesn't speed up the tests much\n PARALLEL: ${{ github.repository == 'ipfs/kubo' && 10 || 3 }}\n - name: Upload coverage report\n uses: codecov/codecov-action@0565863a31f2c772f9f0395002a31e3f06189574 # v5.4.0\n if: failure() || success()\n with:\n name: sharness\n files: kubo/coverage/sharness_tests.coverprofile\n - name: Aggregate results\n run: find kubo/test/sharness/test-results -name 't*-*.sh.*.counts' | kubo/test/sharness/lib/sharness/aggregate-results.sh > kubo/test/sharness/test-results/summary.txt\n - name: 👉️ If this step failed, go to «Summary» (top left) → «HTML Report» → inspect the «Failures» column\n run: |\n cat kubo/test/sharness/test-results/summary.txt &&\n grep 'failed\s*0' kubo/test/sharness/test-results/summary.txt\n - name: Add aggregate results to the summary\n if: failure() || success()\n run: |\n echo "# Summary" >> $GITHUB_STEP_SUMMARY\n echo >> $GITHUB_STEP_SUMMARY\n cat kubo/test/sharness/test-results/summary.txt >> $GITHUB_STEP_SUMMARY\n - name: Generate one-page HTML report\n uses: ipdxco/junit-xml-to-html@v1\n if: failure() || success()\n with:\n mode: no-frames\n input: kubo/test/sharness/test-results/sharness.xml\n output: kubo/test/sharness/test-results/sharness.html\n - name: Upload one-page HTML report to S3\n id: one-page\n uses: ipdxco/custom-github-runners/.github/actions/upload-artifact@main\n if: github.repository == 'ipfs/kubo' && (failure() || success())\n with:\n source: kubo/test/sharness/test-results/sharness.html\n destination: sharness.html\n - name: Upload one-page HTML report\n if: github.repository != 'ipfs/kubo' && (failure() || success())\n uses: actions/upload-artifact@v4\n with:\n name: sharness.html\n path: kubo/test/sharness/test-results/sharness.html\n - name: Generate full HTML report\n uses: ipdxco/junit-xml-to-html@v1\n if: failure() || success()\n with:\n mode: frames\n input: kubo/test/sharness/test-results/sharness.xml\n output: kubo/test/sharness/test-results/sharness-html\n - name: Upload full HTML report to S3\n id: full\n uses: ipdxco/custom-github-runners/.github/actions/upload-artifact@main\n if: github.repository == 'ipfs/kubo' && (failure() || success())\n with:\n source: kubo/test/sharness/test-results/sharness-html\n destination: sharness-html/\n - name: Upload full HTML report\n if: github.repository != 'ipfs/kubo' && (failure() || success())\n uses: actions/upload-artifact@v4\n with:\n name: sharness-html\n path: kubo/test/sharness/test-results/sharness-html\n - name: Add S3 links to the summary\n if: github.repository == 'ipfs/kubo' && (failure() || success())\n run: echo "$MD" >> $GITHUB_STEP_SUMMARY\n env:\n MD: |\n # HTML Reports\n\n - View the [one page HTML report](${{ steps.one-page.outputs.url }})\n - View the [full HTML report](${{ steps.full.outputs.url }}index.html)\n | dataset_sample\yaml\ipfs_kubo\.github\workflows\sharness.yml | sharness.yml | YAML | 4,982 | 0.8 | 0.081301 | 0.016807 | awesome-app | 103 | 2023-11-25T10:52:57.234121 | MIT | false | d28f5c08c97eb96f69361f737f079ba7 |
name: Spell Check\n\non: [push, pull_request]\n\njobs:\n spellcheck:\n runs-on: ubuntu-latest\n\n steps:\n - name: Checkout code\n uses: actions/checkout@v4\n\n - name: Install Codespell\n run: pip install codespell==2.4.0\n\n - name: Run Codespell\n uses: codespell-project/actions-codespell@v2\n with:\n only_warn: 1\n ignore_words_file: .codespell-ignore\n skip: "*.mod,*.sum,*.pdf,./docs/AUTHORS,./test/sharness/t0275-cid-security-data,./test/sharness/t0280-plugin-dag-jose-data,./bin"\n | dataset_sample\yaml\ipfs_kubo\.github\workflows\spellcheck.yml | spellcheck.yml | YAML | 547 | 0.7 | 0 | 0 | react-lib | 4 | 2024-06-08T03:15:31.021815 | GPL-3.0 | false | 81c3c80b75eebf87abdab7079dc0923a |
name: Close and mark stale issue\n\non:\n schedule:\n - cron: '0 0 * * *'\n\npermissions:\n issues: write\n pull-requests: write\n\njobs:\n stale:\n uses: pl-strflt/.github/.github/workflows/reusable-stale-issue.yml@v0.3\n | dataset_sample\yaml\ipfs_kubo\.github\workflows\stale.yml | stale.yml | YAML | 219 | 0.7 | 0 | 0 | awesome-app | 261 | 2024-11-08T15:33:38.165772 | MIT | false | bc57ebe620fc2b8e77d8752759accc84 |
data:\n build: ./data\n volumes:\n - /data\n command: sleep 1000000\n\nbootstrap:\n build: ./bootstrap\n command: daemon --debug --init\n expose:\n - "4011"\n - "4012/udp"\n environment:\n GOLOG_LOG_LEVEL: debug\n\nserver:\n build: ./server\n links:\n - bootstrap\n volumes_from:\n - data\n expose:\n - "4021"\n - "4022/udp"\n environment:\n GOLOG_LOG_LEVEL: debug\n\nclient:\n build: ./client\n links:\n - bootstrap\n volumes_from:\n - data\n expose:\n - "4031"\n - "4032/udp"\n environment:\n GOLOG_LOG_LEVEL: debug\n | dataset_sample\yaml\ipfs_kubo\test\3nodetest\fig.yml | fig.yml | YAML | 630 | 0.7 | 0 | 0 | react-lib | 864 | 2024-07-03T03:29:29.594569 | BSD-3-Clause | true | 06399bf04e6ca1305ce70b52613618ad |
name: Cassandra\nversion: 'trunk'\ndisplay_version: 'trunk'\nprerelease: true\nasciidoc:\n attributes:\n cass_url: 'http://cassandra.apache.org/'\n cass-50: 'Cassandra 5.0'\n cassandra: 'Cassandra' \n product: 'Apache Cassandra'\n\nnav:\n- modules/ROOT/nav.adoc\n- modules/cassandra/nav.adoc | dataset_sample\yaml\java\antora.yml | antora.yml | YAML | 291 | 0.8 | 0 | 0 | awesome-app | 132 | 2024-04-29T14:07:12.072206 | Apache-2.0 | false | dbb403f93f945c180033024b00337878 |
#\n# Licensed to the Apache Software Foundation (ASF) under one\n# or more contributor license agreements. See the NOTICE file\n# distributed with this work for additional information\n# regarding copyright ownership. The ASF licenses this file\n# to you under the Apache License, Version 2.0 (the\n# "License"); you may not use this file except in compliance\n# with the License. You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing,\n# software distributed under the License is distributed on an\n# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY\n# KIND, either express or implied. See the License for the\n# specific language governing permissions and limitations\n# under the License.\n#\n\nINFRA:\n - changed-files:\n - any-glob-to-any-file: [\n '.github/**/*',\n 'tools/**/*',\n 'dev/create-release/**/*',\n '.asf.yaml',\n '.gitattributes',\n '.gitignore',\n 'dev/merge_spark_pr.py'\n ]\n\nBUILD:\n - changed-files:\n - all-globs-to-any-file: [\n 'dev/**/*',\n '!dev/merge_spark_pr.py'\n ]\n - any-glob-to-any-file: [\n 'build/**/*',\n 'project/**/*',\n 'assembly/**/*',\n '**/*pom.xml',\n 'bin/docker-image-tool.sh',\n 'bin/find-spark-home*',\n 'scalastyle-config.xml'\n ]\n\nDOCS:\n - changed-files:\n - any-glob-to-any-file: [\n 'docs/**/*',\n '**/README.md',\n '**/CONTRIBUTING.md',\n 'python/docs/**/*'\n ]\n\nEXAMPLES:\n - changed-files:\n - any-glob-to-any-file: [\n 'examples/**/*',\n 'bin/run-example*'\n ]\n\nCORE:\n - changed-files:\n - all-globs-to-any-file: [\n 'core/**/*',\n '!**/*UI.scala',\n '!**/ui/**/*'\n ]\n - any-glob-to-any-file: [\n 'common/kvstore/**/*',\n 'common/network-common/**/*',\n 'common/network-shuffle/**/*',\n 'python/pyspark/*.py',\n 'python/pyspark/tests/**/*.py'\n ]\n\nSPARK SUBMIT:\n - changed-files:\n - any-glob-to-any-file: [\n 'bin/spark-submit*'\n ]\n\nSPARK SHELL:\n - changed-files:\n - any-glob-to-any-file: [\n 'repl/**/*',\n 'bin/spark-shell*'\n ]\n\nSQL:\n - changed-files:\n - all-globs-to-any-file: [\n '**/sql/**/*',\n '!python/**/avro/**/*',\n '!python/**/protobuf/**/*',\n '!python/**/streaming/**/*'\n ]\n - any-glob-to-any-file: [\n 'common/unsafe/**/*',\n 'common/sketch/**/*',\n 'common/variant/**/*',\n 'bin/spark-sql*',\n 'bin/beeline*',\n 'sbin/*thriftserver*.sh',\n '**/*SQL*.R',\n '**/DataFrame.R',\n '**/*WindowSpec.R',\n '**/*catalog.R',\n '**/*column.R',\n '**/*functions.R',\n '**/*group.R',\n '**/*schema.R',\n '**/*types.R'\n ]\n\nAVRO:\n - changed-files:\n - any-glob-to-any-file: [\n 'connector/avro/**/*',\n 'python/**/avro/**/*'\n ]\n\nDSTREAM:\n - changed-files:\n - any-glob-to-any-file: [\n 'streaming/**/*',\n 'data/streaming/**/*',\n 'connector/kinesis-asl/**/*',\n 'connector/kinesis-asl-assembly/**/*',\n 'connector/kafka-0-10/**/*',\n 'connector/kafka-0-10-assembly/**/*',\n 'connector/kafka-0-10-token-provider/**/*',\n 'python/pyspark/streaming/**/*'\n ]\n\nGRAPHX:\n - changed-files:\n - any-glob-to-any-file: [\n 'graphx/**/*',\n 'data/graphx/**/*'\n ]\n\nML:\n - changed-files:\n - any-glob-to-any-file: [\n '**/ml/**/*',\n '**/*mllib_*.R'\n ]\n\nMLLIB:\n - changed-files:\n - any-glob-to-any-file: [\n '**/mllib/**/*',\n 'mllib-local/**/*'\n ]\n\nSTRUCTURED STREAMING:\n - changed-files:\n - any-glob-to-any-file: [\n '**/sql/**/streaming/**/*',\n 'connector/kafka-0-10-sql/**/*',\n 'python/pyspark/sql/**/streaming/**/*',\n '**/*streaming.R'\n ]\n\nPYTHON:\n - changed-files:\n - any-glob-to-any-file: [\n 'bin/pyspark*',\n '**/python/**/*'\n ]\n\nPANDAS API ON SPARK:\n - changed-files:\n - any-glob-to-any-file: [\n 'python/pyspark/pandas/**/*'\n ]\n\nR:\n - changed-files:\n - any-glob-to-any-file: [\n '**/r/**/*',\n '**/R/**/*',\n 'bin/sparkR*'\n ]\n\nYARN:\n - changed-files:\n - any-glob-to-any-file: [\n 'resource-managers/yarn/**/*'\n ]\n\nKUBERNETES:\n - changed-files:\n - any-glob-to-any-file: [\n 'bin/docker-image-tool.sh',\n 'resource-managers/kubernetes/**/*'\n ]\n\nWINDOWS:\n - changed-files:\n - any-glob-to-any-file: [\n '**/*.cmd',\n 'R/pkg/tests/fulltests/test_Windows.R'\n ]\n\nWEB UI:\n - changed-files:\n - any-glob-to-any-file: [\n '**/ui/**/*',\n '**/*UI.scala'\n ]\n\nDEPLOY:\n - changed-files:\n - any-glob-to-any-file: [\n 'sbin/**/*'\n ]\n\nCONNECT:\n - changed-files:\n - any-glob-to-any-file: [\n 'sql/connect/**/*',\n 'python/**/connect/**/*'\n ]\n\nPROTOBUF:\n - changed-files:\n - any-glob-to-any-file: [\n 'connector/protobuf/**/*',\n 'python/**/protobuf/**/*'\n ]\n | dataset_sample\yaml\java\labeler.yml | labeler.yml | YAML | 4,827 | 0.95 | 0.008547 | 0.085714 | node-utils | 27 | 2024-10-30T00:13:44.537112 | MIT | false | 4bbf02958bb34e49e5583878805ce12f |
#\n# Licensed to the Apache Software Foundation (ASF) under one\n# or more contributor license agreements. See the NOTICE file\n# distributed with this work for additional information\n# regarding copyright ownership. The ASF licenses this file\n# to you under the Apache License, Version 2.0 (the\n# "License"); you may not use this file except in compliance\n# with the License. You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an "AS IS" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n#\n\n | dataset_sample\yaml\java\license.yml | license.yml | YAML | 789 | 0.95 | 0.111111 | 1 | react-lib | 546 | 2024-11-28T21:53:20.528552 | Apache-2.0 | false | ed3e3e8d27b99c8093852dfbbcc51f1c |
site_name: OkHttp\nsite_url: https://square.github.io/okhttp/\nrepo_name: OkHttp\nrepo_url: https://github.com/square/okhttp\nsite_description: "Square’s meticulous HTTP client for the JVM, Android, and GraalVM"\nsite_author: Square, Inc.\nremote_branch: gh-pages\nedit_uri: ""\n\ncopyright: 'Copyright © 2022 Block, Inc.'\n\ntheme:\n name: 'material'\n favicon: assets/images/icon-square.png\n logo: assets/images/icon-square.png\n palette:\n - media: "(prefers-color-scheme: light)"\n scheme: default\n primary: teal\n accent: blue\n toggle:\n icon: octicons/sun-24\n name: "Switch to Dark Mode"\n - media: "(prefers-color-scheme: dark)"\n scheme: slate\n primary: teal\n accent: blue\n toggle:\n icon: octicons/moon-24\n name: "Switch to Light Mode"\n features:\n - navigation.tabs\n\nextra_css:\n - 'assets/css/app.css'\n\nmarkdown_extensions:\n - smarty\n - footnotes\n - meta\n - toc:\n permalink: true\n - attr_list\n - pymdownx.betterem:\n smart_enable: all\n - pymdownx.caret\n - pymdownx.emoji:\n emoji_index: !!python/name:material.extensions.emoji.twemoji\n emoji_generator: !!python/name:material.extensions.emoji.to_svg\n - pymdownx.inlinehilite\n - pymdownx.magiclink\n - pymdownx.smartsymbols\n - pymdownx.superfences\n - pymdownx.tilde\n - pymdownx.tabbed:\n alternate_style: true\n - tables\n\nplugins:\n - search\n - redirects:\n redirect_maps:\n # Redirect all feature pages to features/*\n 'caching.md': 'features/caching.md'\n 'calls.md': 'features/calls.md'\n 'connections.md': 'features/connections.md'\n 'events.md': 'features/events.md'\n 'https.md': 'features/events.md'\n 'interceptors.md': 'features/interceptors.md'\n 'r8_proguard.md': 'features/r8_proguard.md'\n # Redirect all Security pages to security/*\n 'security.md': 'security/security.md'\n 'security_providers.md': 'security/security_providers.md'\n 'tls_configuration_history.md': 'security/tls_configuration_history.md'\n # Redirect all changelog pages to changelog/*\n 'changelog.md': 'changelogs/changelog.md'\n 'upgrading_to_okhttp_4.md': 'changelogs/upgrading_to_okhttp_4.md'\n 'changelog_3x.md': 'changelogs/changelog_3x.md'\n 'changelog_2x.md': 'changelogs/changelog_2x.md'\n 'changelog_1x.md': 'changelogs/changelog_1x.md'\n # Redirect all contributing pages to contribute/*\n 'contributing.md': 'contribute/contributing.md'\n 'code_of_conduct.md': 'contribute/code_of_conduct.md'\n 'concurrency.md': 'contribute/concurrency.md'\n 'debug_logging.md': 'contribute/debug_logging.md'\n\nnav:\n - 'Overview':\n - 'Overview': index.md\n - 'Stack Overflow': https://stackoverflow.com/questions/tagged/okhttp?sort=active\n - 'Features':\n - 'Calls': features/calls.md\n - 'Caching': features/caching.md\n - 'Connections': features/connections.md\n - 'Events': features/events.md\n - 'HTTPS': features/https.md\n - 'Interceptors': features/interceptors.md\n - 'R8/Proguard': features/r8_proguard.md\n - 'Recipes': recipes.md\n - 'Security':\n - 'Security': security/security.md\n - 'Providers': security/security_providers.md\n - 'Configuration History': security/tls_configuration_history.md\n - 'Works with OkHttp': works_with_okhttp.md\n - 'API': 5.x/okhttp/okhttp3/\n - 'Change Logs':\n - 'Change Log': changelogs/changelog.md\n - '4.x Change Log': changelogs/changelog_4x.md\n - 'Upgrading to OkHttp 4': changelogs/upgrading_to_okhttp_4.md\n - '3.x Change Log': changelogs/changelog_3x.md\n - '2.x Change Log': changelogs/changelog_2x.md\n - '1.x Change Log': changelogs/changelog_1x.md\n - 'Contributing':\n - 'Contributing': contribute/contributing.md\n - 'Code of Conduct': contribute/code_of_conduct.md\n - 'Concurrency': contribute/concurrency.md\n - 'Debug Logging': contribute/debug_logging.md\n | dataset_sample\yaml\java\mkdocs.yml | mkdocs.yml | YAML | 3,951 | 0.8 | 0.008547 | 0.036036 | vue-tools | 96 | 2024-08-11T09:37:37.842108 | GPL-3.0 | false | e38996f97aac41ea20d1cf40e8202b16 |
highlighter: rouge\nmarkdown: kramdown\ngems:\n - jekyll-redirect-from\n\n# For some reason kramdown seems to behave differently on different\n# OS/packages wrt encoding. So we hard code this config.\nkramdown:\n entity_output: numeric\n\nplugins:\n - jekyll-redirect-from\n\ninclude:\n - _static\n - _modules\n - _images\n - _sources\n\n# These allow the documentation to be updated with newer releases\n# of Spark, Scala.\nSPARK_VERSION: 4.1.0-SNAPSHOT\nSPARK_VERSION_SHORT: 4.1.0\nSCALA_BINARY_VERSION: "2.13"\nSCALA_VERSION: "2.13.16"\nSPARK_ISSUE_TRACKER_URL: https://issues.apache.org/jira/browse/SPARK\nSPARK_GITHUB_URL: https://github.com/apache/spark\n# Before a new release, we should:\n# 1. update the `version` array for the new Spark documentation\n# on https://github.com/algolia/docsearch-configs/blob/master/configs/apache_spark.json.\n# 2. update the value of `facetFilters.version` in `algoliaOptions` on the new release branch.\n# Otherwise, after release, the search results are always based on the latest documentation\n# (https://spark.apache.org/docs/latest/) even when visiting the documentation of previous releases.\nDOCSEARCH_SCRIPT: |\n docsearch({\n apiKey: 'd62f962a82bc9abb53471cb7b89da35e',\n appId: 'RAI69RXRSK',\n indexName: 'apache_spark',\n inputSelector: '#docsearch-input',\n enhancedSearchInput: true,\n algoliaOptions: {\n 'facetFilters': ["version:latest"]\n },\n debug: false // Set debug to true if you want to inspect the dropdown\n });\n\npermalink: 404.html\n\nexclude: ['README.md']\n | dataset_sample\yaml\java\_config.yml | _config.yml | YAML | 1,547 | 0.8 | 0.040816 | 0.232558 | awesome-app | 351 | 2024-01-25T18:38:26.357454 | MIT | false | 49f8a540e732173b86b860674e3c9992 |
####################################################################################\n####################################################################################\n# Angular Code Ownership #\n####################################################################################\n####################################################################################\n#\n# Configuration of code ownership and review approvals for the angular/angular repo.\n#\n# More info: https://docs.pullapprove.com/\n#\n# =========================================================\n# General rules / philosophy\n# =========================================================\n#\n# - We trust that people do the right thing and won't approve changes they don't feel confident reviewing\n# - We enforce that only approved PRs are merged ensuring that unreviewed code isn't accidentally merged\n# - We distribute approval rights as much as possible to help us scale better\n# - Groups have one or two global approvers groups as fallbacks:\n# - @angular/fw-global-approvers: for approving minor changes, large-scale refactorings, and emergency situations.\n# - @angular/fw-global-approvers-for-docs-only-changes: for approving minor documentation-only changes that don't require engineering review\n# - A small number of file groups have very limited number of reviewers because incorrect changes to the files they guard would have serious consequences (e.g. security, public api)\n#\n# Configuration nuances:\n#\n# - This configuration works in conjunction with the protected branch settings that require all changes to be made via pull requests with at least one approval.\n# - This approval can come from an appropriate codeowner, or any repo collaborator (person with write access) if the PR is authored by a codeowner.\n# - All groups whose pullapprove rules are matched will be required for overall approval.\n#\n# NOTE:\n# In the case of emergency, the repo administrators which include the current angular caretaker\n# can bypass this reviewer approval requirement, this is expected as a last resort and to be\n# done exceedingly sparingly.\n\n####################################################################################\n# GitHub usernames\n####################################################################################\n# See reviewer list under `required-minimum-review` group. Team member names and\n# usernames are managed there.\n\n####################################################################################\n# Approval Groups\n####################################################################################\n# =========================================================\n# @angular/framework-global-approvers\n# =========================================================\n# Used for approving minor changes, large-scale refactorings, and in emergency situations.\n#\n# alxhub\n# jelbourn\n# josephperrott\n#\n# =========================================================\n# @angular/framework-global-approvers-for-docs-only-changes\n# =========================================================\n# Used for approving minor documentation-only changes that don't require engineering review.\n\nversion: 3\n\n# availability:\n# users_unavailable: ['atscott']\n\n# Meta field that goes unused by PullApprove to allow for defining aliases to be\n# used throughout the config.\nmeta:\n # Note: Because all inactive groups start as pending, we are only checking pending and rejected active groups.\n no-groups-above-this-pending: &no-groups-above-this-pending len(groups.active.pending.exclude("required-minimum-review")) == 0\n no-groups-above-this-rejected: &no-groups-above-this-rejected len(groups.active.rejected.exclude("required-minimum-review")) == 0\n\n defaults: &defaults\n reviews:\n # Authors provide their approval implicitly, this approval allows for a reviewer\n # from a group not to need a review specifically for an area of the repository\n # they own. This is coupled with the `required-minimum-review` group which requires\n # that all PRs are reviewed by at least one team member who is not the author of\n # the PR.\n author_value: 1\n\n# turn on 'draft' support\n# https://docs.pullapprove.com/config/github-api-version/\n# https://developer.github.com/v3/previews/#draft-pull-requests\ngithub_api_version: 'shadow-cat-preview'\n\n# https://docs.pullapprove.com/config/overrides/\n# Note that overrides are processed in order.\noverrides:\n # For PRs which are still being worked on, either still in draft mode or indicated through WIP in\n # title or label, PullApprove stays in a pending state until its ready for review.\n - if: "draft or 'WIP' in title or 'PR state: WIP' in labels"\n status: pending\n explanation: 'Waiting to send reviews as PR is WIP'\n # Disable PullApprove on specific PRs by adding the `PullApprove: disable` label\n - if: "'PullApprove: disable' in labels"\n status: success\n explanation: "PullApprove skipped because of 'PullApprove: disable' label"\n # If no file matching based groups are active, report this pull request as failing. Most likely,\n # the PR author would need to update the PullApprove config, or create new group.\n - if: len(groups.active.exclude("required-minimum-review").exclude("global-*")) == 0 and len(groups.approved.include("global-*")) == 0\n status: failure\n explanation: 'At least one group must match this PR. Please update an existing review group, or create a new group.'\n # If any global dev-infra approval is given the status should be passing.\n - if: len(groups.approved.include("global-dev-infra-approvers")) == 1\n status: success\n explanation: 'Passing as globally approved by dev-infra'\n # If any global docs approval is given the status should be passing.\n - if: len(groups.approved.include("global-docs-approvers")) == 1\n status: success\n explanation: 'Passing as globally approved by docs'\n # If any global approval is given the status should be passing.\n - if: len(groups.approved.include("global-approvers")) == 1\n status: success\n explanation: 'Passing as globally approved by global approvers'\n\ngroups:\n # =========================================================\n # Framework: Compiler\n # =========================================================\n fw-compiler:\n <<: *defaults\n conditions:\n - >\n contains_any_globs(files, [\n 'packages/compiler/**/{*,.*}',\n 'packages/examples/compiler/**/{*,.*}',\n 'packages/compiler-cli/**/{*,.*}',\n 'packages/language-service/**/{*,.*}',\n ])\n reviewers:\n users:\n - alxhub\n - AndrewKushnir\n - atscott\n - crisbeto\n - devversion\n - kirjs\n - JoostK\n - mmalerba\n\n # =========================================================\n # Framework: General (most code in our packages)\n # =========================================================\n fw-general:\n <<: *defaults\n conditions:\n - >\n contains_any_globs(files.exclude('packages/core/primitives/*'), [\n 'contributing-docs/public-api-surface.md',\n 'integration/**/{*,.*}',\n 'modules/**/{*,.*}',\n 'packages/animations/**/{*,.*}',\n 'packages/benchpress/**/{*,.*}',\n 'packages/common/**/{*,.*}',\n 'packages/core/**/{*,.*}',\n 'packages/docs/**/{*,.*}',\n 'packages/elements/**/{*,.*}',\n 'packages/examples/**/{*,.*}',\n 'packages/forms/**/{*,.*}',\n 'packages/localize/**/{*,.*}',\n 'packages/misc/**/{*,.*}',\n 'packages/platform-browser/**/{*,.*}',\n 'packages/platform-browser-dynamic/**/{*,.*}',\n 'packages/platform-server/**/{*,.*}',\n 'packages/ssr/**/{*,.*}',\n 'packages/router/**/{*,.*}',\n 'packages/service-worker/**/{*,.*}',\n 'packages/upgrade/**/{*,.*}',\n ])\n reviewers:\n users:\n - alxhub\n - AndrewKushnir\n - atscott\n - crisbeto\n - devversion\n - kirjs\n - ~jelbourn\n - thePunderWoman\n - pkozlowski-opensource\n - mmalerba\n\n # =========================================================\n # Framework: Security-sensitive files which require extra review\n # =========================================================\n fw-security:\n <<: *defaults\n conditions:\n - >\n contains_any_globs(files, [\n 'packages/core/src/sanitization/**/{*,.*}',\n 'packages/core/test/linker/security_integration_spec.ts',\n 'packages/compiler/src/schema/**/{*,.*}',\n 'packages/platform-browser/src/security/**/{*,.*}',\n 'packages/tsconfig-tsec-base.json',\n 'packages/tsec-exemption.json',\n 'tools/tsec.bzl',\n 'adev/src/content/guide/security.md',\n 'adev/src/content/examples/security/**/{*,.*}',\n ])\n reviewers:\n users:\n - alxhub\n - jelbourn\n - josephperrott\n - pkozlowski-opensource\n reviews:\n request: -1 # request reviews from everyone\n required: 2 # require at least 2 approvals\n reviewed_for: required\n\n # =========================================================\n # Bazel\n # =========================================================\n bazel:\n <<: *defaults\n conditions:\n - >\n contains_any_globs(files, [\n 'packages/bazel/**/{*,.*}',\n ])\n reviewers:\n users:\n - devversion\n - josephperrott\n\n # =========================================================\n # zone.js\n # =========================================================\n zone-js:\n <<: *defaults\n conditions:\n - >\n contains_any_globs(files.exclude('yarn.lock'), [\n 'packages/zone.js/**/{*,.*}',\n ])\n reviewers:\n users:\n - JiaLiPassion\n\n # =========================================================\n # Tooling: Compiler API shared with Angular CLI\n #\n # Changing this API might break Angular CLI, so we require\n # the CLI team to approve changes here.\n # =========================================================\n tooling-cli-shared-api:\n conditions:\n - >\n contains_any_globs(files, [\n 'packages/compiler-cli/private/tooling.ts',\n 'packages/localize/tools/index.ts'\n ])\n reviewers:\n users:\n - alan-agius4\n - clydin\n - dgp1130\n reviews:\n request: -1 # request reviews from everyone\n required: 2 # require at least 2 approvals\n reviewed_for: required\n\n # =========================================================\n # Documentation content\n # =========================================================\n angular-dev:\n <<: *defaults\n conditions:\n - >\n contains_any_globs(files, [\n 'adev/**/{*,.*}',\n ])\n reviewers:\n users:\n - alan-agius4\n - alxhub\n - AndrewKushnir\n - atscott\n - bencodezen\n - crisbeto\n - kirjs\n - ~JeanMeche\n - jelbourn\n - thePunderWoman\n - devversion\n - josephperrott\n - pkozlowski-opensource\n - mgechev\n - MarkTechson\n - kirjs\n - mmalerba\n - ~hawkgs\n\n # =========================================================\n # Angular DevTools\n # =========================================================\n devtools:\n <<: *defaults\n conditions:\n - >\n contains_any_globs(files, [\n 'devtools/**/{*,.*}',\n ])\n reviewers:\n users:\n - AleksanderBodurri\n - devversion\n - dgp1130\n - josephperrott\n - mgechev\n - MarkTechson\n - ~JeanMeche\n\n # =========================================================\n # Dev-infra\n # =========================================================\n dev-infra:\n <<: *defaults\n conditions:\n - >\n contains_any_globs(files.exclude('.pullapprove.yml'), [\n '{*,.*}',\n '.devcontainer/**/{*,.*}',\n '.github/**/{*,.*}',\n '.husky/**/{*,.*}',\n '.ng-dev/**/{*,.*}',\n '.vscode/**/{*,.*}',\n '.yarn/**/{*,.*}',\n 'contributing-docs/*.md',\n 'contributing-docs/images/**/{*,.*}',\n 'goldens/{*,.*}',\n 'goldens/public-api/manage.js',\n 'modules/{*,.*}',\n 'packages/{*,.*}',\n 'packages/examples/test-utils/**/{*,.*}',\n 'packages/private/**/{*,.*}',\n 'packages/examples/{*,.*}',\n 'scripts/**/{*,.*}',\n 'third_party/**/{*,.*}',\n 'tools/bazel-repo-patches/**/{*,.*}',\n 'tools/circular_dependency_test/**/{*,.*}',\n 'tools/contributing-stats/**/{*,.*}',\n 'tools/esm-interop/**/{*,.*}',\n 'tools/gulp-tasks/**/{*,.*}',\n 'tools/legacy-saucelabs/**/{*,.*}',\n 'tools/manual_api_docs/**/{*,.*}',\n 'tools/npm-patches/**/{*,.*}',\n 'tools/rxjs/**/{*,.*}',\n 'tools/saucelabs-daemon/**/{*,.*}',\n 'tools/saucelabs/**/{*,.*}',\n 'tools/symbol-extractor/**/{*,.*}',\n 'tools/testing/**/{*,.*}',\n 'tools/tslint/**/{*,.*}',\n 'tools/utils/**/{*,.*}',\n 'tools/yarn/**/{*,.*}',\n 'tools/{*,.*}',\n '**/*.bzl'\n ])\n - author not in ["angular-robot"]\n reviewers:\n users:\n - devversion\n - josephperrott\n\n # =========================================================\n # Renovate Changes\n # =========================================================\n renovate-changes:\n <<: *defaults\n conditions:\n - author in ["angular-robot"]\n reviewers:\n users:\n - ~alan-agius4\n teams:\n - framework-team\n\n # =========================================================\n # Public API\n # =========================================================\n public-api:\n <<: *defaults\n conditions:\n - *no-groups-above-this-pending\n - *no-groups-above-this-rejected\n - >\n contains_any_globs(files.exclude("goldens/public-api/manage.js"), [\n 'goldens/public-api/**/{*,.*}',\n ])\n reviewers:\n users:\n - AndrewKushnir\n - alxhub\n - atscott\n - ~jelbourn\n - thePunderWoman\n - pkozlowski-opensource\n - kirjs\n - mmalerba\n - crisbeto\n - devversion\n - ~iteriani\n - ~tbondwilkinson\n - ~rahatarmanahmed\n reviews:\n request: 3 # Request reviews from 3 people\n required: 2 # Require that 2 people approve\n reviewed_for: required\n\n # ================================================\n # Size tracking\n # ================================================\n size-tracking:\n <<: *defaults\n conditions:\n - *no-groups-above-this-pending\n - *no-groups-above-this-rejected\n - >\n contains_any_globs(files, [\n 'goldens/size-tracking/**/{*,.*}'\n ])\n reviewers:\n users:\n - alxhub\n - AndrewKushnir\n - atscott\n - kirjs\n - ~jelbourn\n - thePunderWoman\n - pkozlowski-opensource\n - mmalerba\n reviews:\n request: 2 # Request reviews from 2 people\n required: 1 # Require that 1 person approve\n reviewed_for: required\n\n ####################################################################################\n # Special Cases\n ####################################################################################\n\n # =========================================================\n # Code Ownership\n # =========================================================\n code-ownership:\n <<: *defaults\n conditions:\n - >\n contains_any_globs(files, [\n '.pullapprove.yml'\n ])\n reviewers:\n users:\n - alxhub\n - AndrewKushnir\n - andrewseguin\n - dgp1130\n - ~jelbourn\n - thePunderWoman\n - josephperrott\n\n # =========================================================\n # Primitives\n # =========================================================\n\n # Angular team required reviews\n primitives:\n <<: *defaults\n conditions:\n - >\n contains_any_globs(files, [\n 'packages/core/primitives/**/{*,.*}',\n ])\n reviewers:\n users:\n - pkozlowski-opensource # Pawel Kozlowski\n - alxhub # Alex Rickabaugh\n - thePunderWoman # Jessica Janiuk\n - AndrewKushnir # Andrew Kushnir\n - atscott # Andrew Scott\n labels:\n pending: 'requires: TGP'\n approved: 'requires: TGP'\n rejected: 'requires: TGP'\n\n # External team required reviews\n primitives-shared:\n <<: *defaults\n conditions:\n - >\n contains_any_globs(files, [\n 'packages/core/primitives/**/{*,.*}',\n ])\n reviewers:\n users:\n - csmick # Cameron Smick\n - mturco # Matt Turco\n - iteriani # Thomas Nguyen\n - tbondwilkinson # Tom Wilkinson\n - rahatarmanahmed # Rahat Ahmed\n - ENAML # Ethan Cline\n labels:\n pending: 'requires: TGP'\n approved: 'requires: TGP'\n rejected: 'requires: TGP'\n\n ####################################################################################\n # Override managed result groups\n #\n # Groups which are only used to determine the value of an override are managed at\n # the bottom of the list as they will set a status on the PR directly, they\n # therefore can always be process last without concern.\n ####################################################################################\n\n # =========================================================\n # Global Approvers\n #\n # All reviews performed for global approvals require using\n # the `Reviewed-for:` specifier to set the approval\n # specificity as documented at:\n # https://docs.pullapprove.com/reviewed-for/\n # =========================================================\n global-approvers:\n type: optional\n reviewers:\n teams:\n - framework-global-approvers\n reviews:\n request: 0\n required: 1\n reviewed_for: required\n\n # =========================================================\n # Global Approvers For Docs\n #\n # All reviews performed for global docs approvals require\n # using the `Reviewed-for:` specifier to set the approval\n # specificity as documented at:\n # https://docs.pullapprove.com/reviewed-for/\n # =========================================================\n global-docs-approvers:\n type: optional\n reviewers:\n teams:\n - framework-global-approvers-for-docs-only-changes\n reviews:\n request: 0\n required: 1\n reviewed_for: required\n\n # =========================================================\n # Global Approvers For Dev-Infra changes\n #\n # All reviews performed for global dev-infra approvals\n # require using the `Reviewed-for:` specifier to set the\n # approval specificity as documented at:\n # https://docs.pullapprove.com/reviewed-for/\n # =========================================================\n global-dev-infra-approvers:\n type: optional\n reviewers:\n teams:\n - dev-infra-framework\n reviews:\n request: 0\n required: 1\n reviewed_for: required\n\n # =========================================================\n # Require review on all PRs\n #\n # All PRs require at least one review. This rule will not\n # request any reviewers, however will require that at least\n # one review is provided before the group is satisfied.\n # =========================================================\n required-minimum-review:\n reviews:\n request: 0 # Do not request any reviews from the reviewer group\n required: 1 # Require that all PRs have approval from at least one of the users in the group\n author_value: 0 # The author of the PR cannot provide an approval for themself\n reviewed_for: ignored # All reviews apply to this group whether noted via Reviewed-for or not\n reviewers:\n teams:\n # Any member of the team can provide a review to perform the minimum review required\n - team\n | dataset_sample\yaml\javascript\.pullapprove.yml | .pullapprove.yml | YAML | 20,213 | 0.95 | 0.053265 | 0.298913 | awesome-app | 743 | 2025-01-04T12:47:57.272717 | BSD-3-Clause | false | 49e2bfff14b68bded45998a27ba564cc |
ci:\n platform:\n npm:\n # Don't run unmainted test on nomnom, it's only used by build tools, not in MUI.\n nomnom:\n tests:\n unmaintained: skip\n # Don't run vulnerabity test on os-locale, it's only used by yargs in build tools, not in MUI.\n os-locale:\n tests:\n vulnerable: skip\n | dataset_sample\yaml\javascript\.tidelift.yml | .tidelift.yml | YAML | 330 | 0.8 | 0 | 0.181818 | react-lib | 81 | 2025-03-07T22:22:45.942294 | Apache-2.0 | false | f46f275966103a231fa204bddd4c3903 |
# Configuration for angular-robot\n\n# options for the size plugin\nsize:\n disabled: true\n maxSizeIncrease: 2000\n circleCiStatusName: 'ci/circleci: test'\n\n# options for the merge plugin\nmerge:\n # the status will be added to your pull requests\n status:\n # set to true to disable\n disabled: false\n # the name of the status\n context: 'ci/angular: merge status'\n # text to show when all checks pass\n successText: 'All checks passed!'\n # text to show when some checks are failing\n failureText: 'The following checks are failing:'\n\n # the g3 status will be added to your pull requests if they include files that match the patterns\n g3Status:\n # set to true to disable\n disabled: true\n # the name of the status\n context: 'google3'\n # text to show when the status is pending, {{PRNumber}} will be replaced by the PR number\n pendingDesc: 'Googler: run g3sync presubmit {{PRNumber}}'\n # text to show when the status is success\n successDesc: 'Does not affect google3'\n # link to use for the details\n url: 'http://go/angular/g3sync'\n # list of patterns to check for the files changed by the PR\n include:\n exclude:\n\n # comment that will be added to a PR when there is a conflict, leave empty or set to false to disable\n mergeConflictComment: "Hi @{{PRAuthor}}! This PR has merge conflicts due to recent upstream merges.\nPlease help to unblock it by resolving these conflicts. Thanks!"\n\n # label to monitor\n mergeLabel: 'action: merge'\n\n # adding any of these labels will also add the merge label\n mergeLinkedLabels:\n - 'action: merge-assistance'\n\n # list of checks that will determine if the merge label can be added\n checks:\n # require that the PR has reviews from all requested reviewers\n #\n # This enables us to request reviews from both eng and tech writers, or multiple eng folks, and prevents accidental merges.\n # Rather than merging PRs with pending reviews, if all approvals are obtained and additional reviews are not needed, any pending reviewers should be removed via GitHub UI (this also leaves an audit trail behind these decisions).\n requireReviews: true,\n\n # whether the PR shouldn't have a conflict with the base branch\n noConflict: true\n # list of labels that a PR needs to have, checked with a regexp.\n requiredLabels:\n - 'target: *'\n\n # list of labels that a PR shouldn't have, checked after the required labels with a regexp\n forbiddenLabels:\n - 'target: TBD'\n - 'action: cleanup'\n - 'action: review'\n - 'state: blocked'\n\n # list of PR statuses that need to be successful\n requiredStatuses:\n - 'google-internal-tests'\n - 'pullapprove'\n\n # the comment that will be added when the merge label is added despite failing checks, leave empty or set to false to disable\n # {{MERGE_LABEL}} will be replaced by the value of the mergeLabel option\n # {{PLACEHOLDER}} will be replaced by the list of failing checks\n mergeRemovedComment: "I see that you just added the `{{MERGE_LABEL}}` label, but the following checks are still failing:\n{{PLACEHOLDER}}\n\n**If you want your PR to be merged, it has to pass all the CI checks.**\n\nIf you can't get the PR to a green state due to flakes or broken `main`, please try rebasing to `main` and/or restarting the CI job. If that fails and you believe that the issue is not due to your change, please contact the caretaker and ask for help."\n\n# options for the triage plugin\ntriage:\n # number of the milestone to apply when the issue has not been triaged yet\n needsTriageMilestone: 83,\n # number of the milestone to apply when the issue is triaged\n defaultMilestone: 82,\n # arrays of labels that determine if an issue has been triaged by the caretaker\n l1TriageLabels:\n - - 'area: *'\n # arrays of labels that determine if an issue has been fully triaged\n l2TriageLabels:\n - - 'P0'\n - 'area: *'\n - - 'P1'\n - 'area: *'\n - - 'P2'\n - 'area: *'\n - - 'P3'\n - 'area: *'\n - - 'P4'\n - 'area: *'\n - - 'P5'\n - 'area: *'\n - - 'feature'\n - 'area: *'\n - - 'discussion'\n - 'area: *'\n - - 'needs clarification'\n - 'area: *'\n - - 'needs reproduction'\n - 'area: *'\n\n# options for the triage PR plugin\ntriagePR:\n # set to true to disable\n disabled: false\n # number of the milestone to apply when the PR has not been triaged yet\n needsTriageMilestone: 83,\n # number of the milestone to apply when the PR is triaged\n defaultMilestone: 82,\n # arrays of labels that determine if a PR has been triaged by the caretaker\n l1TriageLabels:\n - - 'area: *'\n # arrays of labels that determine if a PR has been fully triaged\n l2TriageLabels:\n - - 'area: *'\n\n# options for rerunning CI\nrerunCircleCI:\n # set to true to disable\n disabled: false\n # the label which when added triggers a rerun of the default CircleCI workflow\n triggerRerunLabel: 'action: rerun CI at HEAD'\n | dataset_sample\yaml\javascript\angular-robot.yml | angular-robot.yml | YAML | 4,945 | 0.95 | 0.129771 | 0.376068 | python-kit | 374 | 2025-03-05T14:40:38.199103 | Apache-2.0 | false | e0af91a796f7c4bb525508cc60910337 |
coverage:\n status:\n project:\n default:\n target: auto\n threshold: 1%\n patch: off\ncomment: false\n | dataset_sample\yaml\javascript\codecov.yml | codecov.yml | YAML | 121 | 0.7 | 0 | 0 | react-lib | 329 | 2023-07-21T06:22:29.512680 | BSD-3-Clause | false | 8f15d7e0c99e1ceb09e18dd5aa391ec9 |
version: 2.1\norbs:\n aws-cli: circleci/aws-cli@5.3.1\n aws-s3: circleci/aws-s3@4.1.1\n\nparameters:\n browserstack-force:\n description: Whether to force browserstack usage. We have limited resources on browserstack so the pipeline might decide to skip browserstack if this parameter isn't set to true.\n type: boolean\n default: false\n react-version:\n description: The version of react to be used\n type: string\n default: stable\n workflow:\n description: The name of the workflow to run\n type: string\n default: pipeline\n e2e-base-url:\n description: The base url for running end-to-end test\n type: string\n default: ''\n\ndefault-job: &default-job\n parameters:\n react-version:\n description: The version of react to be used\n type: string\n default: << pipeline.parameters.react-version >>\n test-gate:\n description: A particular type of tests that should be run\n type: string\n default: undefined\n e2e-base-url:\n description: The base url for running end-to-end test\n type: string\n default: << pipeline.parameters.e2e-base-url >>\n environment:\n # expose it globally otherwise we have to thread it from each job to the install command\n BROWSERSTACK_FORCE: << pipeline.parameters.browserstack-force >>\n REACT_VERSION: << parameters.react-version >>\n TEST_GATE: << parameters.test-gate >>\n AWS_REGION_ARTIFACTS: eu-central-1\n COREPACK_ENABLE_DOWNLOAD_PROMPT: '0'\n DANGER_DISABLE_TRANSPILATION: 'true'\n working_directory: /tmp/material-ui\n docker:\n - image: cimg/node:20.17\n\ndefault-context: &default-context\n context:\n - org-global\n\n# CircleCI has disabled the cache across forks for security reasons.\n# Following their official statement, it was a quick solution, they\n# are working on providing this feature back with appropriate security measures.\n# https://discuss.circleci.com/t/saving-cache-stopped-working-warning-skipping-this-step-disabled-in-configuration/24423/21\n#\n# restore_repo: &restore_repo\n# restore_cache:\n# key: v1-repo-{{ .Branch }}-{{ .Revision }}\n\ncommands:\n setup_corepack:\n parameters:\n browsers:\n type: boolean\n default: false\n description: 'Set to true if you intend to any browser (for example with playwright).'\n steps:\n - run:\n name: Set npm registry public signing keys\n command: |\n echo "export COREPACK_INTEGRITY_KEYS='$(curl https://registry.npmjs.org/-/npm/v1/keys | jq -c '{npm: .keys}')'" >> $BASH_ENV\n\n - when:\n condition: << parameters.browsers >>\n steps:\n - run:\n name: Install pnpm package manager\n command: corepack enable\n - when:\n condition:\n not: << parameters.browsers >>\n steps:\n - run:\n name: Install pnpm package manager\n # See https://stackoverflow.com/a/73411601\n command: corepack enable --install-directory ~/bin\n\n - run:\n name: View install environment\n command: |\n node --version\n pnpm --version\n\n install_js:\n parameters:\n browsers:\n type: boolean\n default: false\n description: 'Set to true if you intend to any browser (for example with playwright).'\n\n steps:\n - setup_corepack:\n browsers: << parameters.browsers >>\n\n - run:\n name: Resolve React version\n command: |\n pnpm use-react-version\n # log a patch for maintainers who want to check out this change\n git --no-pager diff HEAD\n\n - run:\n name: Install js dependencies\n command: pnpm install\n\njobs:\n checkout:\n <<: *default-job\n steps:\n - checkout\n - install_js\n - when:\n # Install can be "dirty" when running with non-default versions of React\n condition:\n equal: [<< parameters.react-version >>, stable]\n steps:\n - run:\n name: Should not have any git not staged\n command: git add -A && git diff --exit-code --staged\n - run:\n name: '`pnpm dedupe` was run?'\n command: |\n # #target-branch-reference\n if [[ $(git diff --name-status master | grep -E 'pnpm-workspace\.yaml|pnpm-lock.yaml|package\.json') == "" ]];\n then\n echo "No changes to dependencies detected. Skipping..."\n else\n pnpm dedupe --check\n fi\n test_unit:\n <<: *default-job\n steps:\n - checkout\n - install_js\n - run:\n name: Tests fake browser\n command: pnpm test:coverage:ci\n - run:\n name: Check coverage generated\n command: |\n if ! [[ -s coverage/lcov.info ]]\n then\n exit 1\n fi\n - run:\n name: internal-scripts\n command: |\n # latest commit\n LATEST_COMMIT=$(git rev-parse HEAD)\n\n # latest commit where internal-scripts was changed\n FOLDER_COMMIT=$(git log -1 --format=format:%H --full-diff packages-internal/scripts)\n\n if [ $FOLDER_COMMIT = $LATEST_COMMIT ]; then\n echo "changes, let's run the tests"\n pnpm --filter @mui/internal-scripts test\n else\n echo "no changes"\n fi\n - run:\n name: Coverage\n command: |\n curl -Os https://uploader.codecov.io/latest/linux/codecov\n chmod +x codecov\n ./codecov -t ${CODECOV_TOKEN} -Z -F "$REACT_VERSION-jsdom"\n test_lint:\n <<: *default-job\n steps:\n - checkout\n - install_js\n - run:\n name: Eslint\n command: pnpm eslint:ci\n - run:\n name: Stylelint\n command: pnpm stylelint\n - run:\n name: Lint JSON\n command: pnpm jsonlint\n - run:\n name: Lint Markdown\n command: pnpm markdownlint\n - run:\n # See https://circleci.com/developer/orbs/orb/circleci/vale as reference\n name: Install Vale\n command: |\n #!/bin/bash\n VALE_STR_CLI_VERSION=3.3.0\n\n # set smart sudo\n if [[ $EUID -eq 0 ]]; then export SUDO=""; else export SUDO="sudo"; fi\n\n mkdir /tmp/vale-extract\n cd /tmp/vale-extract\n GZIPPED_OUTPUT="vale.tar.gz"\n BINARY_URL=https://github.com/errata-ai/vale/releases/download/v${VALE_STR_CLI_VERSION}/vale_${VALE_STR_CLI_VERSION}_Linux_64-bit.tar.gz\n curl -sSL "$BINARY_URL" -o "${GZIPPED_OUTPUT}"\n\n if [ ! -s "${GZIPPED_OUTPUT}" ]; then\n echo "Downloaded file is empty"\n rm "${GZIPPED_OUTPUT}"\n exit 1\n fi\n\n tar -xzf "${GZIPPED_OUTPUT}"\n $SUDO mv vale /usr/local/bin\n rm "${GZIPPED_OUTPUT}"\n\n # validate installation\n if [[ -z "$(command -v vale)" ]]; then\n echo "vale installation failed"\n exit 1\n else\n echo "vale installation successful"\n vale --version\n exit 0\n fi\n - run:\n name: Lint writing style\n command: |\n vale sync\n pnpm valelint\n test_static:\n <<: *default-job\n steps:\n - checkout\n - install_js\n - run:\n name: '`pnpm prettier` changes committed?'\n command: pnpm prettier --check\n - run:\n name: Generate PropTypes\n command: pnpm proptypes\n - run:\n name: '`pnpm proptypes` changes committed?'\n command: git add -A && git diff --exit-code --staged\n - run:\n name: Generate the documentation\n command: pnpm docs:api\n - run:\n name: '`pnpm docs:api` changes committed?'\n command: git add -A && git diff --exit-code --staged\n - run:\n name: Update the navigation translations\n command: pnpm docs:i18n\n - run:\n name: '`pnpm docs:i18n` changes committed?'\n command: git add -A && git diff --exit-code --staged\n - run:\n name: '`pnpm extract-error-codes` changes committed?'\n command: |\n pnpm extract-error-codes\n git add -A && git diff --exit-code --staged\n - run:\n name: '`pnpm docs:link-check` changes committed?'\n command: |\n pnpm docs:link-check\n git add -A && git diff --exit-code --staged\n test_types:\n <<: *default-job\n resource_class: 'medium+'\n steps:\n - checkout\n - install_js\n - run:\n name: Transpile TypeScript demos\n command: pnpm docs:typescript:formatted\n - run:\n name: '`pnpm docs:typescript:formatted` changes committed?'\n command: git add -A && git diff --exit-code --staged\n - run:\n name: Tests TypeScript definitions\n command: pnpm typescript:ci\n environment:\n NODE_OPTIONS: --max-old-space-size=3072\n - run:\n name: Test module augmentation\n command: |\n pnpm --filter @mui/material typescript:module-augmentation\n pnpm --filter @mui/joy typescript:module-augmentation\n pnpm --filter @mui/system typescript:module-augmentation\n - run:\n name: Diff declaration files\n command: |\n git add -f packages/mui-material/build || echo '/material declarations do not exist'\n git add -f packages/mui-lab/build || echo '/lab declarations do not exist'\n git add -f packages/mui-utils/build || echo '/utils declarations do not exist'\n pnpm -r build:stable && pnpm -r build:types \n git --no-pager diff\n - run:\n name: Any defect declaration files?\n command: node scripts/testBuiltTypes.mjs\n - save_cache:\n name: Save generated declaration files\n key: typescript-declaration-files-{{ .Branch }}-{{ .Revision }}\n paths:\n # packages with generated declaration files\n - packages/mui-material/build\n - packages/mui-lab/build\n - packages/mui-utils/build\n test_types_next:\n <<: *default-job\n resource_class: 'medium+'\n steps:\n - checkout\n - install_js\n - run:\n name: Resolve typescript version\n command: |\n pnpm update -r typescript@next\n # log a patch for maintainers who want to check out this change\n git --no-pager diff HEAD\n - run:\n name: Tests TypeScript definitions\n command: |\n # ignore build failures\n # it's expected that typescript@next fails since the lines of the errors\n # change frequently. This build is monitored regardless of its status\n set +e\n pnpm typescript:ci\n exit 0\n\n - restore_cache:\n name: Restore generated declaration files\n keys:\n # We assume that the target branch is `next` and that declaration files are persisted in commit order.\n # "If there are multiple matches, the most recently generated cache will be used."\n - typescript-declaration-files-next\n\n - run:\n name: Diff declaration files\n command: |\n git add -f packages/mui-material/build || echo '/core declarations do not exist'\n git add -f packages/mui-lab/build || echo '/lab declarations do not exist'\n git add -f packages/mui-utils/build || echo '/utils declarations do not exist'\n pnpm -r build:types\n git --no-pager diff\n\n - run:\n name: Log defect declaration files\n command: |\n # ignore build failures\n # Fixing these takes some effort that isn't viable to merge in a single PR.\n # We'll simply monitor them for now.\n set +e\n node scripts/testBuiltTypes.mjs\n exit 0\n test_browser:\n <<: *default-job\n resource_class: 'medium+'\n docker:\n - image: mcr.microsoft.com/playwright:v1.51.1-noble\n steps:\n - checkout\n - install_js:\n browsers: true\n - run:\n name: Tests real browsers\n command: pnpm test:karma\n - run:\n name: Check coverage generated\n command: |\n if ! [[ -s coverage/lcov.info ]]\n then\n exit 1\n fi\n - run:\n name: Coverage\n command: |\n curl -Os https://uploader.codecov.io/latest/linux/codecov\n chmod +x codecov\n ./codecov -t ${CODECOV_TOKEN} -Z -F "$REACT_VERSION-browser"\n - store_artifacts:\n # hardcoded in karma-webpack\n path: /tmp/_karma_webpack_\n destination: artifact-file\n test_e2e:\n <<: *default-job\n docker:\n - image: mcr.microsoft.com/playwright:v1.51.1-noble\n steps:\n - checkout\n - install_js:\n browsers: true\n - run:\n name: pnpm test:e2e\n command: pnpm test:e2e\n test_e2e_website:\n # NOTE: This workflow runs after successful docs deploy. See /test/e2e-website/README.md#ci\n <<: *default-job\n docker:\n - image: mcr.microsoft.com/playwright:v1.51.1-noble\n steps:\n - checkout\n - install_js:\n browsers: true\n - run:\n name: pnpm test:e2e-website\n command: pnpm test:e2e-website\n environment:\n PLAYWRIGHT_TEST_BASE_URL: << parameters.e2e-base-url >>\n test_profile:\n <<: *default-job\n docker:\n - image: mcr.microsoft.com/playwright:v1.51.1-noble\n steps:\n - checkout\n - install_js:\n browsers: true\n - run:\n name: Tests real browsers\n # Run a couple of times for a better sample.\n # TODO: hack something together where we can compile once and run multiple times e.g. by abusing watchmode.\n command: |\n # Running on chrome only since actual times are innaccurate anyway\n # The other reason is that browserstack allows little concurrency so it's likely that we're starving other runs.\n pnpm test:karma:profile --browsers chrome,chromeHeadless\n pnpm test:karma:profile --browsers chrome,chromeHeadless\n pnpm test:karma:profile --browsers chrome,chromeHeadless\n pnpm test:karma:profile --browsers chrome,chromeHeadless\n pnpm test:karma:profile --browsers chrome,chromeHeadless\n # Persist reports for inspection in https://mui-dashboard.netlify.app/\n - store_artifacts:\n # see karma.conf.profile.js reactProfilerReporter.outputDir\n path: tmp/react-profiler-report/karma\n destination: react-profiler-report/karma\n test_regressions:\n <<: *default-job\n docker:\n - image: mcr.microsoft.com/playwright:v1.51.1-noble\n steps:\n - checkout\n - install_js:\n browsers: true\n - run:\n name: Run visual regression tests\n command: xvfb-run pnpm test:regressions\n - run:\n name: Build packages for fixtures\n command: pnpm release:build\n - run:\n name: Analyze exported typescript\n command: pnpm test:attw\n - run:\n name: test exported typescript\n command: pnpm --filter @mui-internal/test-module-resolution typescript:all\n - run:\n name: Run visual regression tests using Pigment CSS\n command: xvfb-run pnpm test:regressions-pigment-css\n - run:\n name: Upload screenshots to Argos CI\n command: pnpm test:argos\n test_bundling_prepare:\n <<: *default-job\n steps:\n - checkout\n - install_js\n - run:\n name: Build packages for fixtures\n command: pnpm lerna run --scope "@mui/*" build\n - run:\n name: Pack packages\n command: pnpm release:pack\n - persist_to_workspace:\n root: packed\n paths:\n - '*'\n\n test_bundling_node_cjs:\n <<: *default-job\n working_directory: /tmp/material-ui/test/bundling/fixtures/node-cjs/\n steps:\n - checkout:\n path: /tmp/material-ui\n - attach_workspace:\n at: /tmp/material-ui/packed\n - setup_corepack\n - run:\n name: Install dependencies\n command: pnpm install --ignore-workspace\n - run:\n name: Test fixture\n command: pnpm start\n test_bundling_node_esm:\n <<: *default-job\n working_directory: /tmp/material-ui/test/bundling/fixtures/node-esm/\n steps:\n - checkout:\n path: /tmp/material-ui\n - attach_workspace:\n at: /tmp/material-ui/packed\n - setup_corepack\n - run:\n name: Install dependencies\n command: pnpm install --ignore-workspace\n - run:\n name: Test fixture\n # TODO: Known failure\n command: pnpm start\n test_bundling_next_webpack4:\n <<: *default-job\n docker:\n - image: mcr.microsoft.com/playwright:v1.51.1-noble\n working_directory: /tmp/material-ui/test/bundling/fixtures/next-webpack4/\n steps:\n - checkout:\n path: /tmp/material-ui\n - attach_workspace:\n at: /tmp/material-ui/packed\n - setup_corepack:\n browsers: true\n - run:\n name: Install dependencies\n command: pnpm install --ignore-workspace\n - run:\n name: Test fixture\n command: pnpm start\n test_bundling_next_webpack5:\n <<: *default-job\n docker:\n - image: mcr.microsoft.com/playwright:v1.51.1-noble\n working_directory: /tmp/material-ui/test/bundling/fixtures/next-webpack5/\n steps:\n - checkout:\n path: /tmp/material-ui\n - attach_workspace:\n at: /tmp/material-ui/packed\n - setup_corepack:\n browsers: true\n - run:\n name: Install dependencies\n command: pnpm install --ignore-workspace\n - run:\n name: Test fixture\n command: pnpm start\n test_bundling_create_react_app:\n <<: *default-job\n docker:\n - image: mcr.microsoft.com/playwright:v1.51.1-noble\n working_directory: /tmp/material-ui/test/bundling/fixtures/create-react-app/\n steps:\n - checkout:\n path: /tmp/material-ui\n - attach_workspace:\n at: /tmp/material-ui/packed\n - setup_corepack:\n browsers: true\n - run:\n name: Install dependencies\n command: pnpm install --ignore-workspace\n - run:\n name: Test fixture\n command: pnpm start\n test_bundling_snowpack:\n <<: *default-job\n docker:\n - image: mcr.microsoft.com/playwright:v1.51.1-noble\n working_directory: /tmp/material-ui/test/bundling/fixtures/snowpack/\n steps:\n - checkout:\n path: /tmp/material-ui\n - attach_workspace:\n at: /tmp/material-ui/packed\n - setup_corepack:\n browsers: true\n - run:\n name: Install dependencies\n command: pnpm install --ignore-workspace\n - run:\n name: Test fixture\n command: pnpm start\n test_bundling_vite:\n <<: *default-job\n docker:\n - image: mcr.microsoft.com/playwright:v1.51.1-noble\n working_directory: /tmp/material-ui/test/bundling/fixtures/vite/\n steps:\n - checkout:\n path: /tmp/material-ui\n - attach_workspace:\n at: /tmp/material-ui/packed\n - setup_corepack:\n browsers: true\n - run:\n name: Install dependencies\n command: pnpm install --ignore-workspace\n - run:\n name: Test fixture\n command: pnpm start\n test_bundling_esbuild:\n <<: *default-job\n docker:\n - image: mcr.microsoft.com/playwright:v1.51.1-noble\n working_directory: /tmp/material-ui/test/bundling/fixtures/esbuild/\n steps:\n - checkout:\n path: /tmp/material-ui\n - attach_workspace:\n at: /tmp/material-ui/packed\n - setup_corepack:\n browsers: true\n - run:\n name: Install dependencies\n command: pnpm install --ignore-workspace\n - run:\n name: Test fixture\n command: pnpm start\n test_bundling_gatsby:\n <<: *default-job\n docker:\n - image: mcr.microsoft.com/playwright:v1.51.1-noble\n environment:\n GATSBY_CPU_COUNT: '3'\n working_directory: /tmp/material-ui/test/bundling/fixtures/gatsby/\n steps:\n - checkout:\n path: /tmp/material-ui\n - attach_workspace:\n at: /tmp/material-ui/packed\n - setup_corepack:\n browsers: true\n - run:\n name: Install dependencies\n command: pnpm install --ignore-workspace\n - run:\n name: Test fixture\n command: pnpm start\n\n test_bundle_size_monitor:\n <<: *default-job\n steps:\n - checkout\n - install_js\n - run:\n name: prepare danger on PRs\n command: pnpm danger ci\n environment:\n DANGER_COMMAND: prepareBundleSizeReport\n - setup_corepack\n - run:\n name: build @mui packages\n command: pnpm lerna run --ignore @mui/icons-material --concurrency 6 --scope "@mui/*" build\n - run:\n name: create @mui/material canary distributable\n command: |\n cd packages/mui-material/build\n npm version 0.0.0-canary.${CIRCLE_SHA1} --no-git-tag-version\n npm pack\n mv mui-material-0.0.0-canary.${CIRCLE_SHA1}.tgz ../../../mui-material.tgz\n - when:\n # don't run on PRs\n condition:\n not:\n matches:\n # "^pull/\d+" is not valid YAML\n # "^pull/\\d+" matches neither 'pull/1' nor 'main'\n # Note that we want to include 'pull/1', 'pull/1/head' and ''pull/1/merge'\n pattern: '^pull/.+$'\n value: << pipeline.git.branch >>\n steps:\n - aws-cli/setup:\n aws_access_key_id: $AWS_ACCESS_KEY_ID_ARTIFACTS\n aws_secret_access_key: $AWS_SECRET_ACCESS_KEY_ARTIFACTS\n region: ${AWS_REGION_ARTIFACTS}\n # Upload distributables to S3\n - aws-s3/copy:\n from: mui-material.tgz\n to: s3://mui-org-ci/artifacts/$CIRCLE_BRANCH/$CIRCLE_SHA1/\n - store_artifacts:\n path: mui-material.tgz\n destination: mui-material.tgz\n - run:\n name: create a size snapshot\n command: pnpm size:snapshot\n - store_artifacts:\n name: persist size snapshot as pipeline artifact\n path: size-snapshot.json\n destination: size-snapshot.json\n - when:\n # don't run on PRs\n condition:\n not:\n matches:\n # "^pull/\d+" is not valid YAML\n # "^pull/\\d+" matches neither 'pull/1' nor 'main'\n # Note that we want to include 'pull/1', 'pull/1/head' and ''pull/1/merge'\n pattern: '^pull/.+$'\n value: << pipeline.git.branch >>\n steps:\n - aws-cli/setup:\n aws_access_key_id: $AWS_ACCESS_KEY_ID_ARTIFACTS\n aws_secret_access_key: $AWS_SECRET_ACCESS_KEY_ARTIFACTS\n region: ${AWS_REGION_ARTIFACTS}\n # persist size snapshot on S3\n - aws-s3/copy:\n arguments: --content-type application/json\n from: size-snapshot.json\n to: s3://mui-org-ci/artifacts/$CIRCLE_BRANCH/$CIRCLE_SHA1/\n # symlink size-snapshot to latest\n - aws-s3/copy:\n arguments: --content-type application/json\n from: size-snapshot.json\n to: s3://mui-org-ci/artifacts/$CIRCLE_BRANCH/latest/\n - run:\n name: Run danger on PRs\n command: pnpm danger ci --fail-on-errors\n environment:\n DANGER_COMMAND: reportBundleSize\n test_benchmark:\n <<: *default-job\n docker:\n - image: mcr.microsoft.com/playwright:v1.51.1-noble\n steps:\n - checkout\n - install_js:\n browsers: true\n - run: pnpm benchmark:browser\n - store_artifacts:\n name: Publish benchmark results as a pipeline artifact.\n path: tmp/benchmarks\n destination: benchmarks\nworkflows:\n version: 2\n pipeline:\n when:\n equal: [pipeline, << pipeline.parameters.workflow >>]\n jobs:\n - checkout:\n <<: *default-context\n - test_unit:\n <<: *default-context\n requires:\n - checkout\n - test_lint:\n <<: *default-context\n requires:\n - checkout\n - test_static:\n <<: *default-context\n requires:\n - checkout\n - test_types:\n <<: *default-context\n requires:\n - checkout\n - test_browser:\n <<: *default-context\n requires:\n - checkout\n - test_regressions:\n <<: *default-context\n requires:\n - checkout\n - test_e2e:\n <<: *default-context\n requires:\n - checkout\n - test_bundle_size_monitor:\n <<: *default-context\n requires:\n - checkout\n e2e-website:\n when:\n equal: [e2e-website, << pipeline.parameters.workflow >>]\n jobs:\n - checkout:\n <<: *default-context\n - test_e2e_website:\n <<: *default-context\n requires:\n - checkout\n\n bundling:\n when:\n equal: [bundling, << pipeline.parameters.workflow >>]\n jobs:\n - test_bundling_prepare:\n <<: *default-context\n - test_bundling_node_cjs:\n <<: *default-context\n requires:\n - test_bundling_prepare\n - test_bundling_node_esm:\n <<: *default-context\n requires:\n - test_bundling_prepare\n - test_bundling_create_react_app:\n <<: *default-context\n requires:\n - test_bundling_prepare\n - test_bundling_snowpack:\n <<: *default-context\n requires:\n - test_bundling_prepare\n - test_bundling_vite:\n <<: *default-context\n requires:\n - test_bundling_prepare\n - test_bundling_esbuild:\n <<: *default-context\n requires:\n - test_bundling_prepare\n - test_bundling_gatsby:\n <<: *default-context\n requires:\n - test_bundling_prepare\n - test_bundling_next_webpack4:\n <<: *default-context\n requires:\n - test_bundling_prepare\n - test_bundling_next_webpack5:\n <<: *default-context\n requires:\n - test_bundling_prepare\n\n profile:\n when:\n equal: [profile, << pipeline.parameters.workflow >>]\n jobs:\n - test_profile:\n <<: *default-context\n\n # This workflow can be triggered manually on the PR\n react-17:\n when:\n equal: [react-17, << pipeline.parameters.workflow >>]\n jobs:\n - test_unit:\n <<: *default-context\n react-version: ^17.0.0\n name: test_unit-react@17\n - test_browser:\n <<: *default-context\n react-version: ^17.0.0\n name: test_browser-react@17\n - test_regressions:\n <<: *default-context\n react-version: ^17.0.0\n name: test_regressions-react@17\n - test_e2e:\n <<: *default-context\n react-version: ^17.0.0\n name: test_e2e-react@17\n\n # This workflow is identical to react-17, but scheduled\n # TODO: The v17 tests have deteriorated to the point of no return. Fix for v18 once we\n # deprecate v17, and reenable this workflow.\n # react-17-cron:\n # triggers:\n # - schedule:\n # cron: '0 0 * * *'\n # filters:\n # branches:\n # only:\n # - master\n # - next\n # jobs:\n # - test_unit:\n # <<: *default-context\n # react-version: ^17.0.0\n # name: test_unit-react@17\n # - test_browser:\n # <<: *default-context\n # react-version: ^17.0.0\n # name: test_browser-react@17\n # - test_regressions:\n # <<: *default-context\n # react-version: ^17.0.0\n # name: test_regressions-react@17\n # - test_e2e:\n # <<: *default-context\n # react-version: ^17.0.0\n # name: test_e2e-react@17\n\n # This workflow can be triggered manually on the PR\n react-18:\n when:\n equal: [react-18, << pipeline.parameters.workflow >>]\n jobs:\n - test_unit:\n <<: *default-context\n react-version: ^18.0.0\n name: test_unit-react@18\n - test_browser:\n <<: *default-context\n react-version: ^18.0.0\n name: test_browser-react@18\n - test_regressions:\n <<: *default-context\n react-version: ^18.0.0\n name: test_regressions-react@18\n - test_e2e:\n <<: *default-context\n react-version: ^18.0.0\n name: test_e2e-react@18\n\n # This workflow is identical to react-18, but scheduled\n react-18-cron:\n triggers:\n - schedule:\n cron: '0 0 * * *'\n filters:\n branches:\n only:\n # #target-branch-reference\n - master\n - v5.x\n - v6.x\n jobs:\n - test_unit:\n <<: *default-context\n react-version: ^18.0.0\n name: test_unit-react@18\n - test_browser:\n <<: *default-context\n react-version: ^18.0.0\n name: test_browser-react@18\n - test_regressions:\n <<: *default-context\n react-version: ^18.0.0\n name: test_regressions-react@18\n - test_e2e:\n <<: *default-context\n react-version: ^18.0.0\n name: test_e2e-react@18\n\n # This workflow can be triggered manually on the PR\n react-next:\n when:\n equal: [react-next, << pipeline.parameters.workflow >>]\n jobs:\n - test_unit:\n <<: *default-context\n react-version: next\n name: test_unit-react@next\n - test_browser:\n <<: *default-context\n react-version: next\n name: test_browser-react@next\n - test_regressions:\n <<: *default-context\n react-version: next\n name: test_regressions-react@next\n - test_e2e:\n <<: *default-context\n react-version: next\n name: test_e2e-react@next\n # This workflow is identical to react-next, but scheduled\n react-next-cron:\n triggers:\n - schedule:\n cron: '0 0 * * *'\n filters:\n branches:\n only:\n # #target-branch-reference\n - master\n - v6.x\n jobs:\n - test_unit:\n <<: *default-context\n react-version: next\n name: test_unit-react@next\n - test_browser:\n <<: *default-context\n react-version: next\n name: test_browser-react@next\n - test_regressions:\n <<: *default-context\n react-version: next\n name: test_regressions-react@next\n - test_e2e:\n <<: *default-context\n react-version: next\n name: test_e2e-react@next\n\n typescript-next:\n triggers:\n - schedule:\n cron: '0 0 * * *'\n filters:\n branches:\n only:\n # #target-branch-reference\n - master\n - v6.x\n jobs:\n - test_types_next:\n <<: *default-context\n benchmark:\n when:\n equal: [benchmark, << pipeline.parameters.workflow >>]\n jobs:\n - test_benchmark:\n <<: *default-context\n | dataset_sample\yaml\javascript\config.yml | config.yml | YAML | 31,904 | 0.95 | 0.022817 | 0.088115 | awesome-app | 88 | 2023-09-14T03:52:51.955373 | Apache-2.0 | false | aac8a5e1fae13de23d1ad2eb14b58d68 |
commit_message: '[skip ci]'\npull_request_title: '[docs] Sync translations with Crowdin'\npull_request_labels: [l10n]\nfiles:\n - source: /docs/src/**/*.md\n ignore:\n - /**/%file_name%-%two_letters_code%.md\n - /docs/src/pages/getting-started/templates/*/**/*\n - /docs/src/pages/premium-themes/*/**/*\n - /docs/src/pages/discover-more/backers/*\n - /docs/src/pages/discover-more/roadmap/*\n - /docs/src/pages/company/**/*\n - /docs/pages/careers/**/*\n translation: /%original_path%/%file_name%-%two_letters_code%.%file_extension%\n - source: /docs/data/**/*.md\n ignore:\n - /**/%file_name%-%two_letters_code%.md\n - /docs/data/material/getting-started/templates/*/**/*\n - /docs/data/material/discover-more/backers/*\n - /docs/data/material/discover-more/roadmap/*\n - /docs/data/material/premium-themes/*/**/*\n translation: /%original_path%/%file_name%-%two_letters_code%.%file_extension%\n - source: /docs/translations/**/*.json\n ignore:\n - /**/%file_name%-%two_letters_code%.json\n translation: /%original_path%/%file_name%-%two_letters_code%.%file_extension%\n | dataset_sample\yaml\javascript\crowdin.yml | crowdin.yml | YAML | 1,131 | 0.8 | 0 | 0 | node-utils | 919 | 2024-12-12T02:37:44.454608 | MIT | false | 54479dc254b8bc212cf10a48df38e2fe |
workflows:\n build-adev:\n name: Build adev\n prepare:\n - bazel clean\n workflow:\n - bazel build //adev:build\n | dataset_sample\yaml\javascript\dx-perf-workflows.yml | dx-perf-workflows.yml | YAML | 126 | 0.8 | 0 | 0 | node-utils | 75 | 2025-01-23T20:53:32.981893 | GPL-3.0 | false | a209ec0ca745a977b2272a1da72ccf09 |
# These are supported funding model platforms\n\ngithub: # Replace with up to 4 GitHub Sponsors-enabled usernames e.g., [user1, user2]\npatreon: # Replace with a single Patreon username\nopen_collective: mui-org\nko_fi: # Replace with a single Ko-fi username\ntidelift: npm/@mui/material\ncustom: # Replace with a single custom sponsorship URL\n | dataset_sample\yaml\javascript\FUNDING.yml | FUNDING.yml | YAML | 337 | 0.8 | 0 | 0.142857 | awesome-app | 954 | 2023-12-20T02:13:14.367414 | Apache-2.0 | false | 99945fef03eb7f70f567021676e5bbeb |
version: 1.0.0.{build}\nimage:\n - Visual Studio 2017\npull_requests:\n do_not_increment_build_number: true\n#branches:\n# only:\n# - master\n# - Developer\nskip_branch_with_pr: true\nconfiguration: Release\nbefore_build:\n - cmd: dotnet restore Samples\Senparc.Weixin.MP.Sample.vs2017\Senparc.Weixin.MP.Sample.vs2017.sln\nbuild:\n verbosity: minimal\n parallel: true\nbuild_script:\n - cmd: dotnet build -c Release Samples\Senparc.Weixin.MP.Sample.vs2017\Senparc.Weixin.MP.Sample.vs2017.sln\ntest_script:\n - cmd: dotnet test -c Release Samples\Senparc.Weixin.MP.Sample.vs2017\Senparc.Weixin.MP.Sample.vs2017.sln\nartifacts:\n - path: '**\*.nupkg'\n name: nupkg\n - path: '**\*.snupkg'\n name: snupkg\nnuget:\n disable_publish_on_pr: true\n#deploy:\n# provider: NuGet\n# api_key:\n# secure: <Your secret>\n\n | dataset_sample\yaml\JeffreySu_WeiXinMPSDK\appveyor.yml | appveyor.yml | YAML | 805 | 0.8 | 0 | 0.258065 | node-utils | 554 | 2023-11-14T20:58:38.704946 | BSD-3-Clause | false | 29f5b0d8facbe0ca98efbb442c00ea63 |
# ASP.NET Core (.NET Framework)\n# Build and test ASP.NET Core projects targeting the full .NET Framework.\n# Add steps that publish symbols, save build artifacts, and more:\n# https://docs.microsoft.com/azure/devops/pipelines/languages/dotnet-core\n\ntrigger:\n- master\n\npool:\n vmImage: 'windows-latest'\n\nvariables:\n solution: '**/tools/Senparc.Weixin.CLI/*.sln'\n buildPlatform: 'Any CPU'\n buildConfiguration: 'Release'\n\nsteps:\n- task: NuGetToolInstaller@1\n\n- task: UseDotNet@2\n displayName: 'Use .NET SDK 8'\n inputs:\n version: '8.0.x'\n packageType: 'sdk'\n\n# - task: NuGetCommand@2\n# inputs:\n# restoreSolution: '$(solution)'\n\n\n# - task: VSBuild@1\n# inputs:\n# solution: '$(solution)'\n# msbuildArgs: '/p:DeployOnBuild=true /p:WebPublishMethod=Package /p:PackageAsSingleFile=true /p:SkipInvalidConfigurations=true /p:DesktopBuildPackageLocation="$(build.artifactStagingDirectory)\WebApp.zip" /p:DeployIisAppPath="Default Web Site"'\n# platform: '$(buildPlatform)'\n# configuration: '$(buildConfiguration)'\n\n\n- task: NuGetCommand@2\n inputs:\n command: 'restore'\n restoreSolution: '$(solution)'\n feedsToUse: 'config'\n externalFeedCredentials: 'Nuget Official (V2)'\n \n- task: DotNetCoreCLI@2\n displayName: Build\n inputs:\n command: build\n projects: '$(solution)'\n arguments: '--configuration Release'\n\n- task: NuGetCommand@2\n displayName: Senparc.Weixin.CLI\n inputs:\n command: push\n packagesToPush: '**/Weixin.*.nupkg'\n nuGetFeedType: external\n publishFeedCredentials: 'Nuget Official (V2)'\n continueOnError: true\n condition: succeededOrFailed()\n\n\n- task: NuGetCommand@2\n displayName: Senparc.WeChat.CLI\n inputs:\n command: push\n packagesToPush: '**/WeChat.*.nupkg'\n nuGetFeedType: external\n publishFeedCredentials: 'Nuget Official (V2)'\n continueOnError: true\n condition: succeededOrFailed()\n\n\n# - task: VSTest@2\n# inputs:\n# platform: '$(buildPlatform)'\n# configuration: '$(buildConfiguration)'\n | dataset_sample\yaml\JeffreySu_WeiXinMPSDK\azure-pipelines.yml | azure-pipelines.yml | YAML | 1,996 | 0.8 | 0 | 0.274194 | vue-tools | 711 | 2023-10-14T13:27:23.485637 | GPL-3.0 | false | d453a03a9841064b329898b7041af5bf |
version: "2"\nchecks:\n argument-count:\n enabled: true\n config:\n threshold: 5\n file-lines:\n enabled: true\n config:\n threshold: 300\n method-complexity:\n enabled: true\n config:\n threshold: 15\n method-count:\n enabled: true\n config:\n threshold: 50\n method-lines:\n enabled: true\n config:\n threshold: 30\nplugins:\n fixme:\n enabled: false\n rubocop:\n enabled: true\n channel: rubocop-0-60\n\nexclude_patterns:\n - "*.*"\n - ".*"\n\n - Gemfile\n - LICENSE\n - Rakefile\n\n - benchmark/\n - docs/\n - exe/\n - features/\n - rake/\n - rubocop/\n - script/\n - spec/\n - test/\n - vendor/\n\n - lib/blank_template/\n - lib/site_template/\n - lib/theme_template/\n - lib/jekyll/mime.types\n - lib/jekyll/commands/serve/livereload_assets/livereload.js\n | dataset_sample\yaml\jekyll_jekyll\.codeclimate.yml | .codeclimate.yml | YAML | 800 | 0.7 | 0 | 0 | python-kit | 391 | 2025-04-24T09:39:50.440669 | Apache-2.0 | false | 5ad7de389adf483c9ca5557fbe0f3a1b |
# -----------------------------------------------------------------------------\n# This file has been formatted via a Rake Task configuring cops from\n# RuboCop v1.57.2.\n#\n# To add more cops, paste configurations at the end of the file and run\n# the rake task via `bundle exec rake rubocop:format_config`.\n# -----------------------------------------------------------------------------\n\ninherit_from: ".rubocop_todo.yml"\n\nrequire:\n- "./rubocop/jekyll"\n- rubocop-minitest\n- rubocop-performance\n- rubocop-rake\n- rubocop-rspec\n\n# Directive for all cops\nAllCops:\n TargetRubyVersion: 2.7\n Include:\n - lib/**/*.rb\n - test/**/*.rb\n Exclude:\n - benchmark/**/*\n - bin/**/*\n - exe/**/*\n - script/**/*\n - tmp/**/*\n - vendor/**/*\n\n# Configure custom cop\nJekyll/NoPutsAllowed:\n Exclude:\n - rake/*.rake\n\n# Configure built-in cops\n# =======================\n\n# Gemspec cops\n# ----------------------------------------\nGemspec/DeprecatedAttributeAssignment:\n Enabled: true\nGemspec/DevelopmentDependencies:\n Enabled: false\nGemspec/RequireMFA:\n Enabled: false\n\n# Layout cops\n# ----------------------------------------\nLayout/BeginEndAlignment:\n Enabled: true\nLayout/EmptyComment:\n Enabled: false\nLayout/EmptyLinesAroundAttributeAccessor:\n Enabled: true\nLayout/EndAlignment:\n Severity: error\nLayout/EndOfLine:\n Enabled: false\nLayout/FirstArrayElementIndentation:\n EnforcedStyle: consistent\nLayout/FirstHashElementIndentation:\n EnforcedStyle: consistent\nLayout/HashAlignment:\n EnforcedHashRocketStyle: table\nLayout/IndentationWidth:\n Severity: error\nLayout/LineContinuationLeadingSpace:\n Enabled: true\nLayout/LineContinuationSpacing:\n Enabled: true\nLayout/LineEndStringConcatenationIndentation:\n Enabled: true\nLayout/LineLength:\n Exclude:\n - Gemfile\n - Rakefile\n - features/**/*.rb\n - rake/**/*.rake\n Max: 100\n Severity: warning\nLayout/MultilineMethodCallIndentation:\n EnforcedStyle: indented\nLayout/MultilineOperationIndentation:\n EnforcedStyle: indented\nLayout/SpaceAroundMethodCallOperator:\n Enabled: true\nLayout/SpaceBeforeBrackets:\n Enabled: true\nLayout/SpaceInsideHashLiteralBraces:\n Enabled: true\n Exclude:\n - test/**/*.rb\n\n# Lint cops\n# ----------------------------------------\nLint/AmbiguousAssignment:\n Enabled: true\nLint/AmbiguousOperatorPrecedence:\n Enabled: true\nLint/AmbiguousRange:\n Enabled: true\nLint/BinaryOperatorWithIdenticalOperands:\n Enabled: true\nLint/ConstantDefinitionInBlock:\n Enabled: true\n Exclude:\n - test/**/*.rb\nLint/ConstantOverwrittenInRescue:\n Enabled: true\nLint/DeprecatedConstants:\n Enabled: true\nLint/DeprecatedOpenSSLConstant:\n Enabled: true\nLint/DuplicateBranch:\n Enabled: true\nLint/DuplicateElsifCondition:\n Enabled: true\nLint/DuplicateMagicComment:\n Enabled: true\nLint/DuplicateMatchPattern:\n Enabled: true\nLint/DuplicateRegexpCharacterClassElement:\n Enabled: true\nLint/DuplicateRequire:\n Enabled: true\nLint/DuplicateRescueException:\n Enabled: true\nLint/EmptyBlock:\n Enabled: true\nLint/EmptyClass:\n Enabled: true\nLint/EmptyConditionalBody:\n Enabled: true\nLint/EmptyFile:\n Enabled: true\nLint/FloatComparison:\n Enabled: true\nLint/HashCompareByIdentity:\n Enabled: true\nLint/IdentityComparison:\n Enabled: true\nLint/LambdaWithoutLiteralBlock:\n Enabled: true\nLint/MissingSuper:\n Enabled: false\nLint/MixedCaseRange:\n Enabled: true\nLint/MixedRegexpCaptureTypes:\n Enabled: false\nLint/NestedPercentLiteral:\n Exclude:\n - test/test_site.rb\nLint/NoReturnInBeginEndBlocks:\n Enabled: true\nLint/NumberedParameterAssignment:\n Enabled: true\nLint/OrAssignmentToConstant:\n Enabled: true\nLint/OutOfRangeRegexpRef:\n Enabled: true\nLint/RaiseException:\n Enabled: true\nLint/RedundantDirGlobSort:\n Enabled: true\nLint/RedundantRegexpQuantifiers:\n Enabled: true\nLint/RedundantSafeNavigation:\n Enabled: true\nLint/RefinementImportMethods:\n Enabled: true\nLint/RequireRangeParentheses:\n Enabled: true\nLint/RequireRelativeSelfPath:\n Enabled: true\nLint/SelfAssignment:\n Enabled: true\nLint/StructNewOverride:\n Enabled: true\nLint/SymbolConversion:\n Enabled: true\nLint/ToEnumArguments:\n Enabled: false\nLint/TopLevelReturnWithArgument:\n Enabled: true\nLint/TrailingCommaInAttributeDeclaration:\n Enabled: true\nLint/TripleQuotes:\n Enabled: true\nLint/UnexpectedBlockArity:\n Enabled: true\nLint/UnmodifiedReduceAccumulator:\n Enabled: true\nLint/UnreachableCode:\n Severity: error\nLint/UnreachableLoop:\n Enabled: true\nLint/UselessMethodDefinition:\n Enabled: true\nLint/UselessRuby2Keywords:\n Enabled: true\nLint/UselessTimes:\n Enabled: true\nLint/Void:\n Exclude:\n - lib/jekyll/site.rb\n\n# Metrics cops\n# ----------------------------------------\nMetrics/AbcSize:\n Max: 23\nMetrics/BlockLength:\n Exclude:\n - lib/jekyll/configuration.rb\n - rake/**/*.rake\n - test/**/*.rb\nMetrics/ClassLength:\n Exclude:\n - features/**/*.rb\n - lib/jekyll/commands/serve.rb\n - lib/jekyll/configuration.rb\n - lib/jekyll/document.rb\n - lib/jekyll/site.rb\n - test/**/*.rb\n Max: 240\nMetrics/CollectionLiteralLength:\n Enabled: true\nMetrics/CyclomaticComplexity:\n Exclude:\n - lib/jekyll/commands/serve.rb\n - lib/jekyll/utils.rb\n Max: 11\nMetrics/MethodLength:\n CountComments: false\n Max: 20\n Severity: error\nMetrics/ModuleLength:\n Exclude:\n - lib/jekyll/filters.rb\n Max: 240\nMetrics/ParameterLists:\n Max: 4\nMetrics/PerceivedComplexity:\n Max: 13\n\n# Minitest cops\n# ----------------------------------------\nMinitest/AssertEmptyLiteral:\n Enabled: false\nMinitest/AssertInDelta:\n Enabled: true\nMinitest/AssertKindOf:\n Enabled: true\nMinitest/AssertOperator:\n Enabled: false\nMinitest/AssertOutput:\n Enabled: true\nMinitest/AssertPathExists:\n Enabled: true\nMinitest/AssertPredicate:\n Enabled: false\nMinitest/AssertRaisesWithRegexpArgument:\n Enabled: true\nMinitest/AssertSame:\n Enabled: true\nMinitest/AssertSilent:\n Enabled: true\nMinitest/AssertWithExpectedArgument:\n Enabled: true\nMinitest/AssertionInLifecycleHook:\n Enabled: true\nMinitest/DuplicateTestRun:\n Enabled: true\nMinitest/EmptyLineBeforeAssertionMethods:\n Enabled: false\nMinitest/LifecycleHooksOrder:\n Enabled: true\nMinitest/LiteralAsActualArgument:\n Enabled: true\nMinitest/MultipleAssertions:\n Enabled: true\nMinitest/NonExecutableTestMethod:\n Enabled: true\nMinitest/NonPublicTestMethod:\n Enabled: true\nMinitest/RedundantMessageArgument:\n Enabled: true\nMinitest/RefuteInDelta:\n Enabled: true\nMinitest/RefuteKindOf:\n Enabled: true\nMinitest/RefuteOperator:\n Enabled: false\nMinitest/RefutePathExists:\n Enabled: true\nMinitest/RefutePredicate:\n Enabled: false\nMinitest/RefuteSame:\n Enabled: true\nMinitest/ReturnInTestMethod:\n Enabled: true\nMinitest/SkipEnsure:\n Enabled: true\nMinitest/SkipWithoutReason:\n Enabled: true\nMinitest/TestFileName:\n Enabled: true\n Exclude:\n - test/helper.rb\nMinitest/TestMethodName:\n Enabled: false\nMinitest/UnreachableAssertion:\n Enabled: true\nMinitest/UnspecifiedException:\n Enabled: true\nMinitest/UselessAssertion:\n Enabled: true\n\n# Naming cops\n# ----------------------------------------\nNaming/FileName:\n Enabled: false\nNaming/HeredocDelimiterNaming:\n Exclude:\n - test/**/*.rb\nNaming/MemoizedInstanceVariableName:\n Exclude:\n - lib/jekyll/convertible.rb\n - lib/jekyll/drops/site_drop.rb\n - lib/jekyll/drops/unified_payload_drop.rb\n - lib/jekyll/page_without_a_file.rb\n\n# Performance cops\n# ----------------------------------------\nPerformance/AncestorsInclude:\n Enabled: false\nPerformance/ArraySemiInfiniteRangeSlice:\n Enabled: true\nPerformance/BigDecimalWithNumericArgument:\n Enabled: true\nPerformance/BlockGivenWithExplicitBlock:\n Enabled: true\nPerformance/ChainArrayAllocation:\n Enabled: true\nPerformance/CollectionLiteralInLoop:\n Enabled: true\nPerformance/ConstantRegexp:\n Enabled: true\nPerformance/MapCompact:\n Enabled: true\nPerformance/MapMethodChain:\n Enabled: true\nPerformance/MethodObjectAsBlock:\n Enabled: true\nPerformance/RedundantEqualityComparisonBlock:\n Enabled: false\nPerformance/RedundantSortBlock:\n Enabled: true\nPerformance/RedundantSplitRegexpArgument:\n Enabled: true\nPerformance/RedundantStringChars:\n Enabled: true\nPerformance/ReverseFirst:\n Enabled: true\nPerformance/SortReverse:\n Enabled: false\nPerformance/Squeeze:\n Enabled: true\nPerformance/StringIdentifierArgument:\n Enabled: true\nPerformance/StringInclude:\n Enabled: true\n Exclude:\n - lib/jekyll/utils/platforms.rb\nPerformance/Sum:\n Enabled: true\n\n# Security cops\n# ----------------------------------------\nSecurity/CompoundHash:\n Enabled: true\nSecurity/IoMethods:\n Enabled: true\nSecurity/MarshalLoad:\n Exclude:\n - lib/jekyll/regenerator.rb\n - test/**/*.rb\nSecurity/YAMLLoad:\n Exclude:\n - features/**/*.rb\n - test/**/*.rb\n\n# Style cops\n# ----------------------------------------\nStyle/AccessModifierDeclarations:\n Enabled: false\nStyle/AccessorGrouping:\n Enabled: true\nStyle/Alias:\n EnforcedStyle: prefer_alias_method\nStyle/AndOr:\n Severity: error\nStyle/ArgumentsForwarding:\n Enabled: false\nStyle/ArrayCoercion:\n Enabled: true\nStyle/ArrayIntersect:\n Enabled: true\nStyle/BisectedAttrAccessor:\n Enabled: true\nStyle/CaseLikeIf:\n Enabled: true\nStyle/ClassAndModuleChildren:\n Exclude:\n - test/**/*.rb\nStyle/ClassEqualityComparison:\n Enabled: true\nStyle/CollectionCompact:\n Enabled: true\nStyle/CombinableLoops:\n Enabled: true\nStyle/ConcatArrayLiterals:\n Enabled: true\nStyle/DirEmpty:\n Enabled: true\nStyle/DocumentDynamicEvalDefinition:\n Enabled: true\nStyle/Documentation:\n Enabled: false\nStyle/DoubleNegation:\n Enabled: false\nStyle/EmptyHeredoc:\n Enabled: true\nStyle/EndlessMethod:\n Enabled: true\nStyle/EnvHome:\n Enabled: true\nStyle/ExactRegexpMatch:\n Enabled: true\nStyle/ExplicitBlockArgument:\n Enabled: false\nStyle/ExponentialNotation:\n Enabled: true\nStyle/FetchEnvVar:\n Enabled: false\nStyle/FileEmpty:\n Enabled: true\nStyle/FileRead:\n Enabled: false\nStyle/FileWrite:\n Enabled: true\nStyle/FormatStringToken:\n Exclude:\n - lib/jekyll/liquid_renderer/table.rb\n - lib/jekyll/profiler.rb\n - lib/jekyll/utils/ansi.rb\nStyle/FrozenStringLiteralComment:\n EnforcedStyle: always\nStyle/GlobalStdStream:\n Enabled: true\nStyle/GuardClause:\n Enabled: false\nStyle/HashAsLastArrayItem:\n Enabled: true\nStyle/HashConversion:\n Enabled: true\nStyle/HashEachMethods:\n Enabled: true\nStyle/HashExcept:\n Enabled: true\nStyle/HashLikeCase:\n Enabled: true\nStyle/HashSyntax:\n EnforcedStyle: hash_rockets\n Severity: error\nStyle/HashTransformKeys:\n Enabled: false\nStyle/HashTransformValues:\n Enabled: true\nStyle/IfWithBooleanLiteralBranches:\n Enabled: true\nStyle/KeywordParametersOrder:\n Enabled: true\nStyle/MagicCommentFormat:\n Enabled: true\nStyle/MapCompactWithConditionalBlock:\n Enabled: true\nStyle/MapToHash:\n Enabled: true\nStyle/MinMaxComparison:\n Enabled: true\nStyle/MixinUsage:\n Exclude:\n - test/helper.rb\nStyle/ModuleFunction:\n Enabled: false\nStyle/MultilineTernaryOperator:\n Severity: error\nStyle/NegatedIfElseCondition:\n Enabled: true\nStyle/NestedFileDirname:\n Enabled: true\nStyle/NilLambda:\n Enabled: true\nStyle/ObjectThen:\n Enabled: true\nStyle/OperatorMethodCall:\n Enabled: true\nStyle/OptionalBooleanParameter:\n Enabled: true\n Exclude:\n - lib/jekyll/log_adapter.rb\nStyle/PercentLiteralDelimiters:\n PreferredDelimiters:\n "%Q": "{}"\n "%W": "()"\n "%q": "{}"\n "%r": "!!"\n "%s": "()"\n "%w": "()"\n "%x": "()"\nStyle/QuotedSymbols:\n Enabled: true\nStyle/RedundantArgument:\n Enabled: true\nStyle/RedundantArrayConstructor:\n Enabled: true\nStyle/RedundantAssignment:\n Enabled: true\nStyle/RedundantConstantBase:\n Enabled: true\nStyle/RedundantCurrentDirectoryInPath:\n Enabled: true\nStyle/RedundantDoubleSplatHashBraces:\n Enabled: true\nStyle/RedundantEach:\n Enabled: true\nStyle/RedundantFetchBlock:\n Enabled: false\nStyle/RedundantFileExtensionInRequire:\n Enabled: true\nStyle/RedundantFilterChain:\n Enabled: true\nStyle/RedundantHeredocDelimiterQuotes:\n Enabled: true\nStyle/RedundantInitialize:\n Enabled: true\n Exclude:\n - lib/jekyll/plugin.rb\nStyle/RedundantLineContinuation:\n Enabled: true\nStyle/RedundantRegexpArgument:\n Enabled: true\nStyle/RedundantRegexpCharacterClass:\n Enabled: true\nStyle/RedundantRegexpConstructor:\n Enabled: true\nStyle/RedundantRegexpEscape:\n Enabled: true\nStyle/RedundantSelfAssignment:\n Enabled: true\nStyle/RedundantSelfAssignmentBranch:\n Enabled: true\nStyle/RedundantStringEscape:\n Enabled: true\nStyle/RegexpLiteral:\n EnforcedStyle: percent_r\nStyle/RescueModifier:\n Enabled: false\nStyle/ReturnNilInPredicateMethodDefinition:\n Enabled: true\nStyle/SafeNavigation:\n Exclude:\n - lib/jekyll/document.rb\n - lib/jekyll/page.rb\nStyle/SignalException:\n EnforcedStyle: only_raise\nStyle/SingleArgumentDig:\n Enabled: true\nStyle/SingleLineDoEndBlock:\n Enabled: true\nStyle/SlicingWithRange:\n Enabled: false\nStyle/SoleNestedConditional:\n Enabled: true\nStyle/StringChars:\n Enabled: true\nStyle/StringConcatenation:\n Enabled: true\n Exclude:\n - lib/jekyll/commands/*.rb\n - test/**/*.rb\nStyle/StringLiterals:\n EnforcedStyle: double_quotes\nStyle/StringLiteralsInInterpolation:\n EnforcedStyle: double_quotes\nStyle/SwapValues:\n Enabled: true\nStyle/SymbolArray:\n EnforcedStyle: brackets\nStyle/TrailingCommaInArrayLiteral:\n EnforcedStyleForMultiline: consistent_comma\nStyle/TrailingCommaInHashLiteral:\n EnforcedStyleForMultiline: consistent_comma\n | dataset_sample\yaml\jekyll_jekyll\.rubocop.yml | .rubocop.yml | YAML | 13,222 | 0.95 | 0.001653 | 0.049069 | node-utils | 127 | 2025-01-23T09:28:06.131512 | MIT | false | 2191af43cb66d92d93fe24baf7b13f0c |
# This configuration was generated by\n# `rubocop --auto-gen-config --auto-gen-only-exclude`\n# on 2022-04-06 10:48:47 UTC using RuboCop version 1.26.1.\n# The point is for the user to remove these configuration records\n# one by one as the offenses are removed from the code base.\n# Note that changes in the inspected code, or installation of new\n# versions of RuboCop, may require this file to be generated again.\n\n# Offense count: 1\n# This cop supports safe auto-correction (--auto-correct).\nPerformance/BindCall:\n Exclude:\n - 'test/helper.rb'\n\n# Offense count: 1\nStyle/CombinableLoops:\n Exclude:\n - 'lib/jekyll/tags/post_url.rb'\n\n# Offense count: 1\n# Configuration parameters: AllowedMethods.\n# AllowedMethods: respond_to_missing?\nStyle/OptionalBooleanParameter:\n Exclude:\n - 'lib/jekyll/log_adapter.rb'\n | dataset_sample\yaml\jekyll_jekyll\.rubocop_todo.yml | .rubocop_todo.yml | YAML | 816 | 0.95 | 0.04 | 0.590909 | react-lib | 287 | 2023-08-09T07:00:57.046919 | BSD-3-Clause | false | f7930e722a717263780fef3ab6b15bdc |
linters:\n enable:\n - gofumpt\n - thelper\n - goimports\n - tparallel\n - wastedassign\n - unparam\n - prealloc\n - unconvert\n - exhaustive\n - makezero\n - nakedret\n - copyloopvar\n fast: false\n\nlinters-settings:\n copyloopvar:\n # Check all assigning the loop variable to another variable.\n # Default: false\n # If true, an assignment like `a := x` will be detected as an error.\n check-alias: true\n exhaustive:\n default-signifies-exhaustive: true\n staticcheck:\n # SA1019 is for checking that we're not using fields marked as deprecated\n # in a comment. It decides this in a loose way so I'm silencing it. Also because\n # it's tripping on our own structs.\n checks: ["all", "-SA1019"]\n nakedret:\n # the gods will judge me but I just don't like naked returns at all\n max-func-lines: 0\n\nrun:\n go: "1.24"\n timeout: 10m\n | dataset_sample\yaml\jesseduffield_lazygit\.golangci.yml | .golangci.yml | YAML | 879 | 0.95 | 0.027778 | 0.205882 | vue-tools | 619 | 2024-01-10T19:21:17.474845 | Apache-2.0 | false | 173fb47f9f689993abacb97c8565d2f8 |
# This is an example goreleaser.yaml file with some sane defaults.\n# Make sure to check the documentation at http://goreleaser.com\nbuilds:\n - env:\n - CGO_ENABLED=0\n goos:\n - freebsd\n - windows\n - darwin\n - linux\n goarch:\n - amd64\n - arm\n - arm64\n - '386'\n # Default is `-s -w -X main.version={{.Version}} -X main.commit={{.Commit}} -X main.date={{.Date}}`.\n ldflags:\n - -s -w -X main.version={{.Version}} -X main.commit={{.Commit}} -X main.date={{.Date}} -X main.buildSource=binaryRelease\n\narchives:\n - replacements:\n darwin: Darwin\n linux: Linux\n windows: Windows\n 386: 32-bit\n amd64: x86_64\n format_overrides:\n - goos: windows\n format: zip\nchecksum:\n name_template: 'checksums.txt'\nsnapshot:\n name_template: '{{ .Tag }}-next'\nchangelog:\n use: github-native\n sort: asc\nbrews:\n -\n # Repository to push the tap to.\n tap:\n owner: jesseduffield\n name: homebrew-lazygit\n\n # Your app's homepage.\n # Default is empty.\n homepage: 'https://github.com/jesseduffield/lazygit/'\n\n # Your app's description.\n # Default is empty.\n description: 'A simple terminal UI for git commands, written in Go'\n\n # # Packages your package depends on.\n # dependencies:\n # - git\n # - zsh\n # # Packages that conflict with your package.\n # conflicts:\n # - svn\n # - bash\n | dataset_sample\yaml\jesseduffield_lazygit\.goreleaser.yml | .goreleaser.yml | YAML | 1,415 | 0.8 | 0.016949 | 0.290909 | python-kit | 218 | 2024-10-23T15:22:01.889902 | BSD-3-Clause | false | 17fa4a7ae16161728e33a9f38da6e0a9 |
version: 2\n\nupdates:\n - package-ecosystem: "github-actions"\n directory: "/"\n schedule:\n interval: "daily"\n | dataset_sample\yaml\jgm_cmark\.github\dependabot.yml | dependabot.yml | YAML | 118 | 0.7 | 0 | 0 | node-utils | 966 | 2025-04-27T03:14:06.800780 | BSD-3-Clause | false | 635f49e43e62960f21285bf80143ab95 |
name: CI tests\n\non:\n push:\n branches-ignore:\n - 'dependabot/**'\n pull_request:\n workflow_dispatch:\n\njobs:\n\n linter:\n\n runs-on: ubuntu-latest\n\n steps:\n\n - uses: actions/checkout@v4\n - name: Install clang-tidy\n run: |\n sudo apt-get install -y clang-tidy\n - name: lint with clang-tidy\n run: |\n make lint\n env:\n CC: clang\n CXX: clang++\n\n posix:\n\n strategy:\n fail-fast: false\n matrix:\n os: [linux, macos]\n cc: [clang, gcc]\n build_type: [shared, static]\n sanitizers: ['', ASan]\n\n include:\n # Translate human readable labels\n - os: 'linux'\n image: 'ubuntu-latest'\n - os: 'macos'\n image: 'macos-latest'\n - cc: 'clang'\n cxx: 'clang++'\n - cc: 'gcc'\n cxx: 'g++'\n - build_type: 'shared'\n cmake_shared: 'YES'\n - build_type: 'static'\n cmake_shared: 'NO'\n - sanitizers: 'ASan'\n san_cflags: '-fsanitize=address,undefined -fno-sanitize-recover=all'\n\n # When building shared libraries, they will be loaded\n # dynamically from Python when testing. This means that\n # we have to use a preloaded shared libasan.\n - sanitizers: 'ASan'\n os: 'linux'\n cc: 'gcc'\n build_type: 'shared'\n test_env: 'LD_PRELOAD=$(gcc -print-file-name=libasan.so)'\n - sanitizers: 'ASan'\n os: 'linux'\n cc: 'clang'\n build_type: 'shared'\n # clang defaults to -static-libsasn\n asan_cflags: '-shared-libasan'\n test_env: 'LD_PRELOAD=$(clang -print-file-name=libclang_rt.asan-x86_64.so)'\n\n # We have to disable LeakSanitizer in shared library builds\n # because we get false positives from Python.\n - sanitizers: 'ASan'\n build_type: 'shared'\n asan_opts: 'detect_leaks=0'\n\n # The static build can run with LeakSanitizer.\n # gcc defaults to -shared-libasan and needs -static-libasan\n - sanitizers: 'ASan'\n cc: 'gcc'\n build_type: 'static'\n asan_cflags: '-static-libasan'\n\n exclude:\n # gcc is just an alias for clang on macOS\n - os: 'macos'\n cc: 'gcc'\n # Shared libasan doesn't work with macOS system Python.\n - os: 'macos'\n sanitizers: 'ASan'\n build_type: 'shared'\n\n runs-on: ${{ matrix.image }}\n\n env:\n ASAN_OPTIONS: ${{ matrix.asan_opts }}\n CC: ${{ matrix.cc }}\n CXX: ${{ matrix.cxx }}\n CFLAGS: '${{ matrix.san_cflags }} ${{ matrix.asan_cflags }}'\n CXXFLAGS: '${{ matrix.san_cflags }} ${{ matrix.asan_cflags }}'\n\n steps:\n - uses: actions/checkout@v4\n - name: Build and test\n run: |\n cmake \\n -DBUILD_SHARED_LIBS=${{ matrix.cmake_shared }} \\n -DCMAKE_BUILD_TYPE=Debug \\n -S . -B build\n cmake --build build\n ${{ matrix.test_env }} ctest --test-dir build --output-on-failure\n\n windows:\n\n runs-on: windows-latest\n strategy:\n fail-fast: false\n matrix:\n build_type: [shared, static]\n include:\n - build_type: 'shared'\n cmake_shared: 'YES'\n - build_type: 'static'\n cmake_shared: 'NO'\n\n steps:\n - uses: actions/checkout@v4\n - uses: ilammy/msvc-dev-cmd@v1\n - name: Build and test\n run: |\n cmake ^\n -DBUILD_SHARED_LIBS=${{ matrix.cmake_shared }} ^\n -DCMAKE_BUILD_TYPE=Debug ^\n -S . -B build\n cmake --build build\n ctest --test-dir build -C Debug --output-on-failure\n shell: cmd\n - name: Upload artifact\n if: ${{ matrix.build_type == 'static' }}\n uses: actions/upload-artifact@v4\n with:\n name: cmark windows ${{ matrix.build_type }}\n path: build/src/Debug/cmark.exe\n if-no-files-found: error\n | dataset_sample\yaml\jgm_cmark\.github\workflows\ci.yml | ci.yml | YAML | 4,024 | 0.8 | 0.02069 | 0.087302 | react-lib | 665 | 2024-01-31T18:31:20.294988 | MIT | false | d3ecf40d85f7a464963ed119056bffa9 |
name: CIFuzz\non: [pull_request]\njobs:\n Fuzzing:\n runs-on: ubuntu-latest\n steps:\n - name: Build Fuzzers\n uses: google/oss-fuzz/infra/cifuzz/actions/build_fuzzers@master\n with:\n oss-fuzz-project-name: 'cmark'\n dry-run: false\n - name: Run Fuzzers\n uses: google/oss-fuzz/infra/cifuzz/actions/run_fuzzers@master\n with:\n oss-fuzz-project-name: 'cmark'\n fuzz-seconds: 600\n dry-run: false\n - name: Upload Crash\n uses: actions/upload-artifact@v4\n if: failure()\n with:\n name: artifacts\n path: ./out/artifacts\n | dataset_sample\yaml\jgm_cmark\.github\workflows\fuzz.yml | fuzz.yml | YAML | 599 | 0.7 | 0.043478 | 0 | vue-tools | 222 | 2024-09-04T14:16:32.361827 | BSD-3-Clause | false | af17bdb8da9cccc0702d83b9bf74842a |
image: Visual Studio 2022\nbuild: false\nplatform:\n - x64\nclone_folder: C:\projects\joomla-cms\n\n## Build matrix for lowest and highest possible targets\nenvironment:\n PHPBuild: "x64"\n matrix:\n - php_ver_target: 8.1\n\n\ninit:\n - SET PATH=C:\Tools\php;%PATH%\n - SET COMPOSER_NO_INTERACTION=1\n - SET ANSICON=121x90 (121x90)\n\n## Install PHP and composer, and run the appropriate composer command\ninstall:\n - ps: >-\n appveyor-retry choco install --no-progress --params '""/InstallDir:C:\tools\php""' --ignore-checksums -y php --version ((choco search php --exact --all-versions -r | select-string -pattern $env:php_ver_target | sort { [version]($_ -split '\|' | select -last 1) } -Descending | Select-Object -first 1) -replace '[php|]','')\n - cd C:\tools\php\n - copy php.ini-production php.ini /Y\n - echo date.timezone="UTC" >> php.ini\n - echo extension_dir=ext >> php.ini\n - echo extension=php_openssl.dll >> php.ini\n - echo extension=php_mbstring.dll >> php.ini\n - echo extension=php_fileinfo.dll >> php.ini\n - IF %php_ver_target% LSS 8 echo extension=php_gd2.dll >> php.ini\n - IF %php_ver_target% GEQ 8 echo extension=gd >> php.ini\n - echo extension=php_gmp.dll >> php.ini\n - echo extension=php_pgsql.dll >> php.ini\n - echo extension=php_pdo_pgsql.dll >> php.ini\n - echo extension=php_pdo_mysql.dll >> php.ini\n - echo extension=php_mysqli.dll >> php.ini\n - echo extension=php_curl.dll >> php.ini\n - echo zend_extension=php_opcache.dll >> php.ini\n - echo opcache.enable_cli=1 >> php.ini\n - echo extension=php_ldap.dll >> php.ini\n - choco install composer\n - cd C:\projects\joomla-cms\n - refreshenv\n - echo "TODO Ignore platform reqs till all composer dependencies are compatible with php 8.1"\n - composer install --no-progress --profile --ignore-platform-req=ext-sodium --ignore-platform-reqs\n\nhosts:\n openldap: 127.0.0.1\n\nservices:\n - iis\n\nbefore_test:\n# Run openldap docker image\n - ps: docker pull bitnami/openldap:2.6.3\n - ps: docker run --rm --name openldap --publish 1389:1389 --publish 1636:1636 -v ${pwd}\tests\certs:/certificates --env LDAP_ADMIN_USERNAME=admin --env LDAP_ADMIN_PASSWORD=adminpassword --env LDAP_USERS=customuser --env LDAP_PASSWORDS=custompassword --env LDAP_ENABLE_TLS=yes --env LDAP_TLS_CERT_FILE=/certificates/openldap.crt --env LDAP_TLS_KEY_FILE=/certificates/openldap.key --env LDAP_TLS_CA_FILE=/certificates/CA.crt --env BITNAMI_DEBUG=true --env LDAP_CONFIG_ADMIN_ENABLED=yes --env LDAP_CONFIG_ADMIN_USERNAME=admin --env LDAP_CONFIG_ADMIN_PASSWORD=configpassword -d bitnami/openldap:2.6.3\n# Database setup for MySQL via PowerShell tools\n - ps: Start-Service MySQL80\n - >\n "C:\Program Files\MySQL\MySQL Server 8.0\bin\mysql" -u root -p"Password12!" -e "CREATE DATABASE IF NOT EXISTS test_joomla;"\n# Wait till slapd has started\n - ps: |\n $Counter=0\n $Found=$false\n While ( ! $Found -and $Counter -lt 60 ) {\n $Found = ( docker logs openldap 2>&1 | Select-String -Quiet "\*\* Starting slapd \*\*" )\n Start-Sleep -Seconds 1\n $Counter++\n "$Counter Waiting for slapd"\n }\n if ( ! $Found ) {\n Write-Error -Message "`nERROR: slapd not started (in time)!" -ErrorAction Stop\n exit 1\n }\n\ntest_script:\n - cd C:\projects\joomla-cms\n - libraries/vendor/bin/phpunit --testsuite Unit\n - libraries/vendor/bin/phpunit --testsuite Integration --configuration tests/phpunit-appveyor.xml.dist\n\non_failure:\n - ps: docker logs openldap 2>&1\n | dataset_sample\yaml\joomla_joomla-cms\.appveyor.yml | .appveyor.yml | YAML | 3,516 | 0.8 | 0.04878 | 0.068493 | vue-tools | 857 | 2024-05-19T15:40:39.796400 | Apache-2.0 | false | b74278c8f3ab37fdaa1cc787fb17dab1 |
---\nkind: pipeline\nname: default\n\nclone:\n\nsteps:\n - name: setup\n image: joomlaprojects/docker-images:php8.4\n volumes:\n - name: certificates\n path: /certificates\n commands:\n - cp -v tests/certs/* /certificates/\n\n - name: composer\n image: joomlaprojects/docker-images:php8.4\n volumes:\n - name: composer-cache\n path: /tmp/composer-cache\n commands:\n - composer validate --no-check-all --strict\n - composer install --no-progress --ignore-platform-reqs\n\n - name: phpcs\n image: joomlaprojects/docker-images:php8.1\n depends_on: [ composer ]\n commands:\n - echo $(date)\n - ./libraries/vendor/bin/php-cs-fixer fix -vvv --dry-run --diff\n - ./libraries/vendor/bin/phpcs --extensions=php -p --standard=ruleset.xml .\n - echo $(date)\n\n - name: phpstan\n image: joomlaprojects/docker-images:php8.2\n depends_on: [ phpcs ]\n failure: ignore\n commands:\n - ./libraries/vendor/bin/phpstan\n\n - name: npm\n image: node:20-bullseye-slim\n depends_on: [ phpcs ]\n volumes:\n - name: npm-cache\n path: /tmp/npm-cache\n environment:\n npm_config_cache: /tmp/npm-cache\n commands:\n - npm ci --unsafe-perm\n\n - name: php81-unit\n depends_on: [ phpcs ]\n image: joomlaprojects/docker-images:php8.1\n commands:\n - php -v\n - ./libraries/vendor/bin/phpunit --testsuite Unit\n\n - name: php82-unit\n depends_on: [ phpcs ]\n image: joomlaprojects/docker-images:php8.2\n commands:\n - php -v\n - ./libraries/vendor/bin/phpunit --testsuite Unit\n\n - name: php83-unit\n depends_on: [ phpcs ]\n image: joomlaprojects/docker-images:php8.3\n commands:\n - php -v\n - ./libraries/vendor/bin/phpunit --testsuite Unit\n\n - name: php84-unit\n depends_on: [ phpcs ]\n image: joomlaprojects/docker-images:php8.4\n commands:\n - php -v\n - ./libraries/vendor/bin/phpunit --testsuite Unit\n\n - name: php81-integration\n depends_on: [ npm ]\n image: joomlaprojects/docker-images:php8.1\n commands:\n - php -v\n - ./libraries/vendor/bin/phpunit --testsuite Integration\n\n - name: php82-integration\n depends_on: [ npm ]\n image: joomlaprojects/docker-images:php8.2\n commands:\n - php -v\n - ./libraries/vendor/bin/phpunit --testsuite Integration\n\n - name: php83-integration\n depends_on: [ npm ]\n image: joomlaprojects/docker-images:php8.3\n commands:\n - php -v\n - ./libraries/vendor/bin/phpunit --testsuite Integration\n\n - name: php84-integration\n depends_on: [ npm ]\n image: joomlaprojects/docker-images:php8.4\n commands:\n - php -v\n - ./libraries/vendor/bin/phpunit --testsuite Integration\n\n - name: php81-integration-pgsql\n depends_on: [ npm ]\n image: joomlaprojects/docker-images:php8.1\n commands:\n - php -v\n - ./libraries/vendor/bin/phpunit --testsuite Integration --configuration phpunit-pgsql.xml.dist\n\n - name: php82-integration-pgsql\n depends_on: [ npm ]\n image: joomlaprojects/docker-images:php8.2\n commands:\n - php -v\n - ./libraries/vendor/bin/phpunit --testsuite Integration --configuration phpunit-pgsql.xml.dist\n\n - name: php83-integration-pgsql\n depends_on: [ npm ]\n image: joomlaprojects/docker-images:php8.3\n commands:\n - php -v\n - ./libraries/vendor/bin/phpunit --testsuite Integration --configuration phpunit-pgsql.xml.dist\n\n - name: php84-integration-pgsql\n depends_on: [ npm ]\n image: joomlaprojects/docker-images:php8.4\n commands:\n - php -v\n - ./libraries/vendor/bin/phpunit --testsuite Integration --configuration phpunit-pgsql.xml.dist\n\n - name: scss-cs\n depends_on: [ npm ]\n image: node:current-alpine\n commands:\n - npm run lint:css\n\n - name: javascript-cs\n depends_on: [ npm ]\n image: node:current-alpine\n commands:\n - npm run lint:js\n - npm run lint:testjs\n\n - name: prepare_system_tests\n depends_on:\n - npm\n image: joomlaprojects/docker-images:cypress8.4\n volumes:\n - name: cypress-cache\n path: /root/.cache/Cypress\n environment:\n CYPRESS_VERIFY_TIMEOUT: 100000\n commands:\n - cp cypress.config.dist.mjs cypress.config.mjs\n - npx cypress install\n - npx cypress verify\n\n - name: phpmin-system-mysql\n depends_on:\n - prepare_system_tests\n image: joomlaprojects/docker-images:cypress8.1\n volumes:\n - name: cypress-cache\n path: /root/.cache/Cypress\n environment:\n JOOMLA_INSTALLATION_DISABLE_LOCALHOST_CHECK: 1\n commands:\n - bash tests/System/entrypoint.sh "$(pwd)" cmysql mysqli mysql\n\n - name: phpmax-system-mysql\n depends_on:\n - phpmin-system-mysql\n image: joomlaprojects/docker-images:cypress8.4\n volumes:\n - name: cypress-cache\n path: /root/.cache/Cypress\n environment:\n JOOMLA_INSTALLATION_DISABLE_LOCALHOST_CHECK: 1\n commands:\n - bash tests/System/entrypoint.sh "$(pwd)" cmysqlmax mysqli mysql\n when:\n event:\n exclude:\n - pull_request\n\n\n - name: phpmin-system-postgres\n depends_on:\n - prepare_system_tests\n image: joomlaprojects/docker-images:cypress8.1\n volumes:\n - name: cypress-cache\n path: /root/.cache/Cypress\n environment:\n JOOMLA_INSTALLATION_DISABLE_LOCALHOST_CHECK: 1\n commands:\n - bash tests/System/entrypoint.sh "$(pwd)" cpostgres pgsql postgres\n when:\n event:\n exclude:\n - pull_request\n\n - name: phpmax-system-postgres\n depends_on:\n - phpmin-system-postgres\n image: joomlaprojects/docker-images:cypress8.4\n volumes:\n - name: cypress-cache\n path: /root/.cache/Cypress\n environment:\n JOOMLA_INSTALLATION_DISABLE_LOCALHOST_CHECK: 1\n commands:\n - bash tests/System/entrypoint.sh "$(pwd)" cpostgresmax pgsql postgres\n\n - name: artifacts-system-tests\n image: joomlaprojects/docker-images:packager\n depends_on:\n - phpmax-system-mysql\n - phpmax-system-postgres\n - phpmin-system-mysql\n - phpmin-system-postgres\n environment:\n WEB_SERVER:\n from_secret: webserver\n FTP_KEY:\n from_secret: ftp_key\n FTP_USER:\n from_secret: ftp_user\n FTP_HOST:\n from_secret: ftp_host\n ARTIFACTS_ROOT:\n from_secret: artifacts_root\n GITHUB_TOKEN:\n from_secret: github_token\n commands:\n - export PLUGIN_DEST_DIR=$DRONE_REPO/$DRONE_BRANCH/$DRONE_PULL_REQUEST/system-tests/$DRONE_BUILD_NUMBER\n - echo https://$WEB_SERVER/drone/$PLUGIN_DEST_DIR\n - mkdir -p ~/.ssh\n - eval $(ssh-agent -s)\n - echo -e "Host *\n\tStrictHostKeyChecking no\n\n" > ~/.ssh/config\n - echo "$FTP_KEY" > ~/.ssh/id_rsa\n - chmod 600 ~/.ssh/id_rsa\n - ssh-add\n - rclone config create artifacts sftp host $FTP_HOST user $FTP_USER port 22\n - rclone mkdir artifacts:$ARTIFACTS_ROOT/$DRONE_REPO/$DRONE_BRANCH/$DRONE_PULL_REQUEST/system-tests/$DRONE_BUILD_NUMBER\n - rclone copy tests/System/output/ artifacts:$ARTIFACTS_ROOT/$DRONE_REPO/$DRONE_BRANCH/$DRONE_PULL_REQUEST/system-tests/$DRONE_BUILD_NUMBER\n - 'curl -X POST "https://api.github.com/repos/$DRONE_REPO/statuses/$DRONE_COMMIT" -H "Content-Type: application/json" -H "Authorization: token $GITHUB_TOKEN" -d "{\"state\":\"failure\", \"context\": \"Artifacts from Failure\", \"description\": \"You can find artifacts from the failure of the build here:\", \"target_url\": \"https://$WEB_SERVER/drone/$PLUGIN_DEST_DIR\"}" > /dev/null'\n when:\n status:\n - failure\n\nvolumes:\n - name: composer-cache\n host:\n path: /tmp/composer-cache\n - name: cypress-cache\n host:\n path: /tmp/cypress-cache\n - name: npm-cache\n host:\n path: /tmp/npm-cache\n - name: certificates\n host:\n path: /tmp/certificates\n\nservices:\n - name: mysql\n image: mysql:8.0\n command: ["--default-authentication-plugin=mysql_native_password"]\n environment:\n MYSQL_USER: joomla_ut\n MYSQL_PASSWORD: joomla_ut\n MYSQL_ROOT_PASSWORD: joomla_ut\n MYSQL_DATABASE: test_joomla\n\n - name: postgres\n image: postgres:12-alpine\n ports:\n - 5432\n environment:\n POSTGRES_USER: root\n POSTGRES_PASSWORD: joomla_ut\n POSTGRES_DB: test_joomla\n\n - name: openldap\n image: bitnami/openldap:latest\n ports:\n - 1389\n - 1636\n volumes:\n - name: certificates\n path: /certificates\n environment:\n LDAP_ADMIN_USERNAME: admin\n LDAP_ADMIN_PASSWORD: adminpassword\n LDAP_USERS: customuser\n LDAP_PASSWORDS: custompassword\n LDAP_ENABLE_TLS: yes\n LDAP_TLS_CERT_FILE: /certificates/openldap.crt\n LDAP_TLS_KEY_FILE: /certificates/openldap.key\n LDAP_TLS_CA_FILE: /certificates/CA.crt\n BITNAMI_DEBUG: true\n LDAP_CONFIG_ADMIN_ENABLED: yes\n LDAP_CONFIG_ADMIN_USERNAME: admin\n LDAP_CONFIG_ADMIN_PASSWORD: configpassword\n\n---\nkind: pipeline\nname: package\n\nsteps:\n - name: packager\n image: joomlaprojects/docker-images:packager\n environment:\n HTTP_ROOT: "https://artifacts.joomla.org/drone"\n DRONE_PULL_REQUEST: DRONE_PULL_REQUEST\n DRONE_COMMIT: DRONE_COMMIT\n commands:\n - /bin/drone_prepare_package.sh\n\n - name: upload\n image: joomlaprojects/docker-images:packager\n environment:\n package_key:\n from_secret: package_key\n package_user:\n from_secret: package_user\n package_host:\n from_secret: package_host\n package_root:\n from_secret: package_root\n GITHUB_TOKEN:\n from_secret: github_token\n commands:\n - mkdir -p ~/.ssh\n - eval $(ssh-agent -s)\n - echo -e "Host *\n\tStrictHostKeyChecking no\n\n" > ~/.ssh/config\n - echo "$package_key" > ~/.ssh/id_rsa\n - chmod 600 ~/.ssh/id_rsa\n - ssh-add\n - rclone config create package sftp host $package_host user $package_user port 22\n - rclone copy ./upload/ package:$package_root/$DRONE_REPO/$DRONE_BRANCH/$DRONE_PULL_REQUEST/downloads/$DRONE_BUILD_NUMBER\n - /bin/add_github_status.sh "Download" "Prebuilt packages are available for download." "https://artifacts.joomla.org/drone/${DRONE_REPO}/${DRONE_BRANCH}/${DRONE_PULL_REQUEST}/downloads/${DRONE_BUILD_NUMBER}"\n\ntrigger:\n repo:\n - joomla/joomla-cms\n\n---\nkind: pipeline\nname: nightly_build\n\nsteps:\n - name: prepare\n image: joomlaprojects/docker-images:packager\n commands:\n - export MINORVERSION=${DRONE_BRANCH%-*}\n - composer --version\n - mkdir -p transfer\n - date +%s > transfer/$MINORVERSION-time.txt\n - git rev-parse origin/$MINORVERSION-dev > transfer/$MINORVERSION.txt\n - php build/build.php --remote=origin/$MINORVERSION-dev --exclude-gzip --disable-patch-packages\n - mv build/tmp/packages/* transfer/\n - php build/build.php --remote=origin/$MINORVERSION-dev --exclude-zip --exclude-gzip --exclude-bzip2 --debug-build\n - mv build/tmp/packages/* transfer/\n\n - name: upload\n image: joomlaprojects/docker-images:packager\n environment:\n nightly_key:\n from_secret: nightly_key\n nightly_user:\n from_secret: nightly_user\n nightly_host:\n from_secret: nightly_host\n MATTERMOST_NIGHTLY_HOOK:\n from_secret: mattermost_nightly_hook\n commands:\n - export MINORVERSION=${DRONE_BRANCH%-*}\n - mkdir -p ~/.ssh\n - eval $(ssh-agent -s)\n - echo -e "Host *\n\tStrictHostKeyChecking no\n\n" > ~/.ssh/config\n - echo "$nightly_key" > ~/.ssh/id_rsa\n - chmod 600 ~/.ssh/id_rsa\n - ssh-add\n - rclone config create nightly sftp host $nightly_host user $nightly_user port 22\n - rclone delete nightly:/home/devj/public_html/nightlies/ --include "Joomla_$MINORVERSION.*"\n - rclone delete nightly:/home/devj/public_html/cache/com_content/\n - rclone copy ./transfer/ nightly:/home/devj/public_html/nightlies/\n - curl -i -X POST -H 'Content-Type:application/json' -d '{"text":"Nightly Build for [Joomla ${MINORVERSION}](https://developer.joomla.org/nightly-builds.html) successfully built."}' $MATTERMOST_NIGHTLY_HOOK\n\n - name: buildfailure\n image: joomlaprojects/docker-images:packager\n environment:\n MATTERMOST_NIGHTLY_HOOK:\n from_secret: mattermost_nightly_hook\n commands:\n - export MINORVERSION=${DRONE_BRANCH%-*}\n - curl -i -X POST -H 'Content-Type:application/json' -d '{"text":"Nightly Build for [Joomla ${MINORVERSION}](https://developer.joomla.org/nightly-builds.html) FAILED to built."}' $MATTERMOST_NIGHTLY_HOOK\n when:\n status:\n - failure\n\ntrigger:\n event:\n - cron\n - custom\n repo:\n - joomla/joomla-cms\n\n---\nkind: signature\nhmac: 99da78521b10f37ddf6731e695fe52f85ee3db142bccac66901ae5372e332aae\n\n...\n | dataset_sample\yaml\joomla_joomla-cms\.drone.yml | .drone.yml | YAML | 12,745 | 0.8 | 0.007143 | 0 | react-lib | 786 | 2024-01-02T02:10:44.985538 | Apache-2.0 | false | 51e02f7ad018bfc15e01c0f24fd130c0 |
version: '{branch}_{build}'\npull_requests:\n do_not_increment_build_number: true\nbranches:\n only:\n - master\nskip_non_tags: true\nimage: Visual Studio 2019\nshallow_clone: true\nenvironment:\n LicensedDependenciesUrl:\n secure: ZTSg1DK6QAtwZoCBUDQ94cIaMCKWgSeJMI0dCpVJBPE40+UqG9MNu+UJjjGWjvVa\n OnlineInstallerConfig: https://playnite.link/build/installer_mirrors.txt\nbuild_script:\n- pwsh: .\build\build.ps1 -Package -SdkNuget -LicensedDependenciesUrl $env:LicensedDependenciesUrl -OnlineInstallerConfig $env:OnlineInstallerConfig\ntest: off\nartifacts:\n- path: 'build\Playnite.zip'\ndeploy:\n- provider: Webhook\n url: https://app.signpath.io/API/v1/4028e7e6-f183-4bbf-9640-c4b4d9a6992d/Integrations/AppVeyor?ProjectSlug=Playnite&SigningPolicySlug=release-signing\n authorization:\n secure: ZXw8Ro7eQ0opsdjLpNH7ZgTgS8ycnud7ag2qNOMRvqdxBmvegj+7QHrsh4pSgTSINhB5tbTZwgiMKFDjJV9QHg== | dataset_sample\yaml\JosefNemec_Playnite\appveyor.yml | appveyor.yml | YAML | 878 | 0.8 | 0 | 0 | awesome-app | 313 | 2025-04-20T23:07:13.339270 | BSD-3-Clause | false | 2835fcad08491f4594fbd8130865efab |
files:\n - source: LocSource.xaml\n translation: /%original_path%/%locale_with_underscore%.xaml\n | dataset_sample\yaml\JosefNemec_Playnite\crowdin.yml | crowdin.yml | YAML | 98 | 0.5 | 0 | 0 | react-lib | 752 | 2024-10-24T18:04:57.827018 | BSD-3-Clause | false | 4630c73f66ef211a87372f7a86823b84 |
language: php\n\nphp:\n- 5.4\n- 5.5\n- 5.6\n- 7.0\n- 7.1\n\nbefore_script:\n- travis_retry composer self-update\n- travis_retry composer install --no-interaction --prefer-source\n- npm install -g grunt-cli\n- npm install grunt-contrib-qunit\n- npm install grunt\n\nscript:\n- vendor/bin/phpunit -c lib/php/phpunit.xml.dist\n- npm test\n | dataset_sample\yaml\joypixels_emojione\.travis.yml | .travis.yml | YAML | 317 | 0.7 | 0 | 0 | react-lib | 654 | 2023-08-09T15:32:08.462206 | GPL-3.0 | false | 7555a70651711c6d0cd0182e38937bbb |
version: "3.8"\n\nvolumes:\n cargo_cache:\n pg_data:\n router_build_cache:\n scheduler_build_cache:\n drainer_build_cache:\n redisinsight_store:\n\nnetworks:\n router_net:\n\nservices:\n ### Dependencies\n pg:\n image: postgres:latest\n ports:\n - "5432:5432"\n networks:\n - router_net\n volumes:\n - pg_data:/VAR/LIB/POSTGRESQL/DATA\n environment:\n - POSTGRES_USER=db_user\n - POSTGRES_PASSWORD=db_pass\n - POSTGRES_DB=hyperswitch_db\n healthcheck:\n test: ["CMD-SHELL", "pg_isready -d $${POSTGRES_DB} -U $${POSTGRES_USER}"]\n interval: 5s\n retries: 3\n start_period: 5s\n timeout: 5s\n\n redis-standalone:\n image: redis:7\n networks:\n - router_net\n ports:\n - "6379:6379"\n healthcheck:\n test: ["CMD-SHELL", "redis-cli ping | grep '^PONG$'"]\n interval: 5s\n retries: 3\n start_period: 5s\n timeout: 5s\n\n migration_runner:\n image: rust:latest\n command: "bash -c 'cargo install diesel_cli --no-default-features --features postgres && cargo install just && just migrate'"\n working_dir: /app\n networks:\n - router_net\n volumes:\n - ./:/app\n environment:\n # format -> postgresql://DB_USER:DB_PASSWORD@HOST:PORT/DATABASE_NAME\n - DATABASE_URL=postgresql://db_user:db_pass@pg:5432/hyperswitch_db\n\n ### Application services\n hyperswitch-server:\n build:\n dockerfile_inline: |\n FROM rust:latest\n RUN apt-get update && \\n apt-get install -y protobuf-compiler\n RUN rustup component add rustfmt clippy \n command: cargo run --bin router -- -f ./config/docker_compose.toml\n working_dir: /app\n ports:\n - "8080:8080"\n networks:\n - router_net\n volumes:\n - ./:/app\n - cargo_cache:/cargo_cache\n - router_build_cache:/cargo_build_cache\n environment:\n - CARGO_HOME=/cargo_cache\n - CARGO_TARGET_DIR=/cargo_build_cache\n depends_on:\n pg:\n condition: service_healthy\n redis-standalone:\n condition: service_healthy\n labels:\n logs: "promtail"\n healthcheck:\n test: curl --fail http://localhost:8080/health || exit 1\n interval: 120s\n retries: 4\n start_period: 20s\n timeout: 10s\n\n hyperswitch-producer:\n image: rust:latest\n command: cargo run --bin scheduler -- -f ./config/docker_compose.toml\n working_dir: /app\n networks:\n - router_net\n profiles:\n - scheduler\n volumes:\n - ./:/app\n - cargo_cache:/cargo_cache\n - scheduler_build_cache:/cargo_build_cache\n environment:\n - CARGO_HOME=/cargo_cache\n - CARGO_TARGET_DIR=/cargo_build_cache\n - SCHEDULER_FLOW=producer\n depends_on:\n hyperswitch-consumer:\n condition: service_healthy\n labels:\n logs: "promtail"\n\n hyperswitch-consumer:\n image: rust:latest\n command: cargo run --bin scheduler -- -f ./config/docker_compose.toml\n working_dir: /app\n networks:\n - router_net\n profiles:\n - scheduler\n volumes:\n - ./:/app\n - cargo_cache:/cargo_cache\n - scheduler_build_cache:/cargo_build_cache\n environment:\n - CARGO_HOME=/cargo_cache\n - CARGO_TARGET_DIR=/cargo_build_cache\n - SCHEDULER_FLOW=consumer\n depends_on:\n hyperswitch-server:\n condition: service_started\n labels:\n logs: "promtail"\n healthcheck:\n test: (ps -e | grep scheduler) || exit 1\n interval: 120s\n retries: 4\n start_period: 30s\n timeout: 10s\n\n hyperswitch-drainer:\n image: rust:latest\n command: cargo run --bin drainer -- -f ./config/docker_compose.toml\n working_dir: /app\n deploy:\n replicas: ${DRAINER_INSTANCE_COUNT:-1}\n networks:\n - router_net\n profiles:\n - full_kv\n volumes:\n - ./:/app\n - cargo_cache:/cargo_cache\n - drainer_build_cache:/cargo_build_cache\n environment:\n - CARGO_HOME=/cargo_cache\n - CARGO_TARGET_DIR=/cargo_build_cache\n restart: unless-stopped\n depends_on:\n hyperswitch-server:\n condition: service_started\n labels:\n logs: "promtail"\n\n ### Clustered Redis setup\n redis-cluster:\n image: redis:7\n deploy:\n replicas: ${REDIS_CLUSTER_COUNT:-3}\n command: redis-server /usr/local/etc/redis/redis.conf\n profiles:\n - clustered_redis\n volumes:\n - ./config/redis.conf:/usr/local/etc/redis/redis.conf\n networks:\n - router_net\n ports:\n - "6379"\n - "16379"\n\n redis-init:\n image: redis:7\n profiles:\n - clustered_redis\n depends_on:\n - redis-cluster\n networks:\n - router_net\n command: "bash -c 'export COUNT=${REDIS_CLUSTER_COUNT:-3}\n\n \ if [ $$COUNT -lt 3 ]\n\n \ then\n\n \ echo \"Minimum 3 nodes are needed for redis cluster\"\n\n \ exit 1\n\n \ fi\n\n \ HOSTS=\"\"\n\n \ for ((c=1; c<=$$COUNT;c++))\n\n \ do\n\n \ NODE=$COMPOSE_PROJECT_NAME-redis-cluster-$$c:6379\n\n \ echo $$NODE\n\n \ HOSTS=\"$$HOSTS $$NODE\"\n\n \ done\n\n \ echo Creating a cluster with $$HOSTS\n\n \ redis-cli --cluster create $$HOSTS --cluster-yes\n\n \ '"\n\n ### Monitoring\n grafana:\n image: grafana/grafana:latest\n ports:\n - "3000:3000"\n networks:\n - router_net\n profiles:\n - monitoring\n restart: unless-stopped\n environment:\n - GF_AUTH_ANONYMOUS_ORG_ROLE=Admin\n - GF_AUTH_ANONYMOUS_ENABLED=true\n - GF_AUTH_BASIC_ENABLED=false\n volumes:\n - ./config/grafana.ini:/etc/grafana/grafana.ini\n - ./config/grafana-datasource.yaml:/etc/grafana/provisioning/datasources/datasource.yml\n\n promtail:\n image: grafana/promtail:latest\n volumes:\n - ./logs:/var/log/router\n - ./config:/etc/promtail\n - /var/run/docker.sock:/var/run/docker.sock\n command: -config.file=/etc/promtail/promtail.yaml\n profiles:\n - monitoring\n networks:\n - router_net\n\n loki:\n image: grafana/loki:latest\n ports:\n - "3100"\n command: -config.file=/etc/loki/loki.yaml\n networks:\n - router_net\n profiles:\n - monitoring\n volumes:\n - ./config:/etc/loki\n\n otel-collector:\n image: otel/opentelemetry-collector-contrib:latest\n command: --config=/etc/otel-collector.yaml\n networks:\n - router_net\n profiles:\n - monitoring\n volumes:\n - ./config/otel-collector.yaml:/etc/otel-collector.yaml\n ports:\n - "4317"\n - "8888"\n - "8889"\n\n prometheus:\n image: prom/prometheus:latest\n networks:\n - router_net\n profiles:\n - monitoring\n volumes:\n - ./config/prometheus.yaml:/etc/prometheus/prometheus.yml\n ports:\n - "9090"\n restart: unless-stopped\n\n tempo:\n image: grafana/tempo:latest\n command: -config.file=/etc/tempo.yaml\n volumes:\n - ./config/tempo.yaml:/etc/tempo.yaml\n networks:\n - router_net\n profiles:\n - monitoring\n ports:\n - "3200" # tempo\n - "4317" # otlp grpc\n restart: unless-stopped\n\n redis-insight:\n image: redislabs/redisinsight:latest\n networks:\n - router_net\n profiles:\n - full_kv\n ports:\n - "8001:8001"\n volumes:\n - redisinsight_store:/db\n\n hyperswitch-control-center:\n image: docker.juspay.io/juspaydotin/hyperswitch-control-center:latest\n pull_policy: always\n networks:\n - router_net\n ports:\n - "9000:9000"\n environment:\n apiBaseUrl: http://localhost:8080\n sdkBaseUrl: http://localhost:9050/HyperLoader.js\n\n hyperswitch-web-sdk:\n build:\n dockerfile_inline: |\n FROM node:lts\n RUN git clone https://github.com/juspay/hyperswitch-web.git --depth 1\n WORKDIR hyperswitch-web\n RUN npm i --force\n command: bash -c 'npm run re:build && npx run webpack serve --config webpack.dev.js --host 0.0.0.0'\n ports:\n - "9050:9050"\n environment:\n sdkEnv: local\n envSdkUrl: http://localhost:9050\n envBackendUrl: http://localhost:8080\n envLoggingUrl: http://localhost:8207\n | dataset_sample\yaml\juspay_hyperswitch\docker-compose-development.yml | docker-compose-development.yml | YAML | 7,990 | 0.8 | 0.008746 | 0.016234 | python-kit | 428 | 2023-10-15T00:13:46.664965 | Apache-2.0 | false | 8837c8d80f1e346b06e11f74963f468b |
volumes:\n pg_data:\n redisinsight_store:\n ckh_data:\n\nnetworks:\n router_net:\n\nservices:\n ### Dependencies\n prestart-hook:\n image: curlimages/curl-base:latest\n container_name: prestart-hook\n entrypoint: ["/bin/sh", "-c", "apk add --no-cache bash && /bin/bash /prestart_hook.sh"]\n volumes:\n - ./scripts/prestart_hook.sh:/prestart_hook.sh\n networks:\n - router_net\n\n pg:\n image: postgres:latest\n ports:\n - "5432:5432"\n networks:\n - router_net\n volumes:\n - pg_data:/VAR/LIB/POSTGRESQL/DATA\n environment:\n - POSTGRES_USER=db_user\n - POSTGRES_PASSWORD=db_pass\n - POSTGRES_DB=hyperswitch_db\n healthcheck:\n test: ["CMD-SHELL", "pg_isready -d $${POSTGRES_DB} -U $${POSTGRES_USER}"]\n interval: 5s\n retries: 3\n start_period: 5s\n timeout: 5s\n\n redis-standalone:\n image: redis:7\n networks:\n - router_net\n ports:\n - "6379:6379"\n healthcheck:\n test: ["CMD-SHELL", "redis-cli ping | grep '^PONG$'"]\n interval: 5s\n retries: 3\n start_period: 5s\n timeout: 5s\n\n migration_runner:\n image: rust:latest\n command: "bash -c 'cargo install diesel_cli --no-default-features --features postgres && cargo install just && just migrate'"\n working_dir: /app\n networks:\n - router_net\n volumes:\n - ./:/app\n environment:\n # format -> postgresql://DB_USER:DB_PASSWORD@HOST:PORT/DATABASE_NAME\n - DATABASE_URL=postgresql://db_user:db_pass@pg:5432/hyperswitch_db\n \n mailhog:\n image: mailhog/mailhog\n networks:\n - router_net\n ports:\n - "1025:1025"\n - "8025:8025"\n\n ### Application services\n hyperswitch-server:\n image: docker.juspay.io/juspaydotin/hyperswitch-router:standalone\n pull_policy: always\n command: /local/bin/router -f /local/config/docker_compose.toml\n ports:\n - "8080:8080"\n networks:\n - router_net\n volumes:\n - ./config:/local/config\n - ./files:/local/bin/files\n depends_on:\n pg:\n condition: service_healthy\n redis-standalone:\n condition: service_healthy\n labels:\n logs: "promtail"\n healthcheck:\n test: curl --fail http://localhost:8080/health || exit 1\n interval: 5s\n retries: 3\n start_period: 5s\n timeout: 5s\n\n poststart-hook:\n image: curlimages/curl-base:latest\n container_name: poststart-hook\n depends_on:\n hyperswitch-server:\n condition: service_healthy # Ensures it only starts when `hyperswitch-server` is healthy\n entrypoint: ["/bin/sh", "-c", "apk add --no-cache bash jq && /bin/bash /poststart_hook.sh"]\n volumes:\n - ./scripts/poststart_hook.sh:/poststart_hook.sh\n networks:\n - router_net \n \n hyperswitch-producer:\n image: docker.juspay.io/juspaydotin/hyperswitch-producer:standalone\n pull_policy: always\n command: /local/bin/scheduler -f /local/config/docker_compose.toml\n networks:\n - router_net\n profiles:\n - scheduler\n volumes:\n - ./config:/local/config\n environment:\n - SCHEDULER_FLOW=producer\n depends_on:\n hyperswitch-consumer:\n condition: service_healthy\n labels:\n logs: "promtail"\n\n hyperswitch-consumer:\n image: docker.juspay.io/juspaydotin/hyperswitch-consumer:standalone\n pull_policy: always\n command: /local/bin/scheduler -f /local/config/docker_compose.toml\n networks:\n - router_net\n profiles:\n - scheduler\n volumes:\n - ./config:/local/config\n environment:\n - SCHEDULER_FLOW=consumer\n depends_on:\n hyperswitch-server:\n condition: service_healthy\n labels:\n logs: "promtail"\n healthcheck:\n test: (ps -e | grep scheduler) || exit 1\n interval: 10s\n retries: 3\n start_period: 5s\n timeout: 10s\n\n hyperswitch-drainer:\n image: docker.juspay.io/juspaydotin/hyperswitch-drainer:standalone\n pull_policy: always\n command: /local/bin/drainer -f /local/config/docker_compose.toml\n deploy:\n replicas: ${DRAINER_INSTANCE_COUNT:-1}\n networks:\n - router_net\n profiles:\n - full_kv\n volumes:\n - ./config:/local/config\n restart: unless-stopped\n depends_on:\n hyperswitch-server:\n condition: service_healthy\n labels:\n logs: "promtail"\n\n ### Web Client\n hyperswitch-web:\n image: docker.juspay.io/juspaydotin/hyperswitch-web:latest\n pull_policy: always\n ports:\n - "9050:9050"\n - "9060:9060"\n - "5252:5252"\n networks:\n - router_net\n depends_on:\n hyperswitch-server:\n condition: service_healthy\n environment:\n - HYPERSWITCH_PUBLISHABLE_KEY=placeholder_publishable_key\n - HYPERSWITCH_SECRET_KEY=placeholder_api_key\n - HYPERSWITCH_SERVER_URL=http://localhost:8080\n - HYPERSWITCH_SERVER_URL_FOR_DEMO_APP=http://hyperswitch-server:8080\n - HYPERSWITCH_CLIENT_URL=http://localhost:9050\n - SELF_SERVER_URL=http://localhost:5252\n - SDK_ENV=local\n - ENV_LOGGING_URL=http://localhost:3103\n - ENV_BACKEND_URL=http://localhost:8080\n labels:\n logs: "promtail"\n\n ### Control Center\n hyperswitch-control-center:\n image: docker.juspay.io/juspaydotin/hyperswitch-control-center:latest\n pull_policy: always\n ports:\n - "9000:9000"\n environment:\n - configPath=/tmp/dashboard-config.toml\n volumes:\n - ./config/dashboard.toml:/tmp/dashboard-config.toml\n depends_on:\n hyperswitch-server:\n condition: service_healthy\n hyperswitch-web:\n condition: service_started\n labels:\n logs: "promtail"\n\n ### Clustered Redis setup\n redis-cluster:\n image: redis:7\n deploy:\n replicas: ${REDIS_CLUSTER_COUNT:-3}\n command: redis-server /usr/local/etc/redis/redis.conf\n profiles:\n - clustered_redis\n volumes:\n - ./config/redis.conf:/usr/local/etc/redis/redis.conf\n networks:\n - router_net\n ports:\n - "6379"\n - "16379"\n\n redis-init:\n image: redis:7\n profiles:\n - clustered_redis\n depends_on:\n - redis-cluster\n networks:\n - router_net\n command: |-\n bash -c 'export COUNT=${REDIS_CLUSTER_COUNT:-3}\n if [ $$COUNT -lt 3 ]\n then\n echo \"Minimum 3 nodes are needed for redis cluster\"\n exit 1\n fi\n HOSTS=\"\"\n for ((c=1; c<=$$COUNT;c++))\n do\n NODE=$COMPOSE_PROJECT_NAME-redis-cluster-$$c:6379\n echo $$NODE\n HOSTS=\"$$HOSTS $$NODE\"\n done\n echo Creating a cluster with $$HOSTS\n redis-cli --cluster create $$HOSTS --cluster-yes\n '\n ### Monitoring\n grafana:\n image: grafana/grafana:latest\n ports:\n - "3000:3000"\n networks:\n - router_net\n profiles:\n - monitoring\n restart: unless-stopped\n environment:\n - GF_AUTH_ANONYMOUS_ORG_ROLE=Admin\n - GF_AUTH_ANONYMOUS_ENABLED=true\n - GF_AUTH_BASIC_ENABLED=false\n volumes:\n - ./config/grafana.ini:/etc/grafana/grafana.ini\n - ./config/grafana-datasource.yaml:/etc/grafana/provisioning/datasources/datasource.yml\n\n loki:\n image: grafana/loki:latest\n ports:\n - "3100"\n command: -config.file=/etc/loki/loki.yaml\n networks:\n - router_net\n profiles:\n - monitoring\n volumes:\n - ./config:/etc/loki\n\n otel-collector:\n image: otel/opentelemetry-collector-contrib:latest\n command: --config=/etc/otel-collector.yaml\n networks:\n - router_net\n profiles:\n - monitoring\n volumes:\n - ./config/otel-collector.yaml:/etc/otel-collector.yaml\n ports:\n - "4317"\n - "8888"\n - "8889"\n\n prometheus:\n image: prom/prometheus:latest\n networks:\n - router_net\n profiles:\n - monitoring\n volumes:\n - ./config/prometheus.yaml:/etc/prometheus/prometheus.yml\n ports:\n - "9090"\n restart: unless-stopped\n\n tempo:\n image: grafana/tempo:latest\n command: -config.file=/etc/tempo.yaml\n volumes:\n - ./config/tempo.yaml:/etc/tempo.yaml\n networks:\n - router_net\n profiles:\n - monitoring\n ports:\n - "3200" # tempo\n - "4317" # otlp grpc\n restart: unless-stopped\n\n redis-insight:\n image: redislabs/redisinsight:latest\n networks:\n - router_net\n profiles:\n - monitoring\n ports:\n - "8001:8001"\n volumes:\n - redisinsight_store:/db\n\n kafka0:\n image: confluentinc/cp-kafka:7.0.5\n hostname: kafka0\n networks:\n - router_net\n ports:\n - 9092:9092\n - 9093\n - 9997\n - 29092\n environment:\n KAFKA_BROKER_ID: 1\n KAFKA_LISTENER_SECURITY_PROTOCOL_MAP: PLAINTEXT:PLAINTEXT,CONTROLLER:PLAINTEXT,PLAINTEXT_HOST:PLAINTEXT\n KAFKA_ADVERTISED_LISTENERS: PLAINTEXT://kafka0:29092,PLAINTEXT_HOST://localhost:9092\n KAFKA_INTER_BROKER_LISTENER_NAME: PLAINTEXT\n KAFKA_OFFSETS_TOPIC_REPLICATION_FACTOR: 1\n KAFKA_GROUP_INITIAL_REBALANCE_DELAY_MS: 0\n KAFKA_TRANSACTION_STATE_LOG_MIN_ISR: 1\n KAFKA_TRANSACTION_STATE_LOG_REPLICATION_FACTOR: 1\n KAFKA_PROCESS_ROLES: "broker,controller"\n KAFKA_NODE_ID: 1\n KAFKA_CONTROLLER_QUORUM_VOTERS: "1@kafka0:29093"\n KAFKA_LISTENERS: "PLAINTEXT://kafka0:29092,CONTROLLER://kafka0:29093,PLAINTEXT_HOST://0.0.0.0:9092"\n KAFKA_CONTROLLER_LISTENER_NAMES: "CONTROLLER"\n KAFKA_LOG_DIRS: "/tmp/kraft-combined-logs"\n JMX_PORT: 9997\n KAFKA_JMX_OPTS: -Dcom.sun.management.jmxremote -Dcom.sun.management.jmxremote.authenticate=false -Dcom.sun.management.jmxremote.ssl=false -Djava.rmi.server.hostname=kafka0 -Dcom.sun.management.jmxremote.rmi.port=9997\n profiles:\n - olap\n volumes:\n - ./monitoring/kafka-script.sh:/tmp/update_run.sh\n command: 'bash -c ''if [ ! -f /tmp/update_run.sh ]; then echo "ERROR: Did you forget the update_run.sh file that came with this docker-compose.yml file?" && exit 1 ; else /tmp/update_run.sh && /etc/confluent/docker/run ; fi'''\n\n # Kafka UI for debugging kafka queues\n kafka-ui:\n image: provectuslabs/kafka-ui:latest\n ports:\n - 8090:8080\n networks:\n - router_net\n depends_on:\n - kafka0\n profiles:\n - olap\n environment:\n KAFKA_CLUSTERS_0_NAME: local\n KAFKA_CLUSTERS_0_BOOTSTRAPSERVERS: kafka0:29092\n KAFKA_CLUSTERS_0_JMXPORT: 9997\n\n clickhouse-server:\n image: clickhouse/clickhouse-server:24.3\n networks:\n - router_net\n ports:\n - "9000"\n - "8123:8123"\n volumes:\n - ./crates/analytics/docs/clickhouse/scripts:/docker-entrypoint-initdb.d\n environment:\n - TZ=Asia/Kolkata\n profiles:\n - olap\n ulimits:\n nofile:\n soft: 262144\n hard: 262144\n\n opensearch:\n image: opensearchproject/opensearch:2\n container_name: opensearch\n hostname: opensearch\n environment:\n - "discovery.type=single-node"\n - OPENSEARCH_INITIAL_ADMIN_PASSWORD=0penS3arc#\n - LOG_LEVEL=DEBUG\n profiles:\n - olap\n ports:\n - "9200:9200"\n networks:\n - router_net\n\n opensearch-dashboards:\n image: opensearchproject/opensearch-dashboards:2\n ports:\n - 5601:5601\n profiles:\n - olap\n environment:\n OPENSEARCH_HOSTS: '["https://opensearch:9200"]'\n networks:\n - router_net\n\n vector:\n image: timberio/vector:latest-debian\n ports:\n - "8686"\n - "9598"\n - "3103:3103"\n profiles:\n - olap\n environment:\n KAFKA_HOST: "kafka0:29092"\n networks:\n - router_net\n volumes:\n - ./config/vector.yaml:/etc/vector/vector.yaml\n - /var/run/docker.sock:/var/run/docker.sock\n | dataset_sample\yaml\juspay_hyperswitch\docker-compose.yml | docker-compose.yml | YAML | 11,548 | 0.8 | 0.011261 | 0.019139 | react-lib | 569 | 2024-12-15T16:03:02.051263 | Apache-2.0 | false | 2ff8994d03805cf413be97baa7958ebf |
"base_path": ".."\n"preserve_hierarchy": true\n"files": [\n {\n "source": "/ArchiSteamFarm/Localization/Strings.resx",\n "translation": "/ArchiSteamFarm/Localization/Strings.%locale%.resx",\n "translation_replace": {\n ".bs-BA.resx": ".bs-Latn.resx",\n ".lol-US.resx": ".qps-Ploc.resx",\n ".sr-CS.resx": ".sr-Latn.resx",\n ".zh-CN.resx": ".zh-Hans.resx",\n ".zh-HK.resx": ".zh-Hant-HK.resx",\n ".zh-TW.resx": ".zh-Hant.resx"\n }\n },\n {\n "source": "/ArchiSteamFarm.OfficialPlugins.ItemsMatcher/Localization/Strings.resx",\n "translation": "/ArchiSteamFarm.OfficialPlugins.ItemsMatcher/Localization/Strings.%locale%.resx",\n "translation_replace": {\n ".bs-BA.resx": ".bs-Latn.resx",\n ".lol-US.resx": ".qps-Ploc.resx",\n ".sr-CS.resx": ".sr-Latn.resx",\n ".zh-CN.resx": ".zh-Hans.resx",\n ".zh-HK.resx": ".zh-Hant-HK.resx",\n ".zh-TW.resx": ".zh-Hant.resx"\n }\n },\n {\n "source": "/ArchiSteamFarm.OfficialPlugins.MobileAuthenticator/Localization/Strings.resx",\n "translation": "/ArchiSteamFarm.OfficialPlugins.MobileAuthenticator/Localization/Strings.%locale%.resx",\n "translation_replace": {\n ".bs-BA.resx": ".bs-Latn.resx",\n ".lol-US.resx": ".qps-Ploc.resx",\n ".sr-CS.resx": ".sr-Latn.resx",\n ".zh-CN.resx": ".zh-Hans.resx",\n ".zh-HK.resx": ".zh-Hant-HK.resx",\n ".zh-TW.resx": ".zh-Hant.resx"\n }\n },\n {\n "source": "/ArchiSteamFarm.OfficialPlugins.SteamTokenDumper/Localization/Strings.resx",\n "translation": "/ArchiSteamFarm.OfficialPlugins.SteamTokenDumper/Localization/Strings.%locale%.resx",\n "translation_replace": {\n ".bs-BA.resx": ".bs-Latn.resx",\n ".lol-US.resx": ".qps-Ploc.resx",\n ".sr-CS.resx": ".sr-Latn.resx",\n ".zh-CN.resx": ".zh-Hans.resx",\n ".zh-HK.resx": ".zh-Hant-HK.resx",\n ".zh-TW.resx": ".zh-Hant.resx"\n }\n },\n {\n "source": "/wiki/*.md",\n "translation": "/wiki/locale/%locale%/%file_name%-%locale%.%file_extension%",\n "ignore": [ "/**/_*.md" ]\n },\n {\n "source": "/wiki/_Sidebar.md",\n "translation": "/wiki/locale/%locale%/_Sidebar.md"\n }\n]\n | dataset_sample\yaml\JustArchiNET_ArchiSteamFarm\.github\crowdin.yml | crowdin.yml | YAML | 2,133 | 0.8 | 0 | 0 | react-lib | 848 | 2025-03-27T05:18:37.299833 | GPL-3.0 | false | fd46b87245f7efe92bbc720f3666519f |
github: JustArchi\ncustom: ["https://paypal.me/JustArchi", "https://www.paypal.com/cgi-bin/webscr?cmd=_s-xclick&hosted_button_id=HD2P2P3WGS5Y4", "https://pay.revolut.com/justarchi", "https://steamcommunity.com/tradeoffer/new/?partner=46697991&token=0ix2Ruv_"]\n | dataset_sample\yaml\JustArchiNET_ArchiSteamFarm\.github\FUNDING.yml | FUNDING.yml | YAML | 259 | 0.8 | 0 | 0 | vue-tools | 863 | 2025-05-16T22:50:37.144850 | GPL-3.0 | false | 2beaa07fb1f8cc08dc4245131de01ec7 |
name: 🐛 Bug report\ndescription: Unexpected program behaviour that requires code correction\nlabels: ["🐛 Bug", "👀 Evaluation"]\nbody:\n - type: checkboxes\n id: checklist\n attributes:\n label: Checklist\n description: Ensure that our bug report form is appropriate for you.\n options:\n - label: I read and understood ASF's **[Contributing guidelines](https://github.com/JustArchiNET/ArchiSteamFarm/blob/main/.github/CONTRIBUTING.md)**\n required: true\n - label: I also read **[Setting-up](https://github.com/JustArchiNET/ArchiSteamFarm/wiki/Setting-up)** and **[FAQ](https://github.com/JustArchiNET/ArchiSteamFarm/wiki/FAQ)**, I don't need **[help](https://github.com/JustArchiNET/ArchiSteamFarm/blob/main/.github/SUPPORT.md)**, this is a bug report\n required: true\n - label: This is not a **[duplicate](https://github.com/JustArchiNET/ArchiSteamFarm/issues?q=is%3Aissue)** of an existing issue\n required: true\n - label: I don't own more than **[10 accounts in total](https://github.com/JustArchiNET/ArchiSteamFarm/wiki/FAQ#how-many-bots-can-i-run-with-asf)**\n required: true\n - label: I'm not using **[custom plugins](https://github.com/JustArchiNET/ArchiSteamFarm/wiki/Plugins)**\n required: true\n - label: This is not a **[question](https://github.com/JustArchiNET/ArchiSteamFarm/discussions)**\n required: true\n - label: This is not a **[technical issue](https://github.com/JustArchiNET/ArchiSteamFarm/discussions)**\n required: true\n - label: This is not **[ASF-ui problem](https://github.com/JustArchiNET/ASF-ui/issues/new/choose)**\n required: true\n - type: dropdown\n id: version\n attributes:\n label: ASF version\n description: If you're using a different version than the current **[stable](https://github.com/JustArchiNET/ArchiSteamFarm/releases/latest)** or **[pre-release](https://github.com/JustArchiNET/ArchiSteamFarm/releases)**, ensure that your bug report is reproducible on one of the below.\n options:\n - Latest stable release\n - Latest pre-release\n validations:\n required: true\n - type: dropdown\n id: variant\n attributes:\n label: ASF variant\n description: If you're using a different variant, ensure that your bug report is reproducible on one of the below.\n options:\n - docker-linux/amd64\n - docker-linux/arm/v7\n - docker-linux/arm64\n - generic (with latest .NET runtime)\n - linux-arm\n - linux-arm64\n - linux-x64\n - osx-arm64\n - osx-x64\n - win-arm64\n - win-x64\n validations:\n required: true\n - type: textarea\n id: bug-description\n attributes:\n label: Bug description\n description: Short explanation of what you were going to do, what did you want to accomplish?\n placeholder: |\n I tried to brew a coffee with ASF using `PUT /Api/Coffee` ASF API endpoint, but upon trying the program returned HTTP error: 418 I'm a teapot.\n validations:\n required: true\n - type: textarea\n id: expected-behavior\n attributes:\n label: Expected behavior\n description: What did you expect to happen?\n placeholder: |\n I expected my favourite latte macchiato in a cup put below the machine hosting ASF.\n validations:\n required: true\n - type: textarea\n id: actual-behavior\n attributes:\n label: Actual behavior\n description: What happened instead?\n placeholder: |\n No coffee was brewed, and so I was forced to use a water dispenser instead :/.\n validations:\n required: true\n - type: textarea\n id: steps-to-reproduce\n attributes:\n label: Steps to reproduce\n description: |\n Every command or action that happened after launching ASF, which leads to the bug.\n If launching ASF with provided configs (below) is everything that is needed, then this section is not mandatory.\n Screenshots of the problem and/or steps leading to it could be very useful in particular.\n placeholder: |\n 1. Put cup below the machine hosting ASF.\n 2. Send `PUT /Api/Coffee` request selecting latte macchiato.\n 3. No coffee was brewed.\n - type: textarea\n id: possible-solution\n attributes:\n label: Possible reason/solution\n description: |\n Not mandatory, but you can suggest a fix/reason for the bug, if known to you.\n If you observed something peculiar about your issue that could help us locate and fix the culprit, this is the right place.\n placeholder: |\n Perhaps no coffee was brewed because I was out of milk?\n - type: dropdown\n id: help\n attributes:\n label: Can you help us with this bug report?\n description: |\n ASF is offered for free and our resources are limited.\n Helping us increases the chance of fixing the problem.\n options:\n - Yes, I can code the solution myself and send a pull request\n - Somehow, I can test and offer feedback, but can't code\n - No, I don't have time, skills or willings for any of that\n validations:\n required: true\n - type: textarea\n id: asf-log\n attributes:\n label: Full log.txt recorded during reproducing the problem\n description: |\n You can find `log.txt` file directly in ASF directory.\n If the bug report doesn't come from the last run of the program, you can find logs from previous runs of the program in the `logs` directory instead.\n If no `log.txt` was recorded due to crash at the very early stage, console output should be pasted instead.\n placeholder: |\n 2021-12-16 00:20:43|dotnet-282887|INFO|ASF|InitCore() ArchiSteamFarm V5.2.1.2 (generic/6b492ffa-9927-431d-bae7-7360ab9968a9 | .NET 6.0.0-rtm.21522.10; debian-arm64; Linux 5.15.0-1-arm64 #1 SMP Debian 5.15.3-1 (2021-11-18))\n 2021-12-16 00:20:43|dotnet-282887|INFO|ASF|InitCore() Copyright © 2015-2021 JustArchiNET\n 2021-12-16 00:20:47|dotnet-282887|INFO|ASF|InitPlugins() Initializing Plugins...\n 2021-12-16 00:20:47|dotnet-282887|INFO|ASF|InitPlugins() Loading SteamTokenDumperPlugin V5.2.1.2...\n 2021-12-16 00:20:47|dotnet-282887|INFO|ASF|InitPlugins() SteamTokenDumperPlugin has been loaded successfully!\n 2021-12-16 00:20:47|dotnet-282887|INFO|ASF|UpdateAndRestart() ASF will automatically check for new versions every 1 day.\n 2021-12-16 00:20:52|dotnet-282887|INFO|ASF|Update() Checking for new version...\n 2021-12-16 00:20:54|dotnet-282887|INFO|ASF|Update() Local version: 5.2.1.2 | Remote version: 5.2.1.2\n 2021-12-16 00:20:54|dotnet-282887|INFO|ASF|Load() Loading STD global cache...\n 2021-12-16 00:20:56|dotnet-282887|INFO|ASF|Load() Validating STD global cache integrity...\n 2021-12-16 00:20:56|dotnet-282887|INFO|ASF|OnASFInit() SteamTokenDumperPlugin has been initialized successfully, thank you in advance for your help. The first submission will happen in approximately 47 minutes from now.\n 2021-12-16 00:20:57|dotnet-282887|INFO|ASF|Start() Starting IPC server...\n 2021-12-16 00:20:59|dotnet-282887|INFO|ASF|Start() IPC server ready!\n render: text\n validations:\n required: true\n - type: textarea\n id: global-config\n attributes:\n label: Global ASF.json config file\n description: |\n The config can be found in `config` directory under `ASF.json` name.\n You can leave this field empty if not using one.\n\n Ensure that your config has redacted (but NOT removed) potentially-sensitive properties, such as:\n - IPCPassword (recommended)\n - LicenseID (mandatory)\n - SteamOwnerID (optionally)\n - WebProxy (optionally, if exposing private details)\n - WebProxyPassword (optionally, if exposing private details)\n - WebProxyUsername (optionally, if exposing private details)\n\n Redacting involves replacing sensitive details, for example with stars (***). You should refrain from removing config lines entirely, as their pure existence may be relevant and should be preserved.\n placeholder: |\n {\n "AutoRestart": false,\n "Headless": true,\n "IPCPassword": "********",\n "UpdateChannel": 2,\n "SteamTokenDumperPluginEnabled": true\n }\n render: json\n - type: textarea\n id: bot-config\n attributes:\n label: BotName.json config of all affected bot instances\n description: |\n Bot config files can be found in `config` directory, ending with `json` extension.\n You can leave this field empty if you don't have any defined.\n\n Ensure that your config has redacted (but NOT removed) potentially-sensitive properties, such as:\n - SteamLogin (mandatory)\n - SteamPassword (mandatory)\n - SteamMasterClanID (optionally)\n - SteamParentalCode (optionally)\n - SteamTradeToken (optionally)\n - SteamUserPermissions (optionally, only SteamIDs)\n\n Redacting involves replacing sensitive details, for example with stars (***). You should refrain from removing config lines entirely, as their pure existence may be relevant and should be preserved.\n placeholder: |\n {\n "Enabled": true,\n "SteamLogin": "********",\n "SteamPassword": "********"\n }\n render: json\n - type: textarea\n id: additional-info\n attributes:\n label: Additional info\n description: Everything else you consider worthy that we didn't ask for.\n - type: markdown\n attributes:\n value: |\n ---\n #### Thank you for taking the time to fill out this bug report.\n | dataset_sample\yaml\JustArchiNET_ArchiSteamFarm\.github\ISSUE_TEMPLATE\Bug-report.yml | Bug-report.yml | YAML | 9,722 | 0.95 | 0.084577 | 0.005076 | python-kit | 251 | 2024-08-28T06:19:08.009999 | GPL-3.0 | false | 4c145f565d8ac5547757f27f4c504941 |
blank_issues_enabled: false\ncontact_links:\n - name: 🖥️ ASF-ui issue\n url: https://github.com/JustArchiNET/ASF-ui/issues/new/choose\n about: Please open issue in ASF-ui repo\n - name: 🌐 Localization improvement\n url: https://github.com/JustArchiNET/ArchiSteamFarm/wiki/Localization\n about: Please use our crowdin platform\n - name: ❓ Support question or technical issue\n url: https://github.com/JustArchiNET/ArchiSteamFarm/discussions/categories/support-english\n about: Please use GitHub discussions or other support channels\n - name: 🗯️ Negative feedback, complaints and demands\n url: https://www.youtube.com/watch?v=dQw4w9WgXcQ\n about: We're taking those very seriously\n | dataset_sample\yaml\JustArchiNET_ArchiSteamFarm\.github\ISSUE_TEMPLATE\config.yml | config.yml | YAML | 711 | 0.8 | 0 | 0 | python-kit | 1,000 | 2024-09-02T21:26:10.927555 | GPL-3.0 | false | 19123b244aed73ef87d563c0e3da0381 |
name: ✨ Enhancement idea\ndescription: General idea for improving the project\nlabels: ["✨ Enhancement", "👀 Evaluation"]\nbody:\n - type: checkboxes\n id: checklist\n attributes:\n label: Checklist\n description: Ensure that our enhancement idea form is appropriate for you.\n options:\n - label: I read and understood ASF's **[Contributing guidelines](https://github.com/JustArchiNET/ArchiSteamFarm/blob/main/.github/CONTRIBUTING.md)**\n required: true\n - label: I also read **[Setting-up](https://github.com/JustArchiNET/ArchiSteamFarm/wiki/Setting-up)** and **[FAQ](https://github.com/JustArchiNET/ArchiSteamFarm/wiki/FAQ)**, I don't need **[help](https://github.com/JustArchiNET/ArchiSteamFarm/blob/main/.github/SUPPORT.md)**, this is an enhancement idea\n required: true\n - label: This is not a **[duplicate](https://github.com/JustArchiNET/ArchiSteamFarm/issues?q=is%3Aissue)** of an existing issue\n required: true\n - label: My idea doesn't duplicate existing ASF functionality described on the **[wiki](https://github.com/JustArchiNET/ArchiSteamFarm/wiki)**\n required: true\n - label: I believe that my idea falls into ASF's scope and should be offered as part of ASF built-in functionality\n required: true\n - label: My idea doesn't violate the **[Steam Subscriber Agreement](https://store.steampowered.com/subscriber_agreement)**\n required: true\n - label: My idea doesn't violate the **[Steam Online Conduct](https://store.steampowered.com/online_conduct)**\n required: true\n - label: This is not **[ASF-ui suggestion](https://github.com/JustArchiNET/ASF-ui/issues/new/choose)**\n required: true\n - type: textarea\n id: enhancement-purpose\n attributes:\n label: Enhancement purpose\n description: |\n Purpose of the enhancement - if it solves some problem, precise in particular which. If it benefits the program in some other way, precise in particular why.\n Present the underlying reason why this enhancement makes sense, and what is the context of it.\n placeholder: |\n As of today ASF offers variety of beverages, such as latte macchiato or cappuccino. I'd appreciate if ASF offered some no-milk options as well, for example espresso or ristretto.\n I believe it'd further improve the program offering the users wider selection, which is very convenient.\n validations:\n required: true\n - type: textarea\n id: solution\n attributes:\n label: Solution\n description: What would you like to see as a solution to the purpose specified by you above?\n placeholder: |\n Simply add an option to brew some no-milk types of coffee. The existing logic is fine, we just need wider choice!\n validations:\n required: true\n - type: textarea\n id: why-existing-not-sufficient\n attributes:\n label: Why currently available solutions are not sufficient?\n description: |\n Evaluate the existing solutions in regards to your requirements.\n If something you're suggesting is already possible, then explain to us why the currently available solutions are not sufficient.\n If it's not possible yet, then explain to us why it should be.\n placeholder: |\n I'm allergic to milk, there is currently no option to pick a beverage that doesn't include it.\n Temporarily I'm switching cup mid-brewing as a workaround, but that is suboptimal considering the milk wasted.\n validations:\n required: true\n - type: dropdown\n id: help\n attributes:\n label: Can you help us with this enhancement idea?\n description: |\n ASF is offered for free and our resources are limited.\n Helping us increases the chance of making it happen.\n options:\n - Yes, I can code the solution myself and send a pull request\n - Somehow, I can test and offer feedback, but can't code\n - No, I don't have time, skills or willings for any of that\n validations:\n required: true\n - type: textarea\n id: additional-info\n attributes:\n label: Additional info\n description: Everything else you consider worthy that we didn't ask for.\n - type: markdown\n attributes:\n value: |\n ---\n #### Thank you for taking the time to fill out this enhancement idea.\n | dataset_sample\yaml\JustArchiNET_ArchiSteamFarm\.github\ISSUE_TEMPLATE\Enhancement-idea.yml | Enhancement-idea.yml | YAML | 4,379 | 0.95 | 0.108434 | 0.012048 | react-lib | 497 | 2025-06-03T10:01:53.033398 | BSD-3-Clause | false | ba9c4e391d96e6807d01bea3c74947cb |
name: 📕 Wiki suggestion\ndescription: All issues related to our wiki documentation, mainly corrections and ideas\nlabels: ["📕 Wiki", "👀 Evaluation"]\nbody:\n - type: checkboxes\n id: checklist\n attributes:\n label: Checklist\n description: Ensure that our wiki suggestion form is appropriate for you.\n options:\n - label: I read and understood ASF's **[Contributing guidelines](https://github.com/JustArchiNET/ArchiSteamFarm/blob/main/.github/CONTRIBUTING.md)**\n required: true\n - label: I also read **[Setting-up](https://github.com/JustArchiNET/ArchiSteamFarm/wiki/Setting-up)** and **[FAQ](https://github.com/JustArchiNET/ArchiSteamFarm/wiki/FAQ)**, I don't need **[help](https://github.com/JustArchiNET/ArchiSteamFarm/blob/main/.github/SUPPORT.md)**, this is a wiki suggestion\n required: true\n - label: This is not a **[translation issue](https://github.com/JustArchiNET/ArchiSteamFarm/wiki/Localization)**\n required: true\n - label: This is not **[ASF-ui wiki suggestion](https://github.com/JustArchiNET/ASF-ui/issues/new/choose)**\n required: true\n - type: input\n id: wiki-page\n attributes:\n label: Wiki page\n description: |\n If this is a suggestion regarding an existing wiki page, please link it for reference.\n If the wiki page doesn't exist, suggest its title.\n placeholder: https://github.com/JustArchiNET/ArchiSteamFarm/wiki/???\n validations:\n required: true\n - type: textarea\n id: issue\n attributes:\n label: The issue\n description: |\n Please specify your issue in regards to our wiki documentation.\n If you're reporting a mistake/correction, state what is wrong.\n If you're suggesting an idea, explain the details.\n placeholder: |\n As of today the wiki doesn't explain how to sing famous song composed by Rick Astley - Never Gonna Give You Up.\n I'm sick of googling the lyrics every time I'm opening a complaint on your GitHub, so please consider just adding it along the other stuff.\n validations:\n required: true\n - type: textarea\n id: wrong-text\n attributes:\n label: Wrong text\n description: |\n The existing text on the wiki which you classify as wrong.\n If you're suggesting a new page, paragraph or other addition to the wiki, then this section is not mandatory.\n placeholder: |\n Lack of song lyrics is what's wrong!\n render: markdown\n - type: textarea\n id: suggested-improvement\n attributes:\n label: Suggested improvement\n description: |\n The new or corrected text that would satisfy your issue stated above.\n You may use **[markdown](https://guides.github.com/features/mastering-markdown)** for formatting.\n placeholder: |\n # Never Gonna Give You Up by Rick Astley\n\n ## Verse 1\n We're no strangers to love\n You know the rules and so do I\n A full commitment's what I'm thinking of\n You wouldn't get this from any other guy\n\n ## Pre-Chorus\n I just wanna tell you how I'm feeling\n Gotta make you understand\n\n ## Chorus\n Never gonna give you up\n Never gonna let you down\n Never gonna run around and desert you\n Never gonna make you cry\n Never gonna say goodbye\n Never gonna tell a lie and hurt you\n\n ## More\n See **[full version](https://www.youtube.com/watch?v=dQw4w9WgXcQ)**.\n render: markdown\n validations:\n required: true\n - type: textarea\n id: additional-info\n attributes:\n label: Additional info\n description: Everything else you consider worthy that we didn't ask for.\n - type: markdown\n attributes:\n value: |\n ---\n #### Thank you for taking the time to fill out this wiki suggestion.\n | dataset_sample\yaml\JustArchiNET_ArchiSteamFarm\.github\ISSUE_TEMPLATE\Wiki-suggestion.yml | Wiki-suggestion.yml | YAML | 3,865 | 0.95 | 0.053191 | 0.066667 | python-kit | 616 | 2025-05-03T10:10:22.048338 | MIT | false | c2d908c3df3c2334406712fd83ff9a4f |
name: ASF-ci\n\non: [push, pull_request]\n\nenv:\n DOTNET_CLI_TELEMETRY_OPTOUT: true\n DOTNET_NOLOGO: true\n DOTNET_SDK_VERSION: 9.0\n\npermissions: {}\n\njobs:\n main:\n strategy:\n fail-fast: false\n matrix:\n configuration: [Debug, Release]\n os: [macos-latest, ubuntu-latest, windows-latest]\n\n runs-on: ${{ matrix.os }}\n\n steps:\n - name: Checkout code\n uses: actions/checkout@v4.2.2\n with:\n show-progress: false\n submodules: recursive\n\n - name: Setup .NET Core\n uses: actions/setup-dotnet@v4.3.1\n with:\n dotnet-version: ${{ env.DOTNET_SDK_VERSION }}\n\n - name: Verify .NET Core\n run: dotnet --info\n\n - name: Build ${{ matrix.configuration }} ArchiSteamFarm and other projects\n run: dotnet build -c "${{ matrix.configuration }}" -p:ContinuousIntegrationBuild=true -p:UseAppHost=false --nologo\n\n - name: Run ${{ matrix.configuration }} ArchiSteamFarm.Tests\n run: dotnet test ArchiSteamFarm.Tests -c "${{ matrix.configuration }}" -p:ContinuousIntegrationBuild=true -p:UseAppHost=false --nologo\n | dataset_sample\yaml\JustArchiNET_ArchiSteamFarm\.github\workflows\ci.yml | ci.yml | YAML | 1,085 | 0.7 | 0 | 0 | vue-tools | 174 | 2025-01-12T06:05:09.638294 | BSD-3-Clause | false | ee023b21b1a51570e20e1d1c8ccb46f1 |
name: ASF-crowdin-ci\n\non:\n push:\n branches:\n - main\n\npermissions: {}\n\njobs:\n upload:\n environment: dev-crowdin\n runs-on: ubuntu-latest\n\n steps:\n - name: Checkout code\n uses: actions/checkout@v4.2.2\n with:\n show-progress: false\n submodules: recursive\n\n - name: Upload latest strings for translation on Crowdin\n uses: crowdin/github-action@v2.7.0\n with:\n crowdin_branch_name: main\n config: '.github/crowdin.yml'\n project_id: ${{ secrets.ASF_CROWDIN_PROJECT_ID }}\n token: ${{ secrets.ASF_CROWDIN_API_TOKEN }}\n | dataset_sample\yaml\JustArchiNET_ArchiSteamFarm\.github\workflows\crowdin-ci.yml | crowdin-ci.yml | YAML | 591 | 0.7 | 0.035714 | 0 | node-utils | 95 | 2023-12-07T09:29:19.003671 | GPL-3.0 | false | 199f42aaf55989f3fdca0b6603b61565 |
name: ASF-docker-ci\n\non: [push, pull_request]\n\nenv:\n PLATFORMS: linux/amd64,linux/arm,linux/arm64\n\npermissions: {}\n\njobs:\n main:\n strategy:\n fail-fast: false\n matrix:\n configuration: [Debug, Release]\n file: [Dockerfile, Dockerfile.Service]\n\n runs-on: ubuntu-latest\n\n steps:\n - name: Checkout code\n uses: actions/checkout@v4.2.2\n with:\n show-progress: false\n submodules: recursive\n\n - name: Set up Docker Buildx\n uses: docker/setup-buildx-action@v3.10.0\n\n - name: Build ${{ matrix.configuration }} Docker image from ${{ matrix.file }}\n uses: docker/build-push-action@v6.15.0\n with:\n build-args: CONFIGURATION=${{ matrix.configuration }}\n context: .\n file: ${{ matrix.file }}\n platforms: ${{ env.PLATFORMS }}\n provenance: true\n sbom: true\n | dataset_sample\yaml\JustArchiNET_ArchiSteamFarm\.github\workflows\docker-ci.yml | docker-ci.yml | YAML | 862 | 0.7 | 0 | 0 | vue-tools | 445 | 2024-03-04T10:52:25.699025 | Apache-2.0 | false | 956f208f15886e218d78f66afa9a0d8f |
name: ASF-docker-publish-main\n\non:\n push:\n branches:\n - main\n\nenv:\n PLATFORMS: linux/amd64,linux/arm,linux/arm64\n TAG: main\n\npermissions:\n packages: write\n\njobs:\n main:\n environment: release-docker\n runs-on: ubuntu-latest\n\n steps:\n - name: Checkout code\n uses: actions/checkout@v4.2.2\n with:\n show-progress: false\n submodules: recursive\n\n - name: Set up Docker Buildx\n uses: docker/setup-buildx-action@v3.10.0\n\n - name: Login to ghcr.io\n uses: docker/login-action@v3.4.0\n with:\n registry: ghcr.io\n username: ${{ github.actor }}\n password: ${{ secrets.GITHUB_TOKEN }}\n\n - name: Login to DockerHub\n uses: docker/login-action@v3.4.0\n with:\n username: ${{ secrets.DOCKERHUB_USERNAME }}\n password: ${{ secrets.DOCKERHUB_TOKEN }}\n\n - name: Prepare environment outputs\n shell: sh\n run: |\n set -eu\n\n echo "DATE_ISO8601=$(date --iso-8601=seconds --utc)" >> "$GITHUB_ENV"\n echo "GHCR_REPOSITORY=$(echo ${{ github.repository }} | tr '[:upper:]' '[:lower:]')" >> "$GITHUB_ENV"\n echo "DH_REPOSITORY=$(echo ${{ secrets.DOCKERHUB_USERNAME }}/${{ github.event.repository.name }} | tr '[:upper:]' '[:lower:]')" >> "$GITHUB_ENV"\n\n - name: Build and publish Docker image from Dockerfile\n uses: docker/build-push-action@v6.15.0\n with:\n context: .\n platforms: ${{ env.PLATFORMS }}\n provenance: true\n sbom: true\n secrets: |\n ASF_PRIVATE_SNK=${{ secrets.ASF_PRIVATE_SNK }}\n STEAM_TOKEN_DUMPER_TOKEN=${{ secrets.STEAM_TOKEN_DUMPER_TOKEN }}\n labels: |\n org.opencontainers.image.created=${{ env.DATE_ISO8601 }}\n org.opencontainers.image.version=${{ github.sha }}\n org.opencontainers.image.revision=${{ github.sha }}\n tags: |\n ghcr.io/${{ env.GHCR_REPOSITORY }}:${{ env.TAG }}\n ${{ env.DH_REPOSITORY }}:${{ env.TAG }}\n push: true\n\n - name: Update DockerHub repository description\n uses: peter-evans/dockerhub-description@v4.0.2\n with:\n username: ${{ secrets.DOCKERHUB_USERNAME }}\n password: ${{ secrets.DOCKERHUB_TOKEN }}\n repository: ${{ env.DH_REPOSITORY }}\n short-description: ${{ github.event.repository.description }}\n | dataset_sample\yaml\JustArchiNET_ArchiSteamFarm\.github\workflows\docker-publish-main.yml | docker-publish-main.yml | YAML | 2,326 | 0.7 | 0 | 0 | react-lib | 452 | 2023-10-28T17:22:09.662571 | MIT | false | 4677b0d4e4f847af966133a1270c38a4 |
name: ASF-docker-publish-released\n\non:\n push:\n tags:\n - '*'\n\nenv:\n PLATFORMS: linux/amd64,linux/arm,linux/arm64\n TAG: released\n\npermissions:\n packages: write\n\njobs:\n main:\n environment: release-docker\n runs-on: ubuntu-latest\n\n steps:\n - name: Checkout code\n uses: actions/checkout@v4.2.2\n with:\n show-progress: false\n submodules: recursive\n\n - name: Set up Docker Buildx\n uses: docker/setup-buildx-action@v3.10.0\n\n - name: Login to ghcr.io\n uses: docker/login-action@v3.4.0\n with:\n registry: ghcr.io\n username: ${{ github.actor }}\n password: ${{ secrets.GITHUB_TOKEN }}\n\n - name: Login to DockerHub\n uses: docker/login-action@v3.4.0\n with:\n username: ${{ secrets.DOCKERHUB_USERNAME }}\n password: ${{ secrets.DOCKERHUB_TOKEN }}\n\n - name: Prepare environment outputs\n shell: sh\n run: |\n set -eu\n\n echo "FIXED_TAG=$(echo ${{ github.ref }} | cut -d '/' -f 3)" >> "$GITHUB_ENV"\n echo "DATE_ISO8601=$(date --iso-8601=seconds --utc)" >> "$GITHUB_ENV"\n echo "GHCR_REPOSITORY=$(echo ${{ github.repository }} | tr '[:upper:]' '[:lower:]')" >> "$GITHUB_ENV"\n echo "DH_REPOSITORY=$(echo ${{ secrets.DOCKERHUB_USERNAME }}/${{ github.event.repository.name }} | tr '[:upper:]' '[:lower:]')" >> "$GITHUB_ENV"\n\n - name: Build and publish Docker image from Dockerfile\n uses: docker/build-push-action@v6.15.0\n with:\n context: .\n platforms: ${{ env.PLATFORMS }}\n provenance: true\n sbom: true\n secrets: |\n ASF_PRIVATE_SNK=${{ secrets.ASF_PRIVATE_SNK }}\n STEAM_TOKEN_DUMPER_TOKEN=${{ secrets.STEAM_TOKEN_DUMPER_TOKEN }}\n labels: |\n org.opencontainers.image.created=${{ env.DATE_ISO8601 }}\n org.opencontainers.image.version=${{ env.FIXED_TAG }}\n org.opencontainers.image.revision=${{ github.sha }}\n tags: |\n ghcr.io/${{ env.GHCR_REPOSITORY }}:${{ env.TAG }}\n ghcr.io/${{ env.GHCR_REPOSITORY }}:${{ env.FIXED_TAG }}\n ${{ env.DH_REPOSITORY }}:${{ env.TAG }}\n ${{ env.DH_REPOSITORY }}:${{ env.FIXED_TAG }}\n push: true\n | dataset_sample\yaml\JustArchiNET_ArchiSteamFarm\.github\workflows\docker-publish-released.yml | docker-publish-released.yml | YAML | 2,206 | 0.7 | 0 | 0 | react-lib | 357 | 2025-06-23T03:27:07.180924 | BSD-3-Clause | false | be92593549a336dfe55d2e6e57a98d7e |
name: ASF-lock-threads\n\non:\n schedule:\n - cron: '0 1 * * *'\n workflow_dispatch:\n\npermissions:\n discussions: write\n issues: write\n pull-requests: write\n\nconcurrency:\n group: lock-threads\n\njobs:\n lock:\n runs-on: ubuntu-latest\n steps:\n - name: Lock inactive threads\n uses: dessant/lock-threads@v5.0.1\n with:\n discussion-inactive-days: 90\n issue-inactive-days: 60\n pr-inactive-days: 60\n | dataset_sample\yaml\JustArchiNET_ArchiSteamFarm\.github\workflows\lock-threads.yml | lock-threads.yml | YAML | 444 | 0.7 | 0 | 0 | python-kit | 341 | 2024-05-02T02:08:52.996879 | BSD-3-Clause | false | 6de4353d1b17b2e0b62b166743bf589a |
name: ASF-publish\n\non: [push, pull_request]\n\nenv:\n CONFIGURATION: Release\n DOTNET_CLI_TELEMETRY_OPTOUT: true\n DOTNET_NOLOGO: true\n DOTNET_SDK_VERSION: 9.0\n NODE_JS_VERSION: 'lts/*'\n PLUGINS_BUNDLED: ArchiSteamFarm.OfficialPlugins.ItemsMatcher ArchiSteamFarm.OfficialPlugins.MobileAuthenticator ArchiSteamFarm.OfficialPlugins.SteamTokenDumper\n PLUGINS_INCLUDED: ArchiSteamFarm.OfficialPlugins.Monitoring # Apart from declaring them here, there is certain amount of hardcoding needed below for uploading\n\npermissions: {}\n\njobs:\n publish-asf-ui:\n runs-on: ubuntu-latest\n\n steps:\n - name: Checkout code\n uses: actions/checkout@v4.2.2\n with:\n show-progress: false\n submodules: recursive\n\n - name: Setup Node.js with npm\n uses: actions/setup-node@v4.4.0\n with:\n check-latest: true\n node-version: ${{ env.NODE_JS_VERSION }}\n\n - name: Verify Node.js\n run: node -v\n\n - name: Verify npm\n run: npm -v\n\n - name: Install npm modules for ASF-ui\n run: npm ci --no-progress --prefix ASF-ui\n\n - name: Publish ASF-ui\n run: npm run-script deploy --no-progress --prefix ASF-ui\n\n - name: Upload ASF-ui\n uses: actions/upload-artifact@v4.6.2\n with:\n if-no-files-found: error\n name: ASF-ui\n path: ASF-ui/dist\n\n publish-asf:\n needs: publish-asf-ui\n\n strategy:\n fail-fast: false\n matrix:\n include:\n - os: ubuntu-latest\n variant: generic\n - os: ubuntu-latest\n variant: linux-arm\n - os: ubuntu-latest\n variant: linux-arm64\n - os: ubuntu-latest\n variant: linux-x64\n - os: macos-latest\n variant: osx-arm64\n - os: macos-latest\n variant: osx-x64\n - os: windows-latest\n variant: win-arm64\n - os: windows-latest\n variant: win-x64\n\n environment: build\n runs-on: ${{ matrix.os }}\n\n permissions:\n attestations: write\n id-token: write\n\n steps:\n - name: Checkout code\n uses: actions/checkout@v4.2.2\n with:\n show-progress: false\n\n - name: Setup .NET Core\n uses: actions/setup-dotnet@v4.3.1\n with:\n dotnet-version: ${{ env.DOTNET_SDK_VERSION }}\n\n - name: Verify .NET Core\n run: dotnet --info\n\n - name: Download previously built ASF-ui\n uses: actions/download-artifact@v4.2.1\n with:\n name: ASF-ui\n path: ASF-ui/dist\n\n - name: Prepare private key for signing on Unix\n if: startsWith(matrix.os, 'macos-') || startsWith(matrix.os, 'ubuntu-')\n env:\n ASF_PRIVATE_SNK: ${{ secrets.ASF_PRIVATE_SNK }}\n shell: sh\n run: |\n set -eu\n\n if [ -n "${ASF_PRIVATE_SNK-}" ]; then\n echo "$ASF_PRIVATE_SNK" | base64 -d > "resources/ArchiSteamFarm.snk"\n fi\n\n - name: Prepare private key for signing on Windows\n if: startsWith(matrix.os, 'windows-')\n env:\n ASF_PRIVATE_SNK: ${{ secrets.ASF_PRIVATE_SNK }}\n shell: pwsh\n run: |\n Set-StrictMode -Version Latest\n $ErrorActionPreference = 'Stop'\n $ProgressPreference = 'SilentlyContinue'\n\n if ((Test-Path env:ASF_PRIVATE_SNK) -and ($env:ASF_PRIVATE_SNK)) {\n echo "$env:ASF_PRIVATE_SNK" > "resources\ArchiSteamFarm.snk"\n\n certutil -f -decode "resources\ArchiSteamFarm.snk" "resources\ArchiSteamFarm.snk"\n\n if ($LastExitCode -ne 0) {\n throw "Last command failed."\n }\n }\n\n - name: Prepare ArchiSteamFarm.OfficialPlugins.SteamTokenDumper on Unix\n if: startsWith(matrix.os, 'macos-') || startsWith(matrix.os, 'ubuntu-')\n env:\n STEAM_TOKEN_DUMPER_TOKEN: ${{ secrets.STEAM_TOKEN_DUMPER_TOKEN }}\n shell: sh\n run: |\n set -eu\n\n if [ -n "${STEAM_TOKEN_DUMPER_TOKEN-}" ] && [ -f "ArchiSteamFarm.OfficialPlugins.SteamTokenDumper/SharedInfo.cs" ]; then\n sed "s/STEAM_TOKEN_DUMPER_TOKEN/${STEAM_TOKEN_DUMPER_TOKEN}/g" "ArchiSteamFarm.OfficialPlugins.SteamTokenDumper/SharedInfo.cs" > "ArchiSteamFarm.OfficialPlugins.SteamTokenDumper/SharedInfo.cs.new"\n mv "ArchiSteamFarm.OfficialPlugins.SteamTokenDumper/SharedInfo.cs.new" "ArchiSteamFarm.OfficialPlugins.SteamTokenDumper/SharedInfo.cs"\n fi\n\n - name: Prepare ArchiSteamFarm.OfficialPlugins.SteamTokenDumper on Windows\n if: startsWith(matrix.os, 'windows-')\n env:\n STEAM_TOKEN_DUMPER_TOKEN: ${{ secrets.STEAM_TOKEN_DUMPER_TOKEN }}\n shell: pwsh\n run: |\n Set-StrictMode -Version Latest\n $ErrorActionPreference = 'Stop'\n $ProgressPreference = 'SilentlyContinue'\n\n if ((Test-Path env:STEAM_TOKEN_DUMPER_TOKEN) -and ($env:STEAM_TOKEN_DUMPER_TOKEN) -and (Test-Path "ArchiSteamFarm.OfficialPlugins.SteamTokenDumper\SharedInfo.cs" -PathType Leaf)) {\n (Get-Content "ArchiSteamFarm.OfficialPlugins.SteamTokenDumper\SharedInfo.cs").Replace('STEAM_TOKEN_DUMPER_TOKEN', "$env:STEAM_TOKEN_DUMPER_TOKEN") | Set-Content "ArchiSteamFarm.OfficialPlugins.SteamTokenDumper\SharedInfo.cs"\n }\n\n - name: Publish ASF-${{ matrix.variant }} on Unix\n if: startsWith(matrix.os, 'macos-') || startsWith(matrix.os, 'ubuntu-')\n env:\n VARIANT: ${{ matrix.variant }}\n shell: bash\n run: |\n set -euo pipefail\n\n if [ "$VARIANT" = 'generic' ]; then\n variantArgs="-p:TargetLatestRuntimePatch=false -p:UseAppHost=false"\n else\n variantArgs="-p:PublishSingleFile=true -p:PublishTrimmed=true -r $VARIANT --self-contained"\n fi\n\n dotnet publish ArchiSteamFarm -c "$CONFIGURATION" -o "out/${VARIANT}" "-p:ASFVariant=${VARIANT}" -p:ContinuousIntegrationBuild=true --nologo $variantArgs\n\n # Include .ico file for all platforms, since only Windows script can bundle it inside the exe\n cp "resources/ASF.ico" "out/${VARIANT}/ArchiSteamFarm.ico"\n\n # Include extra logic for builds marked for release\n case "$GITHUB_REF" in\n "refs/tags/"*)\n # Update link in Changelog.html accordingly\n if [ -f "out/${VARIANT}/Changelog.html" ]; then\n tag="$(echo "$GITHUB_REF" | cut -c 11-)"\n\n sed "s/ArchiSteamFarm\/commits\/main/ArchiSteamFarm\/releases\/tag\/${tag}/g" "out/${VARIANT}/Changelog.html" > "out/${VARIANT}/Changelog.html.new"\n mv "out/${VARIANT}/Changelog.html.new" "out/${VARIANT}/Changelog.html"\n fi\n ;;\n esac\n\n - name: Publish ASF-${{ matrix.variant }} on Windows\n if: startsWith(matrix.os, 'windows-')\n env:\n VARIANT: ${{ matrix.variant }}\n shell: pwsh\n run: |\n Set-StrictMode -Version Latest\n $ErrorActionPreference = 'Stop'\n $ProgressPreference = 'SilentlyContinue'\n\n if ($env:VARIANT -like 'generic*') {\n $variantArgs = '-p:TargetLatestRuntimePatch=false', '-p:UseAppHost=false'\n } else {\n $variantArgs = '-p:PublishSingleFile=true', '-p:PublishTrimmed=true', '-r', "$env:VARIANT", '--self-contained'\n }\n\n dotnet publish ArchiSteamFarm -c "$env:CONFIGURATION" -o "out\$env:VARIANT" "-p:ASFVariant=$env:VARIANT" -p:ContinuousIntegrationBuild=true --nologo $variantArgs\n\n if ($LastExitCode -ne 0) {\n throw "Last command failed."\n }\n\n # Icon is available only in .exe Windows builds, we'll bundle the .ico file for other flavours\n if (!(Test-Path "out\$env:VARIANT\ArchiSteamFarm.exe" -PathType Leaf)) {\n Copy-Item 'resources\ASF.ico' "out\$env:VARIANT\ArchiSteamFarm.ico"\n }\n\n # Include extra logic for builds marked for release\n if ($env:GITHUB_REF -like 'refs/tags/*') {\n # Update link in Changelog.html accordingly\n if (Test-Path "out\$env:VARIANT\Changelog.html" -PathType Leaf) {\n $tag = $env:GITHUB_REF.Substring(10)\n\n (Get-Content "out\$env:VARIANT\Changelog.html").Replace('ArchiSteamFarm/commits/main', "ArchiSteamFarm/releases/tag/$tag") | Set-Content "out\$env:VARIANT\Changelog.html"\n }\n }\n\n - name: Publish bundled plugins on Unix\n if: startsWith(matrix.os, 'macos-') || startsWith(matrix.os, 'ubuntu-')\n env:\n VARIANT: ${{ matrix.variant }}\n shell: bash\n run: |\n set -euo pipefail\n\n if [ "$VARIANT" = 'generic' ]; then\n variantArgs="-p:TargetLatestRuntimePatch=false -p:UseAppHost=false"\n else\n variantArgs="-r $VARIANT"\n fi\n\n for plugin in $PLUGINS_BUNDLED; do\n dotnet publish "$plugin" -c "$CONFIGURATION" -o "out/${VARIANT}/plugins/${plugin}" "-p:ASFVariant=${VARIANT}" -p:ContinuousIntegrationBuild=true --nologo $variantArgs\n done\n\n - name: Publish bundled plugins on Windows\n if: startsWith(matrix.os, 'windows-')\n env:\n VARIANT: ${{ matrix.variant }}\n shell: pwsh\n run: |\n Set-StrictMode -Version Latest\n $ErrorActionPreference = 'Stop'\n $ProgressPreference = 'SilentlyContinue'\n\n if ($env:VARIANT -like 'generic*') {\n $variantArgs = '-p:TargetLatestRuntimePatch=false', '-p:UseAppHost=false'\n } else {\n $variantArgs = '-r', "$env:VARIANT"\n }\n\n foreach ($plugin in $env:PLUGINS_BUNDLED.Split([char[]] $null, [System.StringSplitOptions]::RemoveEmptyEntries)) {\n dotnet publish "$plugin" -c "$env:CONFIGURATION" -o "out\$env:VARIANT\plugins\$plugin" "-p:ASFVariant=$env:VARIANT" -p:ContinuousIntegrationBuild=true --nologo $variantArgs\n\n if ($LastExitCode -ne 0) {\n throw "Last command failed."\n }\n }\n\n - name: Zip ASF-${{ matrix.variant }} on Unix\n if: startsWith(matrix.os, 'macos-') || startsWith(matrix.os, 'ubuntu-')\n env:\n VARIANT: ${{ matrix.variant }}\n shell: bash\n run: |\n set -euo pipefail\n\n # By default use fastest compression\n seven_zip_args="-mx=1"\n zip_args="-1"\n\n # Tweak compression args for release publishing\n case "$GITHUB_REF" in\n "refs/tags/"*)\n seven_zip_args="-mx=9 -mfb=258 -mpass=15"\n zip_args="-9"\n ;;\n esac\n\n # Create the final zip file\n case "$(uname -s)" in\n "Darwin")\n # We prefer to use zip on macOS as 7z implementation on that OS doesn't handle file permissions (chmod +x)\n if command -v zip >/dev/null; then\n (\n cd "${GITHUB_WORKSPACE}/out/${VARIANT}"\n zip -q -r $zip_args "../ASF-${VARIANT}.zip" .\n )\n else\n 7z a -bd -slp -tzip -mm=Deflate $seven_zip_args "out/ASF-${VARIANT}.zip" "${GITHUB_WORKSPACE}/out/${VARIANT}/*"\n fi\n ;;\n *)\n if command -v 7z >/dev/null; then\n 7z a -bd -slp -tzip -mm=Deflate $seven_zip_args "out/ASF-${VARIANT}.zip" "${GITHUB_WORKSPACE}/out/${VARIANT}/*"\n else\n (\n cd "${GITHUB_WORKSPACE}/out/${VARIANT}"\n zip -q -r $zip_args "../ASF-${VARIANT}.zip" .\n )\n fi\n ;;\n esac\n\n - name: Zip ASF-${{ matrix.variant }} on Windows\n if: startsWith(matrix.os, 'windows-')\n env:\n VARIANT: ${{ matrix.variant }}\n shell: pwsh\n run: |\n Set-StrictMode -Version Latest\n $ErrorActionPreference = 'Stop'\n $ProgressPreference = 'SilentlyContinue'\n\n # By default use fastest compression\n $compressionArgs = '-mx=1'\n\n # Tweak compression args for release publishing\n if ($env:GITHUB_REF -like 'refs/tags/*') {\n $compressionArgs = '-mx=9', '-mfb=258', '-mpass=15'\n }\n\n # Create the final zip file\n 7z a -bd -slp -tzip -mm=Deflate $compressionArgs "out\ASF-$env:VARIANT.zip" "$env:GITHUB_WORKSPACE\out\$env:VARIANT\*"\n\n if ($LastExitCode -ne 0) {\n throw "Last command failed."\n }\n\n # We can aid non-windows users by adding chmod +x flag to appropriate executables directly in the zip file\n # This is ALMOST a hack, but works reliably enough\n if (Test-Path "tools\zip_exec\zip_exec.exe" -PathType Leaf) {\n $executableFiles = @()\n\n if ($env:VARIANT -like 'generic*') {\n $executableFiles += 'ArchiSteamFarm.sh', 'ArchiSteamFarm-Service.sh'\n } elseif (($env:VARIANT -like 'linux*') -or ($env:VARIANT -like 'osx*')) {\n $executableFiles += 'ArchiSteamFarm', 'ArchiSteamFarm-Service.sh'\n }\n\n foreach ($executableFile in $executableFiles) {\n tools\zip_exec\zip_exec.exe "out\ASF-$env:VARIANT.zip" "$executableFile"\n\n if ($LastExitCode -ne 0) {\n throw "Last command failed."\n }\n }\n }\n\n - name: Generate artifact attestation for ASF-${{ matrix.variant }}.zip\n if: ${{ github.event_name == 'push' }}\n uses: actions/attest-build-provenance@v2.2.3\n with:\n subject-path: out/ASF-${{ matrix.variant }}.zip\n\n - name: Upload ASF-${{ matrix.variant }}\n uses: actions/upload-artifact@v4.6.2\n with:\n if-no-files-found: error\n name: ${{ matrix.os }}_ASF-${{ matrix.variant }}\n path: out/ASF-${{ matrix.variant }}.zip\n\n - name: Publish included plugins on Unix\n if: ${{ matrix.os == 'ubuntu-latest' && matrix.variant == 'generic' }}\n env:\n VARIANT: ${{ matrix.variant }}\n shell: bash\n run: |\n set -euo pipefail\n\n # By default use fastest compression\n seven_zip_args="-mx=1"\n zip_args="-1"\n\n # Tweak compression args for release publishing\n case "$GITHUB_REF" in\n "refs/tags/"*)\n seven_zip_args="-mx=9 -mfb=258 -mpass=15"\n zip_args="-9"\n ;;\n esac\n\n for plugin in $PLUGINS_INCLUDED; do\n dotnet publish "$plugin" -c "$CONFIGURATION" -o "out/${plugin}" "-p:ASFVariant=${VARIANT}" -p:ContinuousIntegrationBuild=true -p:TargetLatestRuntimePatch=false -p:UseAppHost=false --nologo\n\n # Create the final zip file\n if command -v 7z >/dev/null; then\n 7z a -bd -slp -tzip -mm=Deflate $seven_zip_args "out/${plugin}.zip" "${GITHUB_WORKSPACE}/out/${plugin}/*"\n else\n (\n cd "${GITHUB_WORKSPACE}/out/${plugin}"\n zip -q -r $zip_args "../${plugin}.zip" .\n )\n fi\n done\n\n - name: Generate artifact attestation for ArchiSteamFarm.OfficialPlugins.Monitoring\n if: ${{ github.event_name == 'push' && matrix.os == 'ubuntu-latest' && matrix.variant == 'generic' }}\n uses: actions/attest-build-provenance@v2.2.3\n with:\n subject-path: out/ArchiSteamFarm.OfficialPlugins.Monitoring.zip\n\n - name: Upload ArchiSteamFarm.OfficialPlugins.Monitoring\n if: ${{ matrix.os == 'ubuntu-latest' && matrix.variant == 'generic' }}\n uses: actions/upload-artifact@v4.6.2\n with:\n if-no-files-found: error\n name: ArchiSteamFarm.OfficialPlugins.Monitoring\n path: out/ArchiSteamFarm.OfficialPlugins.Monitoring.zip\n\n release:\n if: ${{ github.event_name == 'push' && startsWith(github.ref, 'refs/tags/') }}\n needs: publish-asf\n environment: release-github\n runs-on: ubuntu-latest\n\n permissions:\n attestations: write\n contents: write\n id-token: write\n\n steps:\n - name: Checkout code\n uses: actions/checkout@v4.2.2\n with:\n show-progress: false\n\n - name: Download ASF-generic artifact from ubuntu-latest\n uses: actions/download-artifact@v4.2.1\n with:\n name: ubuntu-latest_ASF-generic\n path: out\n\n - name: Download ASF-linux-arm artifact from ubuntu-latest\n uses: actions/download-artifact@v4.2.1\n with:\n name: ubuntu-latest_ASF-linux-arm\n path: out\n\n - name: Download ASF-linux-arm64 artifact from ubuntu-latest\n uses: actions/download-artifact@v4.2.1\n with:\n name: ubuntu-latest_ASF-linux-arm64\n path: out\n\n - name: Download ASF-linux-x64 artifact from ubuntu-latest\n uses: actions/download-artifact@v4.2.1\n with:\n name: ubuntu-latest_ASF-linux-x64\n path: out\n\n - name: Download ASF-osx-arm64 artifact from macos-latest\n uses: actions/download-artifact@v4.2.1\n with:\n name: macos-latest_ASF-osx-arm64\n path: out\n\n - name: Download ASF-osx-x64 artifact from macos-latest\n uses: actions/download-artifact@v4.2.1\n with:\n name: macos-latest_ASF-osx-x64\n path: out\n\n - name: Download ASF-win-arm64 artifact from windows-latest\n uses: actions/download-artifact@v4.2.1\n with:\n name: windows-latest_ASF-win-arm64\n path: out\n\n - name: Download ASF-win-x64 artifact from windows-latest\n uses: actions/download-artifact@v4.2.1\n with:\n name: windows-latest_ASF-win-x64\n path: out\n\n - name: Download ArchiSteamFarm.OfficialPlugins.Monitoring artifact\n uses: actions/download-artifact@v4.2.1\n with:\n name: ArchiSteamFarm.OfficialPlugins.Monitoring\n path: out\n\n - name: Import GPG key for signing\n uses: crazy-max/ghaction-import-gpg@v6.3.0\n with:\n gpg_private_key: ${{ secrets.ARCHIBOT_GPG_PRIVATE_KEY }}\n\n - name: Generate SHA-512 checksums and signature\n shell: sh\n working-directory: out\n run: |\n set -eu\n\n sha512sum *.zip > SHA512SUMS\n gpg -a -b -o SHA512SUMS.sign SHA512SUMS\n\n - name: Generate artifact attestation for SHA512SUMS\n uses: actions/attest-build-provenance@v2.2.3\n with:\n subject-path: out/SHA512SUMS\n\n - name: Upload SHA512SUMS\n uses: actions/upload-artifact@v4.6.2\n with:\n if-no-files-found: error\n name: SHA512SUMS\n path: out/SHA512SUMS\n\n - name: Generate artifact attestation for SHA512SUMS.sign\n uses: actions/attest-build-provenance@v2.2.3\n with:\n subject-path: out/SHA512SUMS.sign\n\n - name: Upload SHA512SUMS.sign\n uses: actions/upload-artifact@v4.6.2\n with:\n if-no-files-found: error\n name: SHA512SUMS.sign\n path: out/SHA512SUMS.sign\n\n - name: Create ArchiSteamFarm GitHub release\n uses: ncipollo/release-action@v1.16.0\n with:\n allowUpdates: true\n artifactErrorsFailBuild: true\n artifacts: "out/*"\n bodyFile: .github/RELEASE_TEMPLATE.md\n makeLatest: false\n name: ArchiSteamFarm V${{ github.ref_name }}\n prerelease: true\n token: ${{ secrets.ARCHIBOT_GITHUB_TOKEN }}\n updateOnlyUnreleased: true\n | dataset_sample\yaml\JustArchiNET_ArchiSteamFarm\.github\workflows\publish.yml | publish.yml | YAML | 19,017 | 0.95 | 0.115809 | 0.041943 | vue-tools | 703 | 2024-12-30T12:33:48.936062 | BSD-3-Clause | false | 4352b9b75100c657c1c7638443c5e8bb |
name: ASF-translations\n\non:\n schedule:\n - cron: '55 1 * * *'\n workflow_dispatch:\n\npermissions:\n contents: write\n\njobs:\n update:\n environment: dev-crowdin\n runs-on: ubuntu-latest\n\n steps:\n - name: Checkout code\n uses: actions/checkout@v4.2.2\n with:\n show-progress: false\n submodules: recursive\n token: ${{ secrets.ARCHIBOT_GITHUB_TOKEN }}\n\n - name: Reset wiki to follow origin\n shell: sh\n working-directory: wiki\n run: |\n set -eu\n\n git fetch origin master\n git reset --hard origin/master\n\n - name: Download latest translations from Crowdin\n uses: crowdin/github-action@v2.7.0\n with:\n upload_sources: false\n download_translations: true\n skip_untranslated_strings: true\n push_translations: false\n crowdin_branch_name: main\n config: '.github/crowdin.yml'\n project_id: ${{ secrets.ASF_CROWDIN_PROJECT_ID }}\n token: ${{ secrets.ASF_CROWDIN_API_TOKEN }}\n\n - name: Import GPG key for signing\n uses: crazy-max/ghaction-import-gpg@v6.3.0\n with:\n gpg_private_key: ${{ secrets.ARCHIBOT_GPG_PRIVATE_KEY }}\n git_config_global: true\n git_user_signingkey: true\n git_commit_gpgsign: true\n\n - name: Commit and push the changes to wiki\n shell: sh\n working-directory: wiki\n run: |\n set -eu\n\n git add -A "locale"\n\n if ! git diff --cached --quiet; then\n git commit -m "Automatic translations update"\n\n git push origin HEAD:master\n fi\n\n - name: Commit and push the changes to ASF\n shell: sh\n run: |\n set -eu\n\n git add -A "ArchiSteamFarm/Localization" "ArchiSteamFarm.OfficialPlugins.ItemsMatcher/Localization" "ArchiSteamFarm.OfficialPlugins.MobileAuthenticator/Localization" "ArchiSteamFarm.OfficialPlugins.SteamTokenDumper/Localization" "wiki"\n\n if ! git diff --cached --quiet; then\n git commit -m "Automatic translations update"\n\n git push\n fi\n | dataset_sample\yaml\JustArchiNET_ArchiSteamFarm\.github\workflows\translations.yml | translations.yml | YAML | 2,052 | 0.85 | 0.038462 | 0 | awesome-app | 605 | 2024-05-01T19:20:18.621832 | GPL-3.0 | false | eb222d5eb0173c4b50070e694ab0f89a |
# These are supported funding model platforms\n\ngithub: # Replace with up to 4 GitHub Sponsors-enabled usernames e.g., [user1, user2]\npatreon: thorbrigsted\nopen_collective: # Replace with a single Open Collective username\nko_fi: thorbrigsted\ntidelift: # Replace with a single Tidelift platform-name/package-name e.g., npm/babel\ncommunity_bridge: # Replace with a single Community Bridge project-name e.g., cloud-foundry\nliberapay: # Replace with a single Liberapay username\nissuehunt: # Replace with a single IssueHunt username\notechie: # Replace with a single Otechie username\ncustom: # Replace with up to 4 custom sponsorship URLs e.g., ['link1', 'link2']\n | dataset_sample\yaml\jynew_jynew\jyx2\Assets\3rd\xNode-1.8.0\.github\FUNDING.yml | FUNDING.yml | YAML | 657 | 0.8 | 0 | 0.090909 | python-kit | 293 | 2024-11-20T20:09:46.500275 | GPL-3.0 | false | a9e52e634e6f47de96f7e0478bcbc040 |
---\nkind: pipeline\nname: amd64\n\nplatform:\n os: linux\n arch: amd64\n\ntrigger:\n event:\n exclude:\n - cron\n - pull_request\n\nclone:\n retries: 3\n\nsteps:\n- name: build\n image: rancher/dapper:v0.6.0\n secrets: [ AWS_SECRET_ACCESS_KEY-k3s-ci-uploader, AWS_ACCESS_KEY_ID-k3s-ci-uploader, unprivileged_github_token ]\n environment:\n GITHUB_TOKEN:\n from_secret: unprivileged_github_token\n AWS_SECRET_ACCESS_KEY:\n from_secret: AWS_SECRET_ACCESS_KEY-k3s-ci-uploader\n AWS_ACCESS_KEY_ID:\n from_secret: AWS_ACCESS_KEY_ID-k3s-ci-uploader\n commands:\n - dapper ci\n - echo "${DRONE_TAG}-amd64" | sed -e 's/+/-/g' >.tags\n volumes:\n - name: docker\n path: /var/run/docker.sock\n\n- name: fossa\n image: rancher/drone-fossa:latest\n failure: ignore\n settings:\n api_key:\n from_secret: FOSSA_API_KEY\n when:\n instance:\n - drone-publish.k3s.io\n ref:\n include:\n - "refs/heads/master"\n - "refs/heads/release-*"\n event:\n - push\n - tag\n\n- name: github_binary_release\n image: plugins/github-release\n settings:\n api_key:\n from_secret: github_token\n prerelease: true\n checksum:\n - sha256\n checksum_file: CHECKSUMsum-amd64.txt\n checksum_flatten: true\n files:\n - "dist/artifacts/*"\n when:\n instance:\n - drone-publish.k3s.io\n ref:\n - refs/head/master\n - refs/tags/*\n event:\n - tag\n\n- name: docker-publish\n image: plugins/docker\n settings:\n dockerfile: package/Dockerfile\n password:\n from_secret: docker_password\n repo: "rancher/k3s"\n username:\n from_secret: docker_username\n build_args_from_env:\n - DRONE_TAG\n when:\n instance:\n - drone-publish.k3s.io\n ref:\n - refs/head/master\n - refs/tags/*\n event:\n - tag\n\n- name: ghcr-publish\n image: plugins/docker\n settings:\n registry: ghcr.io\n dockerfile: package/Dockerfile\n username: "k3s-io"\n password:\n from_secret: github_token\n repo: "ghcr.io/k3s-io/k3s"\n build_args_from_env:\n - DRONE_TAG\n when:\n instance:\n - drone-publish.k3s.io\n ref:\n - refs/head/master\n - refs/tags/*\n event:\n - tag\n\n- name: test\n image: rancher/dapper:v0.6.0\n secrets: [ AWS_SECRET_ACCESS_KEY-k3s-ci-uploader, AWS_ACCESS_KEY_ID-k3s-ci-uploader ]\n environment:\n ENABLE_REGISTRY: 'true'\n AWS_SECRET_ACCESS_KEY:\n from_secret: AWS_SECRET_ACCESS_KEY-k3s-ci-uploader\n AWS_ACCESS_KEY_ID:\n from_secret: AWS_ACCESS_KEY_ID-k3s-ci-uploader\n commands:\n - docker build --target test-k3s -t k3s:test-${DRONE_STAGE_ARCH}-${DRONE_COMMIT} -f Dockerfile.test .\n - >\n docker run -i -e REPO -e TAG -e DRONE_TAG -e DRONE_BUILD_EVENT -e IMAGE_NAME -e AWS_SECRET_ACCESS_KEY -e AWS_ACCESS_KEY_ID -e SONOBUOY_VERSION -e ENABLE_REGISTRY\n -v /var/run/docker.sock:/var/run/docker.sock --privileged --network host -v /tmp:/tmp k3s:test-${DRONE_STAGE_ARCH}-${DRONE_COMMIT}\n volumes:\n - name: docker\n path: /var/run/docker.sock\n\nvolumes:\n- name: docker\n host:\n path: /var/run/docker.sock\n\n---\nkind: pipeline\nname: conformance\n\nplatform:\n os: linux\n arch: amd64\n\ntrigger:\n event:\n - cron\n cron:\n - nightly\n\nsteps:\n- name: build\n image: rancher/dapper:v0.6.0\n commands:\n - dapper ci\n - echo "${DRONE_TAG}-amd64" | sed -e 's/+/-/g' >.tags\n volumes:\n - name: docker\n path: /var/run/docker.sock\n\n- name: test\n image: rancher/dapper:v0.6.0\n environment:\n ENABLE_REGISTRY: 'true'\n commands:\n - docker build --target test-k3s -t k3s:test-${DRONE_STAGE_ARCH}-${DRONE_COMMIT} -f Dockerfile.test .\n - >\n docker run -i -e REPO -e TAG -e DRONE_TAG -e DRONE_BUILD_EVENT -e IMAGE_NAME -e SONOBUOY_VERSION -e ENABLE_REGISTRY\n -v /var/run/docker.sock:/var/run/docker.sock --privileged --network host -v /tmp:/tmp k3s:test-${DRONE_STAGE_ARCH}-${DRONE_COMMIT}\n volumes:\n - name: docker\n path: /var/run/docker.sock\n\nvolumes:\n- name: docker\n host:\n path: /var/run/docker.sock\n\n---\nkind: pipeline\nname: arm64\n\nplatform:\n os: linux\n arch: arm64\n\ntrigger:\n event:\n exclude:\n - cron\n - pull_request\n\nclone:\n retries: 3\n\nsteps:\n- name: build\n image: rancher/dapper:v0.6.0\n secrets: [ AWS_SECRET_ACCESS_KEY-k3s-ci-uploader, AWS_ACCESS_KEY_ID-k3s-ci-uploader ]\n environment:\n AWS_SECRET_ACCESS_KEY:\n from_secret: AWS_SECRET_ACCESS_KEY-k3s-ci-uploader\n AWS_ACCESS_KEY_ID:\n from_secret: AWS_ACCESS_KEY_ID-k3s-ci-uploader\n commands:\n - dapper ci\n - echo "${DRONE_TAG}-arm64" | sed -e 's/+/-/g' >.tags\n volumes:\n - name: docker\n path: /var/run/docker.sock\n\n- name: github_binary_release\n image: plugins/github-release\n settings:\n api_key:\n from_secret: github_token\n prerelease: true\n checksum:\n - sha256\n checksum_file: CHECKSUMsum-arm64.txt\n checksum_flatten: true\n files:\n - "dist/artifacts/*"\n when:\n instance:\n - drone-publish.k3s.io\n ref:\n - refs/head/master\n - refs/tags/*\n event:\n - tag\n\n- name: docker-publish\n image: plugins/docker\n settings:\n dockerfile: package/Dockerfile\n password:\n from_secret: docker_password\n repo: "rancher/k3s"\n username:\n from_secret: docker_username\n build_args_from_env:\n - DRONE_TAG\n when:\n instance:\n - drone-publish.k3s.io\n ref:\n - refs/head/master\n - refs/tags/*\n event:\n - tag\n\n- name: ghcr-publish\n image: plugins/docker\n settings:\n registry: ghcr.io\n dockerfile: package/Dockerfile\n username: "k3s-io"\n password:\n from_secret: github_token\n repo: "ghcr.io/k3s-io/k3s"\n build_args_from_env:\n - DRONE_TAG\n when:\n instance:\n - drone-publish.k3s.io\n ref:\n - refs/head/master\n - refs/tags/*\n event:\n - tag\n\n- name: test\n image: rancher/dapper:v0.6.0\n secrets: [ AWS_SECRET_ACCESS_KEY-k3s-ci-uploader, AWS_ACCESS_KEY_ID-k3s-ci-uploader ]\n environment:\n ENABLE_REGISTRY: 'true'\n AWS_SECRET_ACCESS_KEY:\n from_secret: AWS_SECRET_ACCESS_KEY-k3s-ci-uploader\n AWS_ACCESS_KEY_ID:\n from_secret: AWS_ACCESS_KEY_ID-k3s-ci-uploader\n commands:\n - docker build --target test-k3s -t k3s:test-${DRONE_STAGE_ARCH}-${DRONE_COMMIT} -f Dockerfile.test .\n - >\n docker run -i -e REPO -e TAG -e DRONE_TAG -e DRONE_BUILD_EVENT -e IMAGE_NAME -e AWS_SECRET_ACCESS_KEY -e AWS_ACCESS_KEY_ID -e SONOBUOY_VERSION -e ENABLE_REGISTRY\n -v /var/run/docker.sock:/var/run/docker.sock --privileged --network host -v /tmp:/tmp k3s:test-${DRONE_STAGE_ARCH}-${DRONE_COMMIT}\n volumes:\n - name: docker\n path: /var/run/docker.sock\n\nvolumes:\n- name: docker\n host:\n path: /var/run/docker.sock\n\n---\nkind: pipeline\nname: arm\n\nplatform:\n os: linux\n arch: arm\n\ntrigger:\n event:\n exclude:\n - cron\n\nclone:\n retries: 3\n\nsteps:\n- name: skipfiles\n image: plugins/git\n commands:\n - export NAME=$(test $DRONE_BUILD_EVENT = pull_request && echo remotes/origin/${DRONE_COMMIT_BRANCH:-master} || echo ${DRONE_COMMIT_SHA}~)\n - export DIFF=$(git --no-pager diff --name-only $NAME | grep -v -f .droneignore);\n - if [ -z "$DIFF" ]; then\n echo "All files in PR are on ignore list";\n exit 78;\n else\n echo "Some files in PR are not ignored, $DIFF";\n fi;\n when:\n event:\n - pull_request\n\n- name: build\n # Keeping Dapper at v0.5.0 for armv7, as newer versions fails with\n # Bad system call on this architecture. xref:\n #\n # https://github.com/k3s-io/k3s/pull/8959#discussion_r1439736566\n # https://drone-pr.k3s.io/k3s-io/k3s/7922/3/3\n image: rancher/dapper:v0.5.0\n secrets: [ AWS_SECRET_ACCESS_KEY-k3s-ci-uploader, AWS_ACCESS_KEY_ID-k3s-ci-uploader ]\n environment:\n AWS_SECRET_ACCESS_KEY:\n from_secret: AWS_SECRET_ACCESS_KEY-k3s-ci-uploader\n AWS_ACCESS_KEY_ID:\n from_secret: AWS_ACCESS_KEY_ID-k3s-ci-uploader\n commands:\n - dapper ci\n - echo "${DRONE_TAG}-arm" | sed -e 's/+/-/g' >.tags\n volumes:\n - name: docker\n path: /var/run/docker.sock\n\n- name: github_binary_release\n image: plugins/github-release:linux-arm\n settings:\n api_key:\n from_secret: github_token\n prerelease: true\n checksum:\n - sha256\n checksum_file: CHECKSUMsum-arm.txt\n checksum_flatten: true\n files:\n - "dist/artifacts/*"\n when:\n instance:\n - drone-publish.k3s.io\n ref:\n - refs/head/master\n - refs/tags/*\n event:\n - tag\n\n- name: docker-publish\n image: plugins/docker:linux-arm\n settings:\n dockerfile: package/Dockerfile\n password:\n from_secret: docker_password\n repo: "rancher/k3s"\n username:\n from_secret: docker_username\n build_args_from_env:\n - DRONE_TAG\n when:\n instance:\n - drone-publish.k3s.io\n ref:\n - refs/head/master\n - refs/tags/*\n event:\n - tag\n\n- name: ghcr-publish\n image: plugins/docker:linux-arm\n settings:\n registry: ghcr.io\n dockerfile: package/Dockerfile\n username: "k3s-io"\n password:\n from_secret: github_token\n repo: "ghcr.io/k3s-io/k3s"\n build_args_from_env:\n - DRONE_TAG\n when:\n instance:\n - drone-publish.k3s.io\n ref:\n - refs/head/master\n - refs/tags/*\n event:\n - tag\n\n- name: test\n # Refer to comment for arm/build.\n image: rancher/dapper:v0.5.0\n secrets: [ AWS_SECRET_ACCESS_KEY-k3s-ci-uploader, AWS_ACCESS_KEY_ID-k3s-ci-uploader ]\n environment:\n ENABLE_REGISTRY: 'true'\n AWS_SECRET_ACCESS_KEY:\n from_secret: AWS_SECRET_ACCESS_KEY-k3s-ci-uploader\n AWS_ACCESS_KEY_ID:\n from_secret: AWS_ACCESS_KEY_ID-k3s-ci-uploader\n commands:\n - docker build --target test-k3s -t k3s:test-${DRONE_STAGE_ARCH}-${DRONE_COMMIT} -f Dockerfile.test .\n - >\n docker run -i -e REPO -e TAG -e DRONE_TAG -e DRONE_BUILD_EVENT -e IMAGE_NAME -e AWS_SECRET_ACCESS_KEY -e AWS_ACCESS_KEY_ID -e SONOBUOY_VERSION -e ENABLE_REGISTRY\n -v /var/run/docker.sock:/var/run/docker.sock --privileged --network host -v /tmp:/tmp k3s:test-${DRONE_STAGE_ARCH}-${DRONE_COMMIT}\n volumes:\n - name: docker\n path: /var/run/docker.sock\n\nvolumes:\n- name: docker\n host:\n path: /var/run/docker.sock\n\n---\nkind: pipeline\nname: manifest\n\nplatform:\n os: linux\n arch: amd64\n\nsteps:\n- name: skipfiles\n image: plugins/git\n commands:\n - export NAME=$(test $DRONE_BUILD_EVENT = pull_request && echo remotes/origin/${DRONE_COMMIT_BRANCH:-master} || echo ${DRONE_COMMIT_SHA}~)\n - export DIFF=$(git --no-pager diff --name-only $NAME | grep -v -f .droneignore);\n - if [ -z "$DIFF" ]; then\n echo "All files in PR are on ignore list";\n exit 78;\n else\n echo "Some files in PR are not ignored, $DIFF";\n fi;\n when:\n event:\n - push\n - pull_request\n\n- name: manifest\n image: plugins/docker\n environment:\n DOCKER_USERNAME:\n from_secret: docker_username\n DOCKER_PASSWORD:\n from_secret: docker_password\n settings:\n dry_run: true\n dockerfile: Dockerfile.manifest\n repo: "rancher/k3s-manifest"\n build_args_from_env:\n - DOCKER_USERNAME\n - DOCKER_PASSWORD\n - DRONE_TAG\ntrigger:\n instance:\n - drone-publish.k3s.io\n ref:\n - refs/head/master\n - refs/tags/*\n event:\n include:\n - tag\n exclude:\n - cron\n\ndepends_on:\n- amd64\n- arm64\n- arm\n\n---\nkind: pipeline\nname: dispatch\n\nplatform:\n os: linux\n arch: amd64\n\nclone:\n retries: 3\n\nsteps:\n- name: skipfiles\n image: plugins/git\n commands:\n - export NAME=$(test $DRONE_BUILD_EVENT = pull_request && echo remotes/origin/${DRONE_COMMIT_BRANCH:-master} || echo ${DRONE_COMMIT_SHA}~)\n - export DIFF=$(git --no-pager diff --name-only $NAME | grep -v -f .droneignore);\n - if [ -z "$DIFF" ]; then\n echo "All files in PR are on ignore list";\n exit 78;\n else\n echo "Some files in PR are not ignored, $DIFF";\n fi;\n when:\n event:\n - push\n - pull_request\n\n- name: dispatch\n image: curlimages/curl:7.74.0\n secrets: [ pat_username, github_token, release_token_k3s ] \n user: root\n environment:\n PAT_USERNAME:\n from_secret: pat_username\n PAT_TOKEN:\n from_secret: github_token\n K3S_RELEASE_TOKEN:\n from_secret: release_token_k3s\n commands:\n - apk -U --no-cache add bash\n - scripts/dispatch\n\ntrigger:\n instance:\n - drone-publish.k3s.io\n ref:\n - refs/head/master\n - refs/tags/*\n event:\n - tag\n\ndepends_on:\n- manifest\n\n---\nkind: pipeline\nname: e2e\ntype: docker\n\nplatform:\n os: linux\n arch: amd64\n\nclone:\n retries: 3\n\nsteps:\n- name: skipfiles\n image: plugins/git\n commands:\n - export NAME=$(test $DRONE_BUILD_EVENT = pull_request && echo remotes/origin/${DRONE_COMMIT_BRANCH:-master} || echo ${DRONE_COMMIT_SHA}~)\n - export DIFF=$(git --no-pager diff --name-only $NAME | grep -v -f .droneignore);\n - if [ -z "$DIFF" ]; then\n echo "All files in PR are on ignore list";\n exit 78;\n else\n echo "Some files in PR are not ignored, $DIFF";\n fi;\n when:\n event:\n - push\n - pull_request\n\n- name: build-e2e-image\n image: docker:25.0.5\n commands:\n - DOCKER_BUILDKIT=1 docker build --target test-e2e -t test-e2e -f Dockerfile.test .\n - apk add make git bash\n - GOCOVER=1 make local\n - cp dist/artifacts/* /tmp/artifacts/\n volumes:\n - name: cache\n path: /tmp/artifacts\n - name: docker\n path: /var/run/docker.sock\n\n- name: test-e2e-validatecluster\n depends_on:\n - build-e2e-image\n image: test-e2e\n pull: never\n resources:\n cpu: 6000\n memory: 10Gi\n environment:\n E2E_REGISTRY: 'true'\n E2E_GOCOVER: 'true'\n commands:\n - mkdir -p dist/artifacts\n - cp /tmp/artifacts/* dist/artifacts/\n # Cleanup VMs that are older than 2h. Happens if a previous test panics or is canceled\n - tests/e2e/scripts/cleanup_vms.sh\n - tests/e2e/scripts/drone_registries.sh\n # Stagger the launch of this test with the parallel splitserver test\n # to prevent conflicts over libvirt network interfaces\n - |\n cd tests/e2e/validatecluster\n ../scripts/cleanup_vms.sh 'validatecluster_([0-9]+)_(server|agent)'\n sleep 15\n go test -v -timeout=45m ./validatecluster_test.go -ci -local\n cp ./coverage.out /tmp/artifacts/validate-coverage.out\n volumes:\n - name: libvirt\n path: /var/run/libvirt/\n - name: docker\n path: /var/run/docker.sock\n - name: cache\n path: /tmp/artifacts\n\n- name: test-e2e-splitserver\n depends_on:\n - build-e2e-image\n image: test-e2e\n pull: never\n resources:\n cpu: 6000\n memory: 10Gi\n environment:\n E2E_REGISTRY: 'true'\n E2E_GOCOVER: 'true'\n commands:\n - mkdir -p dist/artifacts\n - cp /tmp/artifacts/* dist/artifacts/\n - tests/e2e/scripts/drone_registries.sh\n - |\n cd tests/e2e/splitserver\n ../scripts/cleanup_vms.sh 'splitserver_([0-9]+)'\n go test -v -timeout=30m ./splitserver_test.go -ci -local\n cp ./coverage.out /tmp/artifacts/split-coverage.out\n - |\n if [ "$DRONE_BUILD_EVENT" = "pull_request" ]; then\n cd ../upgradecluster\n ../scripts/cleanup_vms.sh 'upgradecluster_([0-9]+)_(server|agent)'\n # Convert release-1.XX branch to v1.XX channel\n if [ "$DRONE_BRANCH" = "master" ]; then\n UPGRADE_CHANNEL="latest"\n else\n UPGRADE_CHANNEL=$(echo $DRONE_BRANCH | sed 's/release-/v/')\n # Check if the UPGRADE_CHANNEL exists, in the case of new minor releases it won't\n if ! curl --head --silent --fail https://update.k3s.io/v1-release/channels/$UPGRADE_CHANNEL; then\n UPGRADE_CHANNEL="latest"\n fi\n fi\n E2E_RELEASE_CHANNEL=$UPGRADE_CHANNEL go test -v -timeout=45m ./upgradecluster_test.go -ci -local -ginkgo.v\n cp ./coverage.out /tmp/artifacts/upgrade-coverage.out\n fi\n \n volumes:\n - name: libvirt\n path: /var/run/libvirt/\n - name: docker\n path: /var/run/docker.sock\n - name: cache\n path: /tmp/artifacts\n\n- name: upload to codecov\n depends_on:\n - test-e2e-validatecluster\n - test-e2e-splitserver\n image: robertstettner/drone-codecov\n settings:\n token: \n from_secret: codecov_token\n files:\n - /tmp/artifacts/validate-coverage.out\n - /tmp/artifacts/split-coverage.out\n - /tmp/artifacts/upgrade-coverage.out\n flags: \n - e2etests\n when:\n event:\n - push\n\n volumes:\n - name: cache\n path: /tmp/artifacts\n\nvolumes:\n- name: docker\n host:\n path: /var/run/docker.sock\n- name: libvirt\n host:\n path: /var/run/libvirt/\n- name: cache\n temp: {}\n | dataset_sample\yaml\k3s-io_k3s\.drone.yml | .drone.yml | YAML | 16,192 | 0.8 | 0.015805 | 0.017296 | node-utils | 518 | 2024-05-10T03:00:32.196288 | MIT | false | d7141770a54de72cd59874917a81ef88 |
# to run define K3S_TOKEN, K3S_VERSION is optional, eg:\n# K3S_TOKEN=${RANDOM}${RANDOM}${RANDOM} docker-compose up\n\nservices:\n\n server:\n image: "rancher/k3s:${K3S_VERSION:-latest}"\n command: server\n tmpfs:\n - /run\n - /var/run\n ulimits:\n nproc: 65535\n nofile:\n soft: 65535\n hard: 65535\n privileged: true\n restart: always\n environment:\n - K3S_TOKEN=${K3S_TOKEN:?err}\n - K3S_KUBECONFIG_OUTPUT=/output/kubeconfig.yaml\n - K3S_KUBECONFIG_MODE=666\n volumes:\n - k3s-server:/var/lib/rancher/k3s\n # This is just so that we get the kubeconfig file out\n - .:/output\n ports:\n - 6443:6443 # Kubernetes API Server\n - 80:80 # Ingress controller port 80\n - 443:443 # Ingress controller port 443\n\n agent:\n image: "rancher/k3s:${K3S_VERSION:-latest}"\n tmpfs:\n - /run\n - /var/run\n ulimits:\n nproc: 65535\n nofile:\n soft: 65535\n hard: 65535\n privileged: true\n restart: always\n environment:\n - K3S_URL=https://server:6443\n - K3S_TOKEN=${K3S_TOKEN:?err}\n volumes:\n - k3s-agent:/var/lib/rancher/k3s\n\nvolumes:\n k3s-server: {}\n k3s-agent: {}\n | dataset_sample\yaml\k3s-io_k3s\docker-compose.yml | docker-compose.yml | YAML | 1,173 | 0.8 | 0 | 0.0625 | awesome-app | 579 | 2024-11-24T16:33:23.347996 | BSD-3-Clause | false | 744f0f39e09f4e2d8133cf6e98b85775 |
version: '2'\nservices:\n kanboard:\n image: kanboard/kanboard:latest\n ports:\n - "80:80"\n - "443:443"\n volumes:\n - kanboard_data:/var/www/app/data\n - kanboard_plugins:/var/www/app/plugins\n - kanboard_ssl:/etc/nginx/ssl\nvolumes:\n kanboard_data:\n driver: local\n kanboard_plugins:\n driver: local\n kanboard_ssl:\n driver: local\n | dataset_sample\yaml\kanboard_kanboard\docker-compose.yml | docker-compose.yml | YAML | 362 | 0.7 | 0 | 0 | vue-tools | 45 | 2025-05-01T11:20:26.074762 | Apache-2.0 | false | 6a27ae26c2074b3ec7c8cf3f3c93fbfa |
# These are supported funding model platforms\n\ngithub: # Replace with up to 4 GitHub Sponsors-enabled usernames e.g., [user1, user2]\npatreon: # Replace with a single Patreon username\nopen_collective: # Replace with a single Open Collective username\nko_fi: # Replace with a single Ko-fi username\ntidelift: # Replace with a single Tidelift platform-name/package-name e.g., npm/babel\ncommunity_bridge: # Replace with a single Community Bridge project-name e.g., cloud-foundry\nliberapay: # Replace with a single Liberapay username\nissuehunt: # Replace with a single IssueHunt username\notechie: # Replace with a single Otechie username\ncustom: ["https://paypal.me/majora2007"]\n | dataset_sample\yaml\Kareadita_Kavita\FUNDING.yml | FUNDING.yml | YAML | 672 | 0.8 | 0 | 0.090909 | vue-tools | 474 | 2023-11-24T18:34:17.510801 | Apache-2.0 | false | 3e1e8ace4d784f63381b2112384a43ef |
name: build-wasm\n\non: [workflow_call, workflow_dispatch]\n\njobs:\n build-wasm:\n runs-on: ubuntu-22.04\n steps:\n - name: Check out repository\n uses: actions/checkout@v3\n\n - name: Install Emscripten\n uses: mymindstorm/setup-emsdk@v13\n with:\n version: 3.1.58\n\n - name: Install Rust\n uses: dtolnay/rust-toolchain@stable\n with:\n toolchain: stable\n\n - name: Install Nightly Rust\n uses: dtolnay/rust-toolchain@stable\n with:\n toolchain: nightly\n components: rustfmt\n\n - name: Install Python\n uses: actions/setup-python@v4\n with:\n python-version: "3.12"\n\n - name: Install Python packages\n run: pip install -r python/requirements.txt\n\n - name: Build wheel\n run: make build-wasm\n\n - name: Upload wheel\n uses: actions/upload-artifact@v4\n with:\n name: pyxel-wasm\n path: dist/*\n | dataset_sample\yaml\kitao_pyxel\.github\workflows\build-wasm.yml | build-wasm.yml | YAML | 962 | 0.95 | 0 | 0 | react-lib | 507 | 2024-12-06T16:39:14.760912 | Apache-2.0 | false | 7c10921451880bbc5ebe3facecf399cb |
name: build\n\non: [workflow_call, workflow_dispatch]\n\njobs:\n build-winmac:\n strategy:\n fail-fast: false\n matrix:\n include:\n - target: x86_64-pc-windows-msvc\n os: windows-2022\n arch: x64\n\n - target: i686-pc-windows-msvc\n os: windows-2022\n arch: x86\n\n - target: x86_64-apple-darwin\n os: macos-13\n arch: x64\n\n - target: aarch64-apple-darwin\n os: macos-13\n arch: x64\n\n runs-on: ${{ matrix.os }}\n steps:\n - name: Check out repository\n uses: actions/checkout@v3\n\n - name: Install Rust\n uses: dtolnay/rust-toolchain@stable\n with:\n toolchain: stable\n\n - name: Install Nightly Rust\n uses: dtolnay/rust-toolchain@stable\n with:\n toolchain: nightly\n components: rustfmt\n\n - name: Install Python\n uses: actions/setup-python@v4\n with:\n python-version: "3.12"\n architecture: ${{ matrix.arch }}\n\n - name: Install Python packages\n run: pip install -r python/requirements.txt\n\n - name: Build wheel\n run: make TARGET=${{ matrix.target }}\n\n - name: Upload wheel\n uses: actions/upload-artifact@v4\n with:\n name: pyxel-${{ matrix.target }}\n path: dist/*\n\n build-linux:\n strategy:\n fail-fast: false\n matrix:\n include:\n - target: x86_64-unknown-linux-gnu\n image: messense/manylinux2014-cross:x86_64\n\n - target: i686-unknown-linux-gnu\n image: messense/manylinux2014-cross:i686\n\n - target: aarch64-unknown-linux-gnu\n image: messense/manylinux2014-cross:aarch64\n\n - target: armv7-unknown-linux-gnueabihf\n image: messense/manylinux2014-cross:armv7l\n\n runs-on: ubuntu-22.04\n container: ${{ matrix.image }}\n\n steps:\n - name: Check out repository\n uses: actions/checkout@v3\n\n - name: Install Rust\n uses: dtolnay/rust-toolchain@stable\n with:\n toolchain: stable\n\n - name: Install Nightly Rust\n uses: dtolnay/rust-toolchain@stable\n with:\n toolchain: nightly\n components: rustfmt\n\n - name: Install Python\n uses: actions/setup-python@v4\n with:\n python-version: "3.12"\n\n - name: Install Python packages\n run: pip install -r python/requirements.txt\n\n - name: Build and install SDL2\n run: |\n SDL2_VERSION=2.0.10\n export CC=$TARGET_CC\n export AR=$TARGET_AR\n export RANLIB=$TARGET_RANLIB\n\n curl -sqLO https://github.com/libsdl-org/SDL/releases/download/release-${SDL2_VERSION}/SDL2-${SDL2_VERSION}.tar.gz\n tar xzf SDL2-${SDL2_VERSION}.tar.gz\n\n cd SDL2-${SDL2_VERSION}\n CFLAGS="-O3 -fPIC" ./configure \\n --build=x86_64-unknown-linux-gnu \\n --host=${{ matrix.target }} \\n --prefix=/usr/${{ matrix.target }}\n make -j4\n make install\n cd ..\n\n rm -rf SDL2-${SDL2_VERSION}\n\n - name: Build wheel\n run: |\n export BINDGENFLAGS="-I/usr/${{ matrix.target }}/include/SDL2 -I/usr/${{ matrix.target }}/${{ matrix.target }}/sysroot/usr/include"\n export RUSTFLAGS="-L/usr/${{ matrix.target }}/lib"\n\n make TARGET=${{ matrix.target }}\n\n - name: Upload wheel\n uses: actions/upload-artifact@v4\n with:\n name: pyxel-${{ matrix.target }}\n path: dist/*\n | dataset_sample\yaml\kitao_pyxel\.github\workflows\build.yml | build.yml | YAML | 3,567 | 0.95 | 0 | 0 | node-utils | 510 | 2024-02-08T21:23:03.963347 | GPL-3.0 | false | fd488be29605d3f15d9730d2bea32acf |
name: release\n\non:\n push:\n tags:\n - "v*.*.*"\n\njobs:\n build:\n uses: kitao/pyxel/.github/workflows/build.yml@main\n\n release:\n runs-on: macos-13\n needs: build\n steps:\n - name: Check out repository\n uses: actions/checkout@v3\n\n - name: Download wheels\n uses: actions/download-artifact@v4\n with:\n path: dist\n\n - name: Flatten wheels\n run: |\n find dist -type f -name "*.whl" -exec mv {} dist/ \;\n find dist -type d -empty -delete\n\n - name: Get Pyxel version\n run: |\n PYXEL_VERSION=$(echo $GITHUB_REF | sed 's/refs\/tags\/v//')\n echo "PYXEL_VERSION=${PYXEL_VERSION}" >> $GITHUB_ENV\n\n - name: Create release\n uses: softprops/action-gh-release@v1\n with:\n name: Pyxel ${{ env.PYXEL_VERSION }}\n tag_name: ${{ github.ref }}\n files: dist/*\n\n - name: Install Python\n uses: actions/setup-python@v4\n with:\n python-version: "3.12"\n\n - name: Publish wheels\n run: |\n pip install twine\n twine upload dist/*\n env:\n TWINE_USERNAME: ${{ secrets.PYPI_USERNAME }}\n TWINE_PASSWORD: ${{ secrets.PYPI_PASSWORD }}\n\n - name: Install Rust\n uses: dtolnay/rust-toolchain@stable\n\n - name: Publish crates\n run: |\n cargo login ${{ secrets.CARGO_REGISTRY_TOKEN }}\n cd rust/pyxel-platform\n cargo publish\n sleep 60\n cd ../pyxel-engine\n cargo publish\n sleep 60\n cd ../pyxel-wrapper\n RUSTFLAGS="-C link-arg=-undefined -C link-arg=dynamic_lookup" cargo publish\n | dataset_sample\yaml\kitao_pyxel\.github\workflows\release.yml | release.yml | YAML | 1,675 | 0.8 | 0 | 0 | node-utils | 42 | 2025-04-30T09:36:51.881299 | MIT | false | 53b48d20d75be41b6260a484eebe3253 |
# This configuration file was automatically generated by Gitpod.\n# Please adjust to your needs (see https://www.gitpod.io/docs/config-gitpod-file)\n# and commit this file to your remote git repository to share the goodness with others.\n\ntasks:\n - init: sudo ./cmd install\n command: ./cmd dev\n\nports:\n - port: 2222\n visibility: public\n - port: 22222\n visibility: public\n | dataset_sample\yaml\kuaifan_dootask\.gitpod.yml | .gitpod.yml | YAML | 380 | 0.8 | 0 | 0.272727 | react-lib | 71 | 2024-02-10T14:45:25.271044 | MIT | false | fdcb96f371097fc92910d739b1ac5f56 |
php:\n preset: laravel\n disabled:\n - no_unused_imports\n finder:\n not-name:\n - index.php\n - server.php\njs:\n finder:\n not-name:\n - webpack.mix.js\ncss: true\n | dataset_sample\yaml\kuaifan_dootask\.styleci.yml | .styleci.yml | YAML | 181 | 0.85 | 0 | 0 | awesome-app | 376 | 2024-06-17T09:09:16.709212 | Apache-2.0 | false | dbfba971f50b26e7514798e185664a40 |
services:\n php:\n container_name: "dootask-php-${APP_ID}"\n image: "kuaifan/php:swoole-8.0.rc18"\n shm_size: 2G\n ulimits:\n core:\n soft: 0\n hard: 0\n volumes:\n - shared_data:/usr/share/dootask\n - ./docker/crontab/crontab.conf:/etc/supervisor/conf.d/crontab.conf\n - ./docker/php/php.conf:/etc/supervisor/conf.d/php.conf\n - ./docker/php/php.ini:/usr/local/etc/php/php.ini\n - ./docker/log/supervisor:/var/log/supervisor\n - ./:/var/www\n environment:\n LANG: "C.UTF-8"\n MODE: "production"\n MYSQL_HOST: "${DB_HOST}"\n MYSQL_PORT: "${DB_PORT}"\n MYSQL_DB_NAME: "${DB_DATABASE}"\n MYSQL_USERNAME: "${DB_USERNAME}"\n MYSQL_PASSWORD: "${DB_PASSWORD}"\n networks:\n extnetwork:\n ipv4_address: "${APP_IPPR}.2"\n depends_on:\n mariadb:\n condition: service_healthy\n redis:\n condition: service_healthy\n restart: unless-stopped\n\n nginx:\n container_name: "dootask-nginx-${APP_ID}"\n image: "nginx:alpine"\n ports:\n - "${APP_PORT}:80"\n - "${APP_SSL_PORT:-}:443"\n volumes:\n - ./docker/nginx:/etc/nginx/conf.d\n - ./public:/var/www/public\n networks:\n extnetwork:\n ipv4_address: "${APP_IPPR}.3"\n restart: unless-stopped\n\n redis:\n container_name: "dootask-redis-${APP_ID}"\n image: "redis:alpine"\n volumes:\n - redis_data:/data\n healthcheck:\n test: ["CMD", "redis-cli", "ping"]\n interval: 5s\n timeout: 5s\n retries: 5\n networks:\n extnetwork:\n ipv4_address: "${APP_IPPR}.4"\n restart: unless-stopped\n\n mariadb:\n container_name: "dootask-mariadb-${APP_ID}"\n image: "mariadb:10.7.3"\n volumes:\n - ./docker/mysql/repassword.sh:/etc/mysql/repassword.sh\n - ./docker/mysql/conf.d:/etc/mysql/conf.d\n - ./docker/mysql/data:/var/lib/mysql\n environment:\n MYSQL_PREFIX: "${DB_PREFIX}"\n MYSQL_ROOT_PASSWORD: "${DB_ROOT_PASSWORD}"\n MYSQL_DATABASE: "${DB_DATABASE}"\n MYSQL_USER: "${DB_USERNAME}"\n MYSQL_PASSWORD: "${DB_PASSWORD}"\n healthcheck:\n test: ["CMD", "mysqladmin", "ping", "-h", "localhost", "-u", "${DB_USERNAME}", "-p${DB_PASSWORD}"]\n interval: 5s\n timeout: 5s\n retries: 5\n networks:\n extnetwork:\n ipv4_address: "${APP_IPPR}.5"\n restart: unless-stopped\n\n office:\n container_name: "dootask-office-${APP_ID}"\n image: "onlyoffice/documentserver:8.2.2.1"\n volumes:\n - ./docker/office/logs:/var/log/onlyoffice\n - ./docker/office/data:/var/www/onlyoffice/Data\n - ./docker/office/etc/documentserver/default.json:/etc/onlyoffice/documentserver/default.json\n - ./docker/office/resources/require.js:/var/www/onlyoffice/documentserver/web-apps/vendor/requirejs/require.js\n - ./docker/office/resources/common/main/resources/img/header:/var/www/onlyoffice/documentserver/web-apps/apps/common/main/resources/img/header\n - ./docker/office/resources/documenteditor/main/resources/css/app.css:/var/www/onlyoffice/documentserver/web-apps/apps/documenteditor/main/resources/css/app.css\n - ./docker/office/resources/documenteditor/mobile/css/526.caf35c11a8d72ca5ac85.css:/var/www/onlyoffice/documentserver/web-apps/apps/documenteditor/mobile/css/526.caf35c11a8d72ca5ac85.css\n - ./docker/office/resources/presentationeditor/main/resources/css/app.css:/var/www/onlyoffice/documentserver/web-apps/apps/presentationeditor/main/resources/css/app.css\n - ./docker/office/resources/presentationeditor/mobile/css/923.f9cf19de1a25c2e7bf8b.css:/var/www/onlyoffice/documentserver/web-apps/apps/presentationeditor/mobile/css/923.f9cf19de1a25c2e7bf8b.css\n - ./docker/office/resources/spreadsheeteditor/main/resources/css/app.css:/var/www/onlyoffice/documentserver/web-apps/apps/spreadsheeteditor/main/resources/css/app.css\n - ./docker/office/resources/spreadsheeteditor/mobile/css/611.1bef49f175e18fc085db.css:/var/www/onlyoffice/documentserver/web-apps/apps/spreadsheeteditor/mobile/css/611.1bef49f175e18fc085db.css\n environment:\n JWT_SECRET: ${APP_KEY}\n networks:\n extnetwork:\n ipv4_address: "${APP_IPPR}.6"\n restart: unless-stopped\n\n fileview:\n container_name: "dootask-fileview-${APP_ID}"\n image: "kuaifan/fileview:4.4.0-4"\n environment:\n KK_CONTEXT_PATH: "/fileview"\n KK_OFFICE_PREVIEW_SWITCH_DISABLED: true\n KK_MEDIA_CONVERT_DISABLE: true\n networks:\n extnetwork:\n ipv4_address: "${APP_IPPR}.7"\n restart: unless-stopped\n\n drawio-webapp:\n container_name: "dootask-drawio-webapp-${APP_ID}"\n image: "jgraph/drawio:24.7.17"\n volumes:\n - ./docker/drawio/webapp/index.html:/usr/local/tomcat/webapps/draw/index.html\n - ./docker/drawio/webapp/stencils:/usr/local/tomcat/webapps/draw/stencils\n - ./docker/drawio/webapp/js/app.min.js:/usr/local/tomcat/webapps/draw/js/app.min.js\n - ./docker/drawio/webapp/js/croppie/croppie.min.css:/usr/local/tomcat/webapps/draw/js/croppie/croppie.min.css\n - ./docker/drawio/webapp/js/diagramly/ElectronApp.js:/usr/local/tomcat/webapps/draw/js/diagramly/ElectronApp.js\n networks:\n extnetwork:\n ipv4_address: "${APP_IPPR}.8"\n restart: unless-stopped\n\n drawio-export:\n container_name: "dootask-drawio-export-${APP_ID}"\n image: "kuaifan/export-server:0.0.1"\n networks:\n extnetwork:\n ipv4_address: "${APP_IPPR}.9"\n volumes:\n - ./docker/drawio/export/fonts:/usr/share/fonts/drawio\n restart: unless-stopped\n\n minder:\n container_name: "dootask-minder-${APP_ID}"\n image: "kuaifan/minder:0.1.3"\n networks:\n extnetwork:\n ipv4_address: "${APP_IPPR}.10"\n restart: unless-stopped\n\n approve:\n container_name: "dootask-approve-${APP_ID}"\n image: "kuaifan/dooapprove:0.1.5"\n environment:\n TZ: "${TIMEZONE:-PRC}"\n MYSQL_HOST: "${DB_HOST}"\n MYSQL_PORT: "${DB_PORT}"\n MYSQL_DBNAME: "${DB_DATABASE}"\n MYSQL_USERNAME: "${DB_USERNAME}"\n MYSQL_PASSWORD: "${DB_PASSWORD}"\n MYSQL_Prefix: "${DB_PREFIX}approve_"\n DEMO_DATA: true\n KEY: ${APP_KEY}\n networks:\n extnetwork:\n ipv4_address: "${APP_IPPR}.11"\n restart: unless-stopped\n\n ai:\n container_name: "dootask-ai-${APP_ID}"\n image: "kuaifan/dootask-ai:0.3.5"\n environment:\n REDIS_HOST: "${REDIS_HOST}"\n REDIS_PORT: "${REDIS_PORT}"\n TIMEOUT: 600\n networks:\n extnetwork:\n ipv4_address: "${APP_IPPR}.12"\n restart: unless-stopped\n\n okr:\n container_name: "dootask-okr-${APP_ID}"\n image: "kuaifan/doookr:0.4.5"\n environment:\n TZ: "${TIMEZONE:-PRC}"\n DOO_TASK_URL: "http://nginx"\n MYSQL_HOST: "${DB_HOST}"\n MYSQL_PORT: "${DB_PORT}"\n MYSQL_DBNAME: "${DB_DATABASE}"\n MYSQL_USERNAME: "${DB_USERNAME}"\n MYSQL_PASSWORD: "${DB_PASSWORD}"\n MYSQL_PREFIX: "${DB_PREFIX}"\n DEMO_DATA: true\n KEY: ${APP_KEY}\n networks:\n extnetwork:\n ipv4_address: "${APP_IPPR}.13"\n restart: unless-stopped\n\n face:\n container_name: "dootask-face-${APP_ID}"\n image: "hitosea2020/dooface:0.0.1"\n ports:\n - "7788:7788"\n environment:\n TZ: "${TIMEZONE:-PRC}"\n STORAGE: mysql\n MYSQL_HOST: "${DB_HOST}"\n MYSQL_PORT: "${DB_PORT}"\n MYSQL_USERNAME: "${DB_USERNAME}"\n MYSQL_PASSWORD: "${DB_PASSWORD}"\n MYSQL_DB_NAME: "${DB_DATABASE}"\n DB_PREFIX: "${DB_PREFIX}"\n REPORT_API: "http://nginx/api/public/checkin/report"\n networks:\n extnetwork:\n ipv4_address: "${APP_IPPR}.14"\n restart: unless-stopped\n\n search:\n container_name: "dootask-search-${APP_ID}"\n image: "public.ecr.aws/zinclabs/zincsearch:0.4.10"\n volumes:\n - search_data:/data\n environment:\n ZINC_DATA_PATH: "/data"\n ZINC_FIRST_ADMIN_USER: "${DB_USERNAME}"\n ZINC_FIRST_ADMIN_PASSWORD: "${DB_PASSWORD}"\n deploy:\n resources:\n limits:\n cpus: '2'\n networks:\n extnetwork:\n ipv4_address: "${APP_IPPR}.15"\n restart: unless-stopped\n\nnetworks:\n extnetwork:\n name: "dootask-networks-${APP_ID}"\n ipam:\n config:\n - subnet: "${APP_IPPR}.0/24"\n gateway: "${APP_IPPR}.1"\n\nvolumes:\n shared_data:\n name: "dootask-shared-data-${APP_ID}"\n redis_data:\n name: "dootask-redis-data-${APP_ID}"\n search_data:\n name: "dootask-search-data-${APP_ID}"\n | dataset_sample\yaml\kuaifan_dootask\docker-compose.yml | docker-compose.yml | YAML | 8,384 | 0.95 | 0 | 0 | vue-tools | 948 | 2025-03-06T11:43:37.046335 | BSD-3-Clause | false | 79b3999c9cb159d2b11a15e1cd994f3c |
# Copyright 2017 The Kubernetes Authors.\n#\n# Licensed under the Apache License, Version 2.0 (the "License");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an "AS IS" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n#\ncodecov:\n require_ci_to_pass: true\n\ncoverage:\n precision: 2\n round: down\n range: "70...90"\n\n status:\n project:\n default:\n target: 75%\n threshold: 1%\n # We need codecov purely for information purposes. Due to coverage calculation issue we have to disable checks.\n informational: true\n # We use codecov for information described as above. So it is enough to check only `project` as of now.\n patch: off\n changes: false\n\nparsers:\n gcov:\n branch_detection:\n conditional: yes\n loop: yes\n method: no\n macro: no\n\ncomment:\n layout: "header, diff"\n behavior: default\n require_changes: false\n | dataset_sample\yaml\kubernetes_dashboard\.codecov.yml | .codecov.yml | YAML | 1,243 | 0.95 | 0.066667 | 0.390244 | react-lib | 506 | 2023-08-01T15:51:23.306506 | MIT | false | 206da937b2222675ea5b35e2b8477fda |
# Copyright 2017 The Kubernetes Authors.\n#\n# Licensed under the Apache License, Version 2.0 (the "License");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an "AS IS" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n#\n\nchart-repos:\n - metrics-server=https://kubernetes-sigs.github.io/metrics-server/\n - cert-manager=https://charts.jetstack.io\n - ingress-nginx=https://kubernetes.github.io/ingress-nginx\n - kong=https://charts.konghq.com\ndebug: false\n# additional-commands:\n# - helm kubeconform {{ .Path }} --config charts/helm-chart/.kubeconform\n | dataset_sample\yaml\kubernetes_dashboard\.ct.yml | .ct.yml | YAML | 918 | 0.95 | 0.043478 | 0.727273 | python-kit | 727 | 2025-06-28T14:02:02.390563 | Apache-2.0 | false | b1d1824ef5fc9b144459078770337bc3 |
# Copyright 2017 The Kubernetes Authors.\n#\n# Licensed under the Apache License, Version 2.0 (the "License");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an "AS IS" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n#\nrun:\n modules-download-mode: readonly\n allow-parallel-runners: true\n timeout: 30m\n issues-exit-code: 1\n tests: true\nlinters:\n disable-all: true\n enable:\n # default linters\n - errcheck\n - gosimple\n - govet\n - ineffassign\n - staticcheck\n - typecheck\n - unused\n - gocyclo\n\n # additional linters\n - errorlint\n - goimports\n - misspell\n | dataset_sample\yaml\kubernetes_dashboard\.golangci.yml | .golangci.yml | YAML | 961 | 0.95 | 0.027027 | 0.444444 | python-kit | 818 | 2025-01-04T10:17:33.625577 | Apache-2.0 | false | 64d75cc3dcb4ca89fe3669e0d155a2b0 |
version: "2"\nrun:\n allow-parallel-runners: true\nlinters:\n default: none\n enable:\n - asasalint\n - asciicheck\n - bidichk\n - bodyclose\n - contextcheck\n - decorder\n - dogsled\n - dupl\n - durationcheck\n - errcheck\n - errchkjson\n - errname\n - ginkgolinter\n - gocheckcompilerdirectives\n - goconst\n - gocritic\n - gocyclo\n - godox\n - goheader\n - gomoddirectives\n - gomodguard\n - goprintffuncname\n - gosec\n - govet\n - grouper\n - importas\n - ineffassign\n - loggercheck\n - makezero\n - misspell\n - musttag\n - nakedret\n - nolintlint\n - nosprintfhostport\n - prealloc\n - predeclared\n - promlinter\n - reassign\n - revive\n - rowserrcheck\n - sqlclosecheck\n - staticcheck\n - testableexamples\n - unconvert\n - unparam\n - unused\n - usestdlibvars\n - whitespace\n settings:\n errcheck:\n check-type-assertions: true\n check-blank: true\n gocritic:\n enabled-checks:\n - appendAssign\n - argOrder\n - badCall\n - badCond\n - badLock\n - badRegexp\n - badSorting\n - builtinShadowDecl\n - caseOrder\n - codegenComment\n - commentedOutCode\n - deferInLoop\n - deprecatedComment\n - dupArg\n - dupBranchBody\n - dupCase\n - dupSubExpr\n - dynamicFmtString\n - emptyDecl\n - evalOrder\n - exitAfterDefer\n - externalErrorReassign\n - filepathJoin\n - flagDeref\n - flagName\n - mapKey\n - nilValReturn\n - offBy1\n - regexpPattern\n - returnAfterHttpError\n - sloppyReassign\n - sloppyTypeAssert\n - sortSlice\n - sprintfQuotedString\n - sqlQuery\n - syncMapLoadAndDelete\n - truncateCmp\n - unnecessaryDefer\n - weakCond\n - appendCombine\n - equalFold\n - hugeParam\n - indexAlloc\n - preferDecodeRune\n - preferFprint\n - preferStringWriter\n - preferWriteByte\n - rangeExprCopy\n - rangeValCopy\n - sliceClear\n - stringXbytes\n - assignOp\n - boolExprSimplify\n - captLocal\n - commentFormatting\n - commentedOutImport\n - defaultCaseOrder\n - deferUnlambda\n - docStub\n - dupImport\n - elseif\n - emptyFallthrough\n - emptyStringTest\n - exposedSyncMutex\n - hexLiteral\n - httpNoBody\n - ifElseChain\n - methodExprCall\n - newDeref\n - octalLiteral\n - preferFilepathJoin\n - redundantSprint\n - regexpMust\n - regexpSimplify\n - ruleguard\n - singleCaseSwitch\n - sloppyLen\n - stringConcatSimplify\n - stringsCompare\n - switchTrue\n - timeExprSimplify\n - tooManyResultsChecker\n - typeAssertChain\n - typeDefFirst\n - typeSwitchVar\n - underef\n - unlabelStmt\n - unlambda\n - unslice\n - valSwap\n - whyNoLint\n - wrapperFunc\n - yodaStyleExpr\n - builtinShadow\n - importShadow\n - initClause\n - nestingReduce\n - paramTypeCombine\n - ptrToRefParam\n - typeUnparen\n - unnamedResult\n - unnecessaryBlock\n gocyclo:\n min-complexity: 40\n godox:\n keywords:\n - BUG\n - FIXME\n - HACK\n nolintlint:\n require-explanation: false\n require-specific: true\n allow-unused: false\n exclusions:\n generated: lax\n presets:\n - comments\n - common-false-positives\n - legacy\n - std-error-handling\n paths:\n - third_party$\n - builtin$\n - examples$\nissues:\n max-issues-per-linter: 0\n max-same-issues: 0\nformatters:\n enable:\n - gofmt\n - gofumpt\n - goimports\n exclusions:\n generated: lax\n paths:\n - third_party$\n - builtin$\n - examples$\n | dataset_sample\yaml\kubernetes_ingress-nginx\.golangci.yml | .golangci.yml | YAML | 3,999 | 0.85 | 0 | 0 | vue-tools | 288 | 2023-11-03T18:02:18.852358 | MIT | false | 91566f0c2d93e85db8ebe8cf0a888fea |
site_name: Ingress-Nginx Controller\nrepo_name: "kubernetes/ingress-nginx"\nrepo_url: https://github.com/kubernetes/ingress-nginx\nsite_url: https://kubernetes.github.io/ingress-nginx\nedit_uri: edit/main/docs/\n\n# Extensions\nmarkdown_extensions:\n - admonition\n - abbr\n - attr_list\n - def_list\n - footnotes\n - meta\n - md_in_html\n - toc:\n # insert a blank space before the character\n permalink: " ¶"\n - pymdownx.arithmatex:\n generic: true\n - pymdownx.betterem:\n smart_enable: all\n - pymdownx.caret\n - pymdownx.critic\n - pymdownx.details\n - pymdownx.emoji:\n emoji_index: !!python/name:materialx.emoji.twemoji\n emoji_generator: !!python/name:materialx.emoji.to_svg\n - pymdownx.highlight\n - pymdownx.inlinehilite\n - pymdownx.keys\n - pymdownx.mark\n - pymdownx.smartsymbols\n - pymdownx.snippets:\n check_paths: true\n - pymdownx.superfences:\n custom_fences:\n - name: mermaid\n class: mermaid\n format: !!python/name:pymdownx.superfences.fence_code_format\n - pymdownx.tabbed\n - pymdownx.tasklist:\n custom_checkbox: true\n - pymdownx.tilde\n\ntheme:\n name: material\n features:\n - navigation.tabs\n - navigation.tabs.sticky\n - navigation.instant\n - navigation.sections\n - content.code.copy\n\n palette:\n primary: "teal"\n accent: "green"\n\n include_sidebar: true\n\nplugins:\n - search\n - awesome-pages\n - minify:\n minify_html: true\n\nextra_css: [extra.css]\n\n\nnav:\n - Welcome:\n - Welcome: "index.md"\n - How it works: "how-it-works.md"\n - Troubleshooting: "troubleshooting.md"\n - kubectl plugin: "kubectl-plugin.md"\n - Deployment:\n - Installation Guide: "deploy/index.md"\n - Bare-metal considerations: "deploy/baremetal.md"\n - Role Based Access Control (RBAC): "deploy/rbac.md"\n - Upgrade: "deploy/upgrade.md"\n - Hardening guide: "deploy/hardening-guide.md"\n - User Guide:\n - NGINX Configuration:\n - Introduction: "user-guide/nginx-configuration/index.md"\n - Basic usage: "user-guide/basic-usage.md"\n - Annotations: "user-guide/nginx-configuration/annotations.md"\n - Annotations Risks: "user-guide/nginx-configuration/annotations-risk.md"\n - ConfigMap: "user-guide/nginx-configuration/configmap.md"\n - Custom NGINX template: "user-guide/nginx-configuration/custom-template.md"\n - Log format: "user-guide/nginx-configuration/log-format.md"\n - Command line arguments: "user-guide/cli-arguments.md"\n - Custom errors: "user-guide/custom-errors.md"\n - Default backend: "user-guide/default-backend.md"\n - Exposing TCP and UDP services: "user-guide/exposing-tcp-udp-services.md"\n - Exposing FCGI services: "user-guide/fcgi-services.md"\n - Regular expressions in paths: user-guide/ingress-path-matching.md\n - External Articles: "user-guide/external-articles.md"\n - Miscellaneous: "user-guide/miscellaneous.md"\n - Prometheus and Grafana installation: "user-guide/monitoring.md"\n - Multiple Ingress controllers: "user-guide/multiple-ingress.md"\n - TLS/HTTPS: "user-guide/tls.md"\n - Third party addons:\n - ModSecurity Web Application Firewall: "user-guide/third-party-addons/modsecurity.md"\n - OpenTelemetry: "user-guide/third-party-addons/opentelemetry.md"\n - Examples:\n - Introduction: "examples/index.md"\n - Prerequisites: "examples/PREREQUISITES.md"\n - Sticky Sessions: "examples/affinity/cookie/README.md"\n - Auth:\n - Basic Authentication: "examples/auth/basic/README.md"\n - Client Certificate Authentication: "examples/auth/client-certs/README.md"\n - External Basic Authentication: "examples/auth/external-auth/README.md"\n - External OAUTH Authentication: "examples/auth/oauth-external-auth/README.md"\n - Customization:\n - Configuration Snippets: "examples/customization/configuration-snippets/README.md"\n - Custom Configuration: "examples/customization/custom-configuration/README.md"\n - Custom Errors: "examples/customization/custom-errors/README.md"\n - Custom Headers: "examples/customization/custom-headers/README.md"\n - External authentication: "examples/customization/external-auth-headers/README.md"\n - Custom DH parameters for perfect forward secrecy: "examples/customization/ssl-dh-param/README.md"\n - Sysctl tuning: "examples/customization/sysctl/README.md"\n - Docker registry: "examples/docker-registry/README.md"\n - gRPC: "examples/grpc/README.md"\n - Multi TLS certificate termination: "examples/multi-tls/README.md"\n - Rewrite: "examples/rewrite/README.md"\n - Static IPs: "examples/static-ip/README.md"\n - TLS termination: "examples/tls-termination/README.md"\n - Open Policy Agent rules: "examples/openpolicyagent/README.md"\n - Canary Deployments: "examples/canary/README.md"\n - Developer Guide:\n - Getting Started: "developer-guide/getting-started.md"\n - Code Overview: "developer-guide/code-overview.md"\n - FAQ: "faq.md"\n | dataset_sample\yaml\kubernetes_ingress-nginx\mkdocs.yml | mkdocs.yml | YAML | 5,099 | 0.8 | 0.015038 | 0.016 | python-kit | 622 | 2024-07-04T16:37:31.097223 | MIT | false | 9ad9013edf93adfe09aa0bccc57318b2 |
version: 2\nupdates:\n - package-ecosystem: "github-actions"\n directory: "/"\n schedule:\n interval: "weekly"\n labels:\n - "area/dependency"\n - "release-note-none"\n - "ok-to-test"\n groups:\n actions:\n update-types:\n - "minor"\n - "patch"\n - package-ecosystem: "docker"\n directories:\n - "**/rootfs"\n schedule:\n interval: "weekly"\n labels:\n - "area/dependency"\n - "release-note-none"\n - "ok-to-test"\n groups:\n docker:\n update-types:\n - "minor"\n - "patch"\n - package-ecosystem: "gomod"\n directories:\n - "/"\n - "**/rootfs"\n schedule:\n interval: "weekly"\n labels:\n - "area/dependency"\n - "release-note-none"\n - "ok-to-test"\n groups:\n go:\n update-types:\n - "patch"\n | dataset_sample\yaml\kubernetes_ingress-nginx\.github\dependabot.yml | dependabot.yml | YAML | 847 | 0.7 | 0 | 0 | awesome-app | 82 | 2024-03-20T16:28:15.170645 | Apache-2.0 | false | 1e18f657ece072d69d87ee6d7e668fb5 |
# action.yml\nname: 'Deploy MkDocs'\ndescription: 'Deploys MkDocs site'\nbranding:\n icon: 'arrow-up-circle'\n color: 'orange'\nruns:\n using: 'docker'\n image: 'Dockerfile'\n | dataset_sample\yaml\kubernetes_ingress-nginx\.github\actions\mkdocs\action.yml | action.yml | YAML | 170 | 0.8 | 0 | 0.111111 | react-lib | 152 | 2023-10-22T10:46:59.541250 | GPL-3.0 | false | 37245a7f4825999b5eaa14e9fe3cac0b |
contact_links:\n - name: Support Request/Question \n url: https://kubernetes.slack.com/messages/ingress-nginx/\n about: Support request or question relating to Ingress NGINX in Kubernetes Slack\n | dataset_sample\yaml\kubernetes_ingress-nginx\.github\ISSUE_TEMPLATE\config.yml | config.yml | YAML | 198 | 0.8 | 0 | 0 | awesome-app | 954 | 2023-11-09T18:08:29.758827 | Apache-2.0 | false | db38050872b9c1f6e603e05a514a3ab2 |
name: golangci-lint\n\non:\n pull_request:\n paths:\n - '**/*.go'\n - '.github/workflows/golangci-lint.yml'\n\npermissions:\n contents: read\n\njobs:\n golangci:\n name: lint\n runs-on: ubuntu-latest\n steps:\n - name: Checkout\n uses: actions/checkout@11bd71901bbe5b1630ceea73d27597364c9af683 # v4.2.2\n\n - name: Get go version\n run: echo "GOLANG_VERSION=$(cat GOLANG_VERSION)" >> $GITHUB_ENV\n\n - name: Set up Go\n id: go\n uses: actions/setup-go@0aaccfd150d50ccaeb58ebd88d36e91967a5f35b # v5.4.0\n with:\n go-version: ${{ env.GOLANG_VERSION }}\n check-latest: true\n\n - name: golangci-lint\n uses: golangci/golangci-lint-action@1481404843c368bc19ca9406f87d6e0fc97bdcfd # v7.0.0\n with:\n only-new-issues: true\n | dataset_sample\yaml\kubernetes_ingress-nginx\.github\workflows\golangci-lint.yml | golangci-lint.yml | YAML | 806 | 0.8 | 0 | 0 | awesome-app | 848 | 2024-11-06T17:46:24.549204 | MIT | false | ce275b63364f65f89e42a4290235b1ba |
name: Adds all issues\n\non:\n issues:\n types:\n - opened\n\njobs:\n add-to-project:\n name: Add issue to project\n runs-on: ubuntu-latest\n permissions:\n repository-projects: write\n issues: write\n steps:\n - uses: actions/add-to-project@244f685bbc3b7adfa8466e08b698b5577571133e # v1.0.2\n with:\n project-url: https://github.com/orgs/kubernetes/projects/104\n github-token: ${{ secrets.PROJECT_WRITER }}\n | dataset_sample\yaml\kubernetes_ingress-nginx\.github\workflows\project.yml | project.yml | YAML | 453 | 0.8 | 0 | 0 | node-utils | 905 | 2025-07-05T14:33:28.926248 | BSD-3-Clause | false | 3bd3307b29a0f85b3890d56843166c52 |
name: Scorecards supply-chain security\n\non:\n # Only the default branch is supported.\n branch_protection_rule:\n schedule:\n - cron: '20 11 * * 5'\n push:\n branches:\n - "main"\n\n# Declare default permissions as read only.\npermissions: read-all\n\njobs:\n analysis:\n name: Scorecards analysis\n runs-on: ubuntu-latest\n permissions:\n # Needed to upload the results to code-scanning dashboard.\n security-events: write\n # Used to receive a badge. (Upcoming feature)\n id-token: write\n # Needs for private repositories.\n contents: read\n actions: read\n\n steps:\n - name: "Checkout code"\n uses: actions/checkout@11bd71901bbe5b1630ceea73d27597364c9af683 # v4.2.2\n with:\n persist-credentials: false\n\n - name: "Run analysis"\n uses: ossf/scorecard-action@f49aabe0b5af0936a0987cfb85d86b75731b0186 # v2.4.1\n with:\n results_file: results.sarif\n results_format: sarif\n # (Optional) Read-only PAT token. Uncomment the `repo_token` line below if:\n # - you want to enable the Branch-Protection check on a *public* repository, or\n # - you are installing Scorecards on a *private* repository\n # To create the PAT, follow the steps in https://github.com/ossf/scorecard-action#authentication-with-pat.\n # repo_token: ${{ secrets.SCORECARD_READ_TOKEN }}\n\n # Publish the results for public repositories to enable scorecard badges. For more details, see\n # https://github.com/ossf/scorecard-action#publishing-results.\n # For private repositories, `publish_results` will automatically be set to `false`, regardless\n # of the value entered here.\n publish_results: true\n\n # Upload the results as artifacts (optional). Commenting out will disable uploads of run results in SARIF\n # format to the repository Actions tab.\n - name: "Upload artifact"\n uses: actions/upload-artifact@ea165f8d65b6e75b540449e92b4886f43607fa02 # v4.6.2\n with:\n name: SARIF file\n path: results.sarif\n retention-days: 5\n\n # Upload the results to GitHub's code scanning dashboard.\n - name: "Upload to code-scanning"\n uses: github/codeql-action/upload-sarif@45775bd8235c68ba998cffa5171334d58593da47 # v3.28.15\n with:\n sarif_file: results.sarif\n | dataset_sample\yaml\kubernetes_ingress-nginx\.github\workflows\scorecards.yml | scorecards.yml | YAML | 2,382 | 0.8 | 0.046875 | 0.303571 | vue-tools | 649 | 2024-07-20T13:40:22.263867 | Apache-2.0 | false | ae84db63af5c346644f8f61ba6b08b36 |
site_name: kOps - Kubernetes Operations\n# strict: true\nrepo_name: 'kubernetes/kops'\nrepo_url: 'https://github.com/kubernetes/kops'\nsite_url: 'https://kops.sigs.k8s.io'\nmarkdown_extensions:\n - admonition\n - codehilite\n - pymdownx.inlinehilite\n - pymdownx.tasklist:\n custom_checkbox: true\n - pymdownx.superfences\n - pymdownx.tilde\n - toc:\n permalink: ' ¶'\ntheme:\n name: material\n features:\n - navigation.tabs\n icon:\n logo: 'material/cloud-outline'\n favicon: 'img/logo-notext.svg'\n palette:\n # Palette toggle for automatic mode\n - media: "(prefers-color-scheme)"\n toggle:\n icon: material/brightness-auto\n name: Switch to light mode\n\n # Palette toggle for light mode\n - media: "(prefers-color-scheme: light)"\n scheme: default\n primary: 'teal'\n accent: 'green'\n toggle:\n icon: material/brightness-7\n name: Switch to dark mode\n\n # Palette toggle for dark mode\n - media: "(prefers-color-scheme: dark)"\n scheme: slate\n toggle:\n icon: material/brightness-4\n name: Match OS theme (Automatic)\n\nplugins:\n - search\n - macros:\n module_name: hack/mkdocs_macros/feature_stability_table\nextra_css: [extra.css]\n\nnav:\n - Welcome:\n - Welcome: "index.md"\n - Releases & Versioning: "welcome/releases.md"\n - Office Hours: "welcome/office_hours.md"\n - Values: "values.md"\n - Getting Started:\n - Installing: "getting_started/install.md"\n - Deploying to AWS: "getting_started/aws.md"\n - Deploying to GCE: "getting_started/gce.md"\n - Deploying to Digital Ocean - Beta: "getting_started/digitalocean.md"\n - Deploying to Hetzner - Beta: "getting_started/hetzner.md"\n - Deploying to OpenStack - Beta: "getting_started/openstack.md"\n - Deploying to Azure - Alpha: "getting_started/azure.md"\n - Deploying to Spot Ocean - Alpha: "getting_started/spot-ocean.md"\n - kOps Commands: "getting_started/commands.md"\n - kOps Arguments: "getting_started/arguments.md"\n - kubectl usage: "getting_started/kubectl.md"\n - Production setup: "getting_started/production.md"\n - CLI:\n - kops: "cli/kops.md"\n - kops completion: "cli/kops_completion.md"\n - kops create: "cli/kops_create.md"\n - kops delete: "cli/kops_delete.md"\n - kops distrust: "cli/kops_distrust.md"\n - kops edit: "cli/kops_edit.md"\n - kops export: "cli/kops_export.md"\n - kops get: "cli/kops_get.md"\n - kops promote: "cli/kops_promote.md"\n - kops replace: "cli/kops_replace.md"\n - kops rolling-update: "cli/kops_rolling-update.md"\n - kops toolbox: "cli/kops_toolbox.md"\n - kops trust: "cli/kops_trust.md"\n - kops update: "cli/kops_update.md"\n - kops upgrade: "cli/kops_upgrade.md"\n - kops validate: "cli/kops_validate.md"\n - kops version: "cli/kops_version.md"\n - API:\n - Cluster Resource: "cluster_spec.md"\n - InstanceGroup Resource: "instance_groups.md"\n - Addons:\n - Addons: "addons.md"\n - Operations:\n - Updates & Upgrades: "operations/updates_and_upgrades.md"\n - Rolling Updates: "operations/rolling-update.md"\n - Working with Instance Groups: "tutorial/working-with-instancegroups.md"\n - Using Manifests and Customizing: "manifests_and_customizing_via_api.md"\n - High Availability: "operations/high_availability.md"\n - Scaling: "operations/scaling.md"\n - Karpenter: "operations/karpenter.md"\n - Local asset repositories: "operations/asset-repository.md"\n - Instancegroup images: "operations/images.md"\n - Cluster configuration management: "changing_configuration.md"\n - Cluster Templating: "operations/cluster_template.md"\n - GPU setup: "gpu.md"\n - Label management: "labels.md"\n - Rotate Secrets: "operations/rotate-secrets.md"\n - Service Account Issuer Migration: "operations/service_account_issuer_migration.md"\n - Service Account Token Volume: "operations/service_account_token_volumes.md"\n - Moving from a Single Master to Multiple HA Masters: "single-to-multi-master.md"\n - Running kOps in a CI environment: "continuous_integration.md"\n - Gossip DNS: "gossip.md"\n - etcd:\n - etcd administration: "operations/etcd_administration.md"\n - etcd backup, restore and encryption: "operations/etcd_backup_restore_encryption.md"\n - Moving from a Single Master to Multiple HA Masters: "single-to-multi-master.md"\n - etcd3 Migration: "etcd3-migration.md"\n - Troubleshooting: "operations/troubleshoot.md"\n\n - Networking:\n - Networking Overview: "networking.md"\n - CNI:\n - AWS VPC: "networking/aws-vpc.md"\n - Calico: "networking/calico.md"\n - Canal: "networking/canal.md"\n - Cilium: "networking/cilium.md"\n - Flannel: "networking/flannel.md"\n - Kube-Router: "networking/kube-router.md"\n - IPv6: "networking/ipv6.md"\n - Run kOps in an existing VPC: "run_in_existing_vpc.md"\n - Supported network topologies: "topology.md"\n - Subdomain setup: "creating_subdomain.md"\n - Security:\n - Security: "security.md"\n - Advisories: "advisories/README.md"\n - Bastion setup: "bastion.md"\n - Instance IAM roles: "iam_roles.md"\n - MFA setup: "mfa.md"\n - Security Groups: "security_groups.md"\n - Advanced:\n - Download Config: "advanced/download_config.md"\n - Subdomain NS Records: "advanced/ns.md"\n - Experimental: "advanced/experimental.md"\n - Cluster boot sequence: "boot-sequence.md"\n - Philosophy: "philosophy.md"\n - State store: "state.md"\n - AWS China: "aws-china.md"\n - Custom CA: "custom_ca.md"\n - Horizontal Pod Autoscaling: "horizontal_pod_autoscaling.md"\n - Egress Proxy: "http_proxy.md"\n - Node Resource Allocation: "node_resource_handling.md"\n - Terraform: "terraform.md"\n - Authentication: "authentication.md"\n - Contributing:\n - Getting Involved and Contributing: "contributing/index.md"\n - New Kubernetes Version: "contributing/new_kubernetes_version.md"\n - Our Release Process: "contributing/release-process.md"\n - Releasing With Homebrew: "contributing/homebrew.md"\n - Updating The Default Base AMI: "contributing/update_ami_versions.md"\n - Building: "contributing/building.md"\n - Adding a feature: "contributing/adding_a_feature.md"\n - Testing: "contributing/testing.md"\n - Testing preview versions: "contributing/test_versions.md"\n - Developing using Docker: "contributing/Docker.md"\n - Documentation Guidelines: "contributing/documentation.md"\n - Hack Directory: "contributing/hack.md"\n - How to update kOps API: "contributing/api_updates.md"\n - Low level description on how kOps works: "contributing/how_it_works.md"\n - Notes on Gossip design: "contributing/gossip.md"\n - Notes on master instance sizing: "contributing/instancesizes.md"\n - Vendoring: "contributing/vendoring.md"\n - Ports: "contributing/ports.md"\n - Cluster Addons & Manager : "contributing/addons.md"\n - Releases:\n - "1.30": releases/1.30-NOTES.md\n - "1.29": releases/1.29-NOTES.md\n - "1.28": releases/1.28-NOTES.md\n - "1.27": releases/1.27-NOTES.md\n - "1.26": releases/1.26-NOTES.md\n - "1.25": releases/1.25-NOTES.md\n - "1.24": releases/1.24-NOTES.md\n - "1.23": releases/1.23-NOTES.md\n - "1.22": releases/1.22-NOTES.md\n - "1.21": releases/1.21-NOTES.md\n - "1.20": releases/1.20-NOTES.md\n - "1.19": releases/1.19-NOTES.md\n - "1.18": releases/1.18-NOTES.md\n - "1.17": releases/1.17-NOTES.md\n - "1.16": releases/1.16-NOTES.md\n - "1.15": releases/1.15-NOTES.md\n - "1.14": releases/1.14-NOTES.md\n - "1.13": releases/1.13-NOTES.md\n - "1.12": releases/1.12-NOTES.md\n - "1.11": releases/1.11-NOTES.md\n - "1.10": releases/1.10-NOTES.md\n - "1.9": releases/1.9-NOTES.md\n - "1.8": releases/1.8-NOTES.md\n - "1.7": releases/1.7-NOTES.md\n - "1.6": releases/1.6-NOTES.md\n - "1.4": releases/1.4-NOTES.md\n | dataset_sample\yaml\kubernetes_kops\mkdocs.yml | mkdocs.yml | YAML | 7,848 | 0.8 | 0.014925 | 0.020408 | vue-tools | 308 | 2024-10-13T11:39:30.899045 | GPL-3.0 | false | 32576d98a262dd404ecad761d78beb7f |
version: 2\nupdates:\n- package-ecosystem: "github-actions"\n directory: "/"\n schedule:\n interval: "weekly"\n day: "sunday"\n labels:\n - "ok-to-test"\n | dataset_sample\yaml\kubernetes_kops\.github\dependabot.yml | dependabot.yml | YAML | 155 | 0.7 | 0 | 0 | python-kit | 447 | 2025-04-15T04:31:14.900328 | Apache-2.0 | false | 343780eb350cc79ff86eff121cfa946c |
---\nname: e2e\n\n'on':\n - push\n - pull_request\n\nenv:\n GOPROXY: https://proxy.golang.org\n GOPATH: ${{ github.workspace }}/go\n\npermissions:\n contents: read\n\njobs:\n tests-e2e-scenarios-bare-metal:\n runs-on: ubuntu-24.04\n timeout-minutes: 70\n steps:\n - uses: actions/checkout@11bd71901bbe5b1630ceea73d27597364c9af683\n with:\n path: ${{ env.GOPATH }}/src/k8s.io/kops\n\n - name: Set up go\n uses: actions/setup-go@0aaccfd150d50ccaeb58ebd88d36e91967a5f35b\n with:\n go-version-file: '${{ env.GOPATH }}/src/k8s.io/kops/go.mod'\n\n - name: tests/e2e/scenarios/bare-metal/run-test\n working-directory: ${{ env.GOPATH }}/src/k8s.io/kops\n run: |\n timeout 60m tests/e2e/scenarios/bare-metal/run-test\n env:\n ARTIFACTS: /tmp/artifacts\n - name: Archive production artifacts\n if: always()\n uses: actions/upload-artifact@v4\n with:\n name: tests-e2e-scenarios-bare-metal\n path: /tmp/artifacts/\n | dataset_sample\yaml\kubernetes_kops\.github\workflows\e2e.yml | e2e.yml | YAML | 1,016 | 0.8 | 0.025 | 0 | python-kit | 409 | 2024-10-22T03:04:56.268958 | MIT | false | a589e50e21740975f167b7eb860d35e3 |
---\nname: CI\n\n"on":\n - push\n - pull_request\n\nenv:\n GOPROXY: https://proxy.golang.org\n GOPATH: ${{ github.workspace }}/go\n\npermissions:\n contents: read\n\njobs:\n build-linux:\n strategy:\n matrix:\n os:\n - name: Linux-amd64\n runs-on: ubuntu-24.04\n arch: amd64\n - name: Linux-arm64\n runs-on: ubuntu-24.04-arm\n arch: arm64\n name: build-linux-${{ matrix.os.arch }}\n runs-on: ${{ matrix.os.runs-on }}\n steps:\n - uses: actions/checkout@11bd71901bbe5b1630ceea73d27597364c9af683\n with:\n path: ${{ env.GOPATH }}/src/k8s.io/kops\n\n - name: Set up go\n uses: actions/setup-go@0aaccfd150d50ccaeb58ebd88d36e91967a5f35b\n with:\n go-version-file: "${{ env.GOPATH }}/src/k8s.io/kops/go.mod"\n\n - name: make all examples test\n working-directory: ${{ env.GOPATH }}/src/k8s.io/kops\n run: |\n make all examples test\n\n build-macos-amd64:\n runs-on: macos-latest\n steps:\n - uses: actions/checkout@11bd71901bbe5b1630ceea73d27597364c9af683\n with:\n path: ${{ env.GOPATH }}/src/k8s.io/kops\n\n - name: Set up go\n uses: actions/setup-go@0aaccfd150d50ccaeb58ebd88d36e91967a5f35b\n with:\n go-version-file: "${{ env.GOPATH }}/src/k8s.io/kops/go.mod"\n\n - name: make kops examples test\n working-directory: ${{ env.GOPATH }}/src/k8s.io/kops\n run: |\n make kops examples test\n\n build-windows-amd64:\n runs-on: windows-2022\n steps:\n - uses: actions/checkout@11bd71901bbe5b1630ceea73d27597364c9af683\n with:\n path: ${{ env.GOPATH }}/src/k8s.io/kops\n\n - name: Set up go\n uses: actions/setup-go@0aaccfd150d50ccaeb58ebd88d36e91967a5f35b\n with:\n go-version-file: "${{ env.GOPATH }}/src/k8s.io/kops/go.mod"\n\n - name: make kops examples test\n working-directory: ${{ env.GOPATH }}/src/k8s.io/kops\n run: |\n make kops examples test-windows\n\n verify:\n strategy:\n matrix:\n os:\n - name: Linux-amd64\n runs-on: ubuntu-24.04\n arch: amd64\n - name: Linux-arm64\n runs-on: ubuntu-24.04-arm\n arch: arm64\n name: verify-${{ matrix.os.arch }}\n runs-on: ${{ matrix.os.runs-on }}\n steps:\n - uses: actions/checkout@11bd71901bbe5b1630ceea73d27597364c9af683\n with:\n path: ${{ env.GOPATH }}/src/k8s.io/kops\n\n - name: Set up go\n uses: actions/setup-go@0aaccfd150d50ccaeb58ebd88d36e91967a5f35b\n with:\n go-version-file: "${{ env.GOPATH }}/src/k8s.io/kops/go.mod"\n\n - name: make quick-ci\n working-directory: ${{ env.GOPATH }}/src/k8s.io/kops\n run: |\n make quick-ci\n | dataset_sample\yaml\kubernetes_kops\.github\workflows\main.yml | main.yml | YAML | 2,780 | 0.8 | 0 | 0 | awesome-app | 472 | 2024-06-10T22:38:50.538035 | Apache-2.0 | false | 0e967143779fa212e69c985f4924072f |
name: 'Tag Release'\n\non:\n push:\n branches:\n - master\n - 'release-*'\n paths:\n - kops-version.go\n\njobs:\n tag-release:\n if: ${{ github.repository == 'kubernetes/kops' }}\n runs-on: ubuntu-24.04\n\n permissions:\n contents: write\n\n steps:\n - uses: actions/checkout@11bd71901bbe5b1630ceea73d27597364c9af683\n - run: /usr/bin/git config --global user.email actions@github.com\n - run: /usr/bin/git config --global user.name 'GitHub Actions Release Tagger'\n - run: hack/tag-release.sh\n | dataset_sample\yaml\kubernetes_kops\.github\workflows\tag-release.yml | tag-release.yml | YAML | 532 | 0.7 | 0.043478 | 0 | python-kit | 696 | 2023-11-13T12:37:06.258198 | Apache-2.0 | false | ee3ee0bfd663442f08d1b5c8f4413da9 |
language: go\narch:\n - amd64\n - ppc64le\ninstall:\n - go get -t\n - go get golang.org/x/tools/cmd/cover\n - go get github.com/mattn/goveralls\nscript:\n - go test -race -v ./...\nafter_script:\n - $HOME/gopath/bin/goveralls -service=travis-ci -repotoken $COVERALLS_TOKEN\n | dataset_sample\yaml\kubernetes_kops\vendor\dario.cat\mergo\.travis.yml | .travis.yml | YAML | 273 | 0.7 | 0 | 0 | node-utils | 637 | 2024-03-23T02:14:55.174289 | MIT | false | 326cc9a952683630b4cdd8b1eb75897f |
language: go\n\ngo:\n - "1.x"\n\nenv:\n - GO111MODULE=on\n\ninstall:\n - go get ./...\n\nscript:\n - go test ./...\n | dataset_sample\yaml\kubernetes_kops\vendor\github.com\armon\go-metrics\.travis.yml | .travis.yml | YAML | 107 | 0.7 | 0 | 0 | react-lib | 173 | 2024-04-26T09:18:12.316597 | BSD-3-Clause | false | f6c5d59e8922b63e6bd0c153fc7e6e3d |
language: go\n\nos:\n - linux\n - osx\n - windows\n\ngo:\n - go1.13.x\n - go1.x\n\nservices:\n - xvfb\n\nbefore_install:\n - export DISPLAY=:99.0\n\nscript:\n - if [ "$TRAVIS_OS_NAME" = "linux" ]; then sudo apt-get install xsel; fi\n - go test -v .\n - if [ "$TRAVIS_OS_NAME" = "linux" ]; then sudo apt-get install xclip; fi\n - go test -v .\n | dataset_sample\yaml\kubernetes_kops\vendor\github.com\atotto\clipboard\.travis.yml | .travis.yml | YAML | 321 | 0.7 | 0.090909 | 0 | node-utils | 324 | 2024-04-23T19:52:19.307275 | MIT | false | a07df666b82ed543e17afd7ed9f8806f |
language: go\nsudo: true\ndist: bionic\n\nbranches:\n only:\n - main\n\nos:\n - linux\n - osx\n # Travis doesn't work with windows and Go tip\n #- windows\n\ngo:\n - tip\n\nmatrix:\n allow_failures:\n - go: tip\n\nbefore_install:\n - if [ "$TRAVIS_OS_NAME" = "windows" ]; then choco install make; fi\n - (cd /tmp/; go get golang.org/x/lint/golint)\n\nscript:\n - make go test -v ./...;\n\n | dataset_sample\yaml\kubernetes_kops\vendor\github.com\aws\smithy-go\.travis.yml | .travis.yml | YAML | 377 | 0.8 | 0.035714 | 0.095238 | python-kit | 685 | 2025-05-21T07:51:24.612323 | BSD-3-Clause | false | b615be337e1992846d32c080fc02c3fe |
# NOTE: Please refer to https://aka.ms/azsdk/engsys/ci-yaml before editing this file.\ntrigger:\n branches:\n include:\n - main\n - feature/*\n - hotfix/*\n - release/*\n paths:\n include:\n - sdk/azcore/\n - eng/\n\npr:\n branches:\n include:\n - main\n - feature/*\n - hotfix/*\n - release/*\n paths:\n include:\n - sdk/azcore/\n - eng/\n\nextends:\n template: /eng/pipelines/templates/jobs/archetype-sdk-client.yml\n parameters:\n ServiceDirectory: azcore\n | dataset_sample\yaml\kubernetes_kops\vendor\github.com\Azure\azure-sdk-for-go\sdk\azcore\ci.yml | ci.yml | YAML | 504 | 0.8 | 0 | 0.037037 | python-kit | 218 | 2024-03-09T17:14:31.314092 | Apache-2.0 | false | 8c91b6a80332e71c293a8aa5448000f1 |
# NOTE: Please refer to https://aka.ms/azsdk/engsys/ci-yaml before editing this file.\ntrigger:\n branches:\n include:\n - main\n - feature/*\n - hotfix/*\n - release/*\n paths:\n include:\n - sdk/azidentity/\n\npr:\n branches:\n include:\n - main\n - feature/*\n - hotfix/*\n - release/*\n paths:\n include:\n - sdk/azidentity/\n\nextends:\n template: /eng/pipelines/templates/jobs/archetype-sdk-client.yml\n parameters:\n CloudConfig:\n Public:\n SubscriptionConfigurations:\n - $(sub-config-identity-test-resources)\n EnableRaceDetector: true\n Location: westus2\n RunLiveTests: true\n ServiceDirectory: azidentity\n UsePipelineProxy: false\n\n ${{ if endsWith(variables['Build.DefinitionName'], 'weekly') }}:\n PersistOidcToken: true\n MatrixConfigs:\n - Name: managed_identity_matrix\n GenerateVMJobs: true\n Path: sdk/azidentity/managed-identity-matrix.json\n Selection: sparse\n MatrixReplace:\n - Pool=.*LINUXPOOL.*/azsdk-pool-mms-ubuntu-2204-identitymsi\n - OSVmImage=.*LINUXNEXTVMIMAGE.*/azsdk-pool-mms-ubuntu-2204-1espt\n | dataset_sample\yaml\kubernetes_kops\vendor\github.com\Azure\azure-sdk-for-go\sdk\azidentity\ci.yml | ci.yml | YAML | 1,201 | 0.8 | 0.021739 | 0.023256 | react-lib | 908 | 2024-07-13T12:38:10.647005 | MIT | false | 1a377d92573c3c389b22c94165c1e87b |
# NOTE: Please refer to https://aka.ms/azsdk/engsys/ci-yaml before editing this file.\ntrigger:\n branches:\n include:\n - main\n - feature/*\n - hotfix/*\n - release/*\n paths:\n include:\n - sdk/resourcemanager/authorization/armauthorization/\n\npr:\n branches:\n include:\n - main\n - feature/*\n - hotfix/*\n - release/*\n paths:\n include:\n - sdk/resourcemanager/authorization/armauthorization/\n\nstages:\n- template: /eng/pipelines/templates/jobs/archetype-sdk-client.yml\n parameters:\n IncludeRelease: true\n ServiceDirectory: 'resourcemanager/authorization/armauthorization'\n | dataset_sample\yaml\kubernetes_kops\vendor\github.com\Azure\azure-sdk-for-go\sdk\resourcemanager\authorization\armauthorization\v3\ci.yml | ci.yml | YAML | 628 | 0.8 | 0 | 0.038462 | vue-tools | 275 | 2024-01-05T11:59:37.790827 | MIT | false | f062bcd17e9feaea9db004ca08e035e8 |
# NOTE: Please refer to https://aka.ms/azsdk/engsys/ci-yaml before editing this file.\ntrigger:\n branches:\n include:\n - main\n - feature/*\n - hotfix/*\n - release/*\n paths:\n include:\n - sdk/resourcemanager/compute/armcompute/\n\npr:\n branches:\n include:\n - main\n - feature/*\n - hotfix/*\n - release/*\n paths:\n include:\n - sdk/resourcemanager/compute/armcompute/\n\nstages:\n- template: /eng/pipelines/templates/jobs/archetype-sdk-client.yml\n parameters:\n IncludeRelease: true\n ServiceDirectory: 'resourcemanager/compute/armcompute'\n | dataset_sample\yaml\kubernetes_kops\vendor\github.com\Azure\azure-sdk-for-go\sdk\resourcemanager\compute\armcompute\ci.yml | ci.yml | YAML | 592 | 0.8 | 0 | 0.038462 | python-kit | 889 | 2023-11-30T07:42:15.380930 | BSD-3-Clause | false | d32ca955000d19430061e3d95883fa3f |
# NOTE: Please refer to https://aka.ms/azsdk/engsys/ci-yaml before editing this file.\ntrigger:\n branches:\n include:\n - main\n - feature/*\n - hotfix/*\n - release/*\n paths:\n include:\n - sdk/resourcemanager/network/armnetwork/\n\npr:\n branches:\n include:\n - main\n - feature/*\n - hotfix/*\n - release/*\n paths:\n include:\n - sdk/resourcemanager/network/armnetwork/\n\nstages:\n- template: /eng/pipelines/templates/jobs/archetype-sdk-client.yml\n parameters:\n IncludeRelease: true\n ServiceDirectory: 'resourcemanager/network/armnetwork'\n | dataset_sample\yaml\kubernetes_kops\vendor\github.com\Azure\azure-sdk-for-go\sdk\resourcemanager\network\armnetwork\ci.yml | ci.yml | YAML | 592 | 0.8 | 0 | 0.038462 | node-utils | 374 | 2024-10-17T05:58:52.270095 | MIT | false | 0a5cfefb9a071850cde6bb79d82f8c62 |
# NOTE: Please refer to https://aka.ms/azsdk/engsys/ci-yaml before editing this file.\ntrigger:\n branches:\n include:\n - main\n - feature/*\n - hotfix/*\n - release/*\n paths:\n include:\n - sdk/resourcemanager/resources/armresources/\n\npr:\n branches:\n include:\n - main\n - feature/*\n - hotfix/*\n - release/*\n paths:\n include:\n - sdk/resourcemanager/resources/armresources/\n\nstages:\n- template: /eng/pipelines/templates/jobs/archetype-sdk-client.yml\n parameters:\n IncludeRelease: true\n ServiceDirectory: 'resourcemanager/resources/armresources'\n | dataset_sample\yaml\kubernetes_kops\vendor\github.com\Azure\azure-sdk-for-go\sdk\resourcemanager\resources\armresources\ci.yml | ci.yml | YAML | 604 | 0.8 | 0 | 0.038462 | awesome-app | 357 | 2023-12-25T11:42:35.588139 | BSD-3-Clause | false | d8a442fc56e1c378fa393c2afa2f3986 |
# NOTE: Please refer to https://aka.ms/azsdk/engsys/ci-yaml before editing this file.\ntrigger:\n branches:\n include:\n - main\n - feature/*\n - hotfix/*\n - release/*\n paths:\n include:\n - sdk/resourcemanager/storage/armstorage/\n\npr:\n branches:\n include:\n - main\n - feature/*\n - hotfix/*\n - release/*\n paths:\n include:\n - sdk/resourcemanager/storage/armstorage/\n\nextends:\n template: /eng/pipelines/templates/jobs/archetype-sdk-client.yml\n parameters:\n IncludeRelease: true\n ServiceDirectory: 'resourcemanager/storage/armstorage'\n UsePipelineProxy: false\n | dataset_sample\yaml\kubernetes_kops\vendor\github.com\Azure\azure-sdk-for-go\sdk\resourcemanager\storage\armstorage\ci.yml | ci.yml | YAML | 621 | 0.8 | 0 | 0.037037 | react-lib | 569 | 2023-08-09T05:23:19.985283 | MIT | false | eaf2092564c0c78a9096dc26ea2a5046 |
trigger:\n branches:\n include:\n - main\n - feature/*\n - hotfix/*\n - release/*\n paths:\n include:\n - sdk/storage/azblob\n\npr:\n branches:\n include:\n - main\n - feature/*\n - hotfix/*\n - release/*\n paths:\n include:\n - sdk/storage/azblob\n\n\nextends:\n template: /eng/pipelines/templates/jobs/archetype-sdk-client.yml\n parameters:\n ServiceDirectory: 'storage/azblob'\n RunLiveTests: true\n UsePipelineProxy: false\n EnvVars:\n AZURE_CLIENT_ID: $(AZBLOB_CLIENT_ID)\n AZURE_TENANT_ID: $(AZBLOB_TENANT_ID)\n AZURE_CLIENT_SECRET: $(AZBLOB_CLIENT_SECRET)\n AZURE_SUBSCRIPTION_ID: $(AZBLOB_SUBSCRIPTION_ID)\n | dataset_sample\yaml\kubernetes_kops\vendor\github.com\Azure\azure-sdk-for-go\sdk\storage\azblob\ci.yml | ci.yml | YAML | 701 | 0.8 | 0 | 0 | python-kit | 840 | 2025-05-17T20:05:05.973042 | BSD-3-Clause | false | c6a7d61310749dd70aa764d988a0708f |
run:\n tests: false\n\nissues:\n include:\n - EXC0001\n - EXC0005\n - EXC0011\n - EXC0012\n - EXC0013\n\n max-issues-per-linter: 0\n max-same-issues: 0\n\nlinters:\n enable:\n # - dupl\n - exhaustive\n # - exhaustivestruct\n - goconst\n - godot\n - godox\n - gomnd\n - gomoddirectives\n - goprintffuncname\n - ifshort\n # - lll\n - misspell\n - nakedret\n - nestif\n - noctx\n - nolintlint\n - prealloc\n - wrapcheck\n\n # disable default linters, they are already enabled in .golangci.yml\n disable:\n - deadcode\n - errcheck\n - gosimple\n - govet\n - ineffassign\n - staticcheck\n - structcheck\n - typecheck\n - unused\n - varcheck\n | dataset_sample\yaml\kubernetes_kops\vendor\github.com\charmbracelet\bubbletea\.golangci-soft.yml | .golangci-soft.yml | YAML | 696 | 0.8 | 0 | 0.093023 | node-utils | 811 | 2024-02-12T16:03:44.459828 | Apache-2.0 | false | 3c5c67b803f54b20e099c09faaefa787 |
run:\n tests: false\n\nissues:\n include:\n - EXC0001\n - EXC0005\n - EXC0011\n - EXC0012\n - EXC0013\n\n max-issues-per-linter: 0\n max-same-issues: 0\n\nlinters:\n enable:\n - bodyclose\n - exportloopref\n - goimports\n - gosec\n - nilerr\n - predeclared\n - revive\n - rowserrcheck\n - sqlclosecheck\n - tparallel\n - unconvert\n - unparam\n - whitespace\n | dataset_sample\yaml\kubernetes_kops\vendor\github.com\charmbracelet\bubbletea\.golangci.yml | .golangci.yml | YAML | 389 | 0.85 | 0 | 0 | python-kit | 482 | 2025-06-21T13:58:12.927888 | BSD-3-Clause | false | b9f89c65f4f6ecf49fd60c797e1f598e |
run:\n tests: false\n\nissues:\n include:\n - EXC0001\n - EXC0005\n - EXC0011\n - EXC0012\n - EXC0013\n\n max-issues-per-linter: 0\n max-same-issues: 0\n\nlinters:\n enable:\n # - dupl\n - exhaustive\n # - exhaustivestruct\n - goconst\n - godot\n - godox\n - gomnd\n - gomoddirectives\n - goprintffuncname\n - ifshort\n # - lll\n - misspell\n - nakedret\n - nestif\n - noctx\n - nolintlint\n - prealloc\n - wrapcheck\n\n # disable default linters, they are already enabled in .golangci.yml\n disable:\n - deadcode\n - errcheck\n - gosimple\n - govet\n - ineffassign\n - staticcheck\n - structcheck\n - typecheck\n - unused\n - varcheck\n | dataset_sample\yaml\kubernetes_kops\vendor\github.com\charmbracelet\lipgloss\.golangci-soft.yml | .golangci-soft.yml | YAML | 696 | 0.8 | 0 | 0.093023 | react-lib | 375 | 2025-03-24T11:17:29.438947 | MIT | false | 3c5c67b803f54b20e099c09faaefa787 |
run:\n tests: false\n\nissues:\n include:\n - EXC0001\n - EXC0005\n - EXC0011\n - EXC0012\n - EXC0013\n\n max-issues-per-linter: 0\n max-same-issues: 0\n\nlinters:\n enable:\n - bodyclose\n - exportloopref\n - goimports\n - gosec\n - nilerr\n - predeclared\n - revive\n - rowserrcheck\n - sqlclosecheck\n - tparallel\n - unconvert\n - unparam\n - whitespace\n | dataset_sample\yaml\kubernetes_kops\vendor\github.com\charmbracelet\lipgloss\.golangci.yml | .golangci.yml | YAML | 389 | 0.85 | 0 | 0 | react-lib | 933 | 2024-12-17T13:27:26.609885 | Apache-2.0 | false | b9f89c65f4f6ecf49fd60c797e1f598e |
linters:\n enable:\n - gofmt\n - goimports\n - ineffassign\n - misspell\n - revive\n - staticcheck\n - structcheck\n - unconvert\n - unused\n - varcheck\n - vet\n disable:\n - errcheck\n\nrun:\n timeout: 3m\n skip-dirs:\n - vendor\n | dataset_sample\yaml\kubernetes_kops\vendor\github.com\containerd\console\.golangci.yml | .golangci.yml | YAML | 255 | 0.85 | 0 | 0 | node-utils | 29 | 2023-10-20T22:11:12.203929 | Apache-2.0 | false | a11b4c6573b3396d5a5cf92605c2721a |
linters:\n enable:\n - exportloopref # Checks for pointers to enclosing loop variables\n - gofmt\n - goimports\n - gosec\n - ineffassign\n - misspell\n - nolintlint\n - revive\n - staticcheck\n - tenv # Detects using os.Setenv instead of t.Setenv since Go 1.17\n - unconvert\n - unused\n - vet\n - dupword # Checks for duplicate words in the source code\n disable:\n - errcheck\n\nrun:\n timeout: 5m\n skip-dirs:\n - api\n - cluster\n - design\n - docs\n - docs/man\n - releases\n - reports\n - test # e2e scripts\n | dataset_sample\yaml\kubernetes_kops\vendor\github.com\containerd\log\.golangci.yml | .golangci.yml | YAML | 558 | 0.95 | 0.066667 | 0 | react-lib | 684 | 2025-02-01T12:31:17.550207 | GPL-3.0 | false | b8375607b0dbb7da9b703c9efacd207f |
linters:\n enable:\n - exportloopref # Checks for pointers to enclosing loop variables\n - gofmt\n - goimports\n - gosec\n - ineffassign\n - misspell\n - nolintlint\n - revive\n - staticcheck\n - tenv # Detects using os.Setenv instead of t.Setenv since Go 1.17\n - unconvert\n - unused\n - vet\n - dupword # Checks for duplicate words in the source code\n disable:\n - errcheck\n\nrun:\n timeout: 5m\n skip-dirs:\n - api\n - cluster\n - design\n - docs\n - docs/man\n - releases\n - reports\n - test # e2e scripts\n | dataset_sample\yaml\kubernetes_kops\vendor\github.com\containerd\platforms\.golangci.yml | .golangci.yml | YAML | 558 | 0.95 | 0.066667 | 0 | node-utils | 462 | 2025-05-01T03:52:13.379516 | BSD-3-Clause | false | b8375607b0dbb7da9b703c9efacd207f |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.