content stringlengths 1 103k ⌀ | path stringlengths 8 216 | filename stringlengths 2 179 | language stringclasses 15
values | size_bytes int64 2 189k | quality_score float64 0.5 0.95 | complexity float64 0 1 | documentation_ratio float64 0 1 | repository stringclasses 5
values | stars int64 0 1k | created_date stringdate 2023-07-10 19:21:08 2025-07-09 19:11:45 | license stringclasses 4
values | is_test bool 2
classes | file_hash stringlengths 32 32 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
\nversion: '2.2'\n\nservices:\n# Example of PostgreSQL configuration, uncomment if you want to use postgres\n# postgres:\n# image: postgres:12.4\n# environment:\n# POSTGRES_USER: 'test'\n# POSTGRES_PASSWORD: 'test'\n# POSTGRES_DB: 'test'\n\n cube:\n image: cubejs/cube:latest\n #depends_on:\n # - postgres\n #links:\n # - postgres\n ports:\n - 4000:4000 # Cube.js API and Developer Playground\n - 3000:3000 # Dashboard app, if created\n env_file: .env\n volumes:\n # If you are going to use own dependencies, for example axios/vault or anything else for getting configuration\n # - .:/cube/conf\n - ./dashboard-app:/cube/conf/dashboard-app\n - ./cube.js:/cube/conf/cube.js\n - ./schema:/cube/conf/schema\n | dataset_sample\yaml\cube-js_cube\examples\react-muze\docker-compose.yml | docker-compose.yml | YAML | 763 | 0.8 | 0.142857 | 0.52 | python-kit | 346 | 2025-05-15T08:12:44.835964 | GPL-3.0 | false | ad5ffcd6643373c5a838ae2837e7dfc5 |
\nversion: '2.2'\n\nservices:\n cube:\n image: cubejs/cube:v0.26.60\n ports:\n - 4000:4000 # Cube.js API and Developer Playground\n - 3000:3000 # Dashboard app, if created\n env_file: .env\n volumes:\n - .:/cube/conf\n # We ignore Cube.js deps, because they are built-in inside the official Docker image\n - .empty:/cube/conf/node_modules/@cubejs-backend/\n | dataset_sample\yaml\cube-js_cube\examples\react-pivot-table\docker-compose.yml | docker-compose.yml | YAML | 383 | 0.8 | 0.071429 | 0.083333 | python-kit | 653 | 2025-06-10T20:45:37.602968 | Apache-2.0 | false | 6872ee9954c9e6f727b56c27425c113a |
version: '2.2'\n\nservices:\n cube:\n image: cubejs/cube:latest\n ports:\n - 4000:4000\n - 3000:3000\n env_file: .env\n volumes:\n - .:/cube/conf\n\n query:\n image: cfmanteiga/alpine-bash-curl-jq\n depends_on:\n - cube\n volumes:\n - .:/query\n entrypoint: ["sh", "query/queries/run.sh"]\n | dataset_sample\yaml\cube-js_cube\examples\recipes\active-users\docker-compose.yml | docker-compose.yml | YAML | 322 | 0.7 | 0 | 0 | vue-tools | 677 | 2023-10-17T18:01:49.764997 | MIT | false | f98aec9b754e9c666ab8f6dc73d094bc |
\nversion: '2.2'\n\nservices:\n cube:\n image: cubejs/cube:v0.31.4\n ports:\n # It's better to use random port binding for 4000/3000 ports\n # without it you will not able to start multiple projects inside docker\n - 4000:4000 # Cube.js API and Developer Playground\n - 3000:3000 # Dashboard app, if created\n env_file: .env\n volumes:\n - .:/cube/conf\n # We ignore Cube.js deps, because they are built-in inside the official Docker image\n - .empty:/cube/conf/node_modules/@cubejs-backend/\n | dataset_sample\yaml\cube-js_cube\examples\recipes\changing-visibility-of-cubes-or-views\docker-compose.yml | docker-compose.yml | YAML | 527 | 0.8 | 0.125 | 0.214286 | python-kit | 458 | 2024-03-13T21:37:12.876917 | Apache-2.0 | false | cd584be031382c82aeaaf230881d6b1a |
version: '2.2'\n\nservices:\n cube:\n image: cubejs/cube:latest\n ports:\n - 4000:4000\n - 3000:3000\n env_file: .env\n volumes:\n - .:/cube/conf\n\n query:\n image: cfmanteiga/alpine-bash-curl-jq\n depends_on:\n - cube\n volumes:\n - .:/query\n entrypoint: ["sh", "query/queries/run.sh"]\n | dataset_sample\yaml\cube-js_cube\examples\recipes\column-based-access\docker-compose.yml | docker-compose.yml | YAML | 322 | 0.7 | 0 | 0 | node-utils | 896 | 2025-03-20T15:50:58.395593 | Apache-2.0 | false | f98aec9b754e9c666ab8f6dc73d094bc |
version: '2.2'\n\nservices:\n cube:\n build:\n context: .\n dockerfile: Dockerfile\n ports:\n - 4000:4000\n - 3000:3000\n env_file: .env\n volumes:\n - .:/cube/conf\n\n query:\n image: cfmanteiga/alpine-bash-curl-jq\n depends_on:\n - cube\n volumes:\n - .:/query\n entrypoint: ["sh", "query/queries/run.sh"]\n | dataset_sample\yaml\cube-js_cube\examples\recipes\entity-attribute-value\docker-compose.yml | docker-compose.yml | YAML | 349 | 0.7 | 0 | 0 | node-utils | 969 | 2025-01-17T16:43:33.161808 | BSD-3-Clause | false | 217d7b532ac357ee962f23c975a501c6 |
version: '2.2'\n\nservices:\n cube:\n image: cubejs/cube:latest\n ports:\n - 4000:4000\n - 3000:3000\n env_file: .env\n volumes:\n - .:/cube/conf\n\n query:\n image: cfmanteiga/alpine-bash-curl-jq\n depends_on:\n - cube\n volumes:\n - .:/query\n entrypoint: ["sh", "query/queries/run.sh"]\n | dataset_sample\yaml\cube-js_cube\examples\recipes\getting-unique-values-for-a-field\docker-compose.yml | docker-compose.yml | YAML | 322 | 0.7 | 0 | 0 | react-lib | 858 | 2024-05-31T18:24:05.168202 | Apache-2.0 | false | f98aec9b754e9c666ab8f6dc73d094bc |
version: '2.2'\n\nservices:\n cubestore:\n image: cubejs/cubestore:latest\n environment:\n - CUBESTORE_REMOTE_DIR=/cube/data\n volumes:\n - .cubestore:/cube/data\n \n cube:\n image: cubejs/cube:latest\n ports:\n - 4000:4000\n - 3000:3000\n env_file: .env\n volumes:\n - .:/cube/conf\n depends_on:\n - cubestore\n links:\n - cubestore\n\n query:\n image: cfmanteiga/alpine-bash-curl-jq\n depends_on:\n - cube\n volumes:\n - .:/query\n entrypoint: ["sh", "query/queries/run.sh"]\n | dataset_sample\yaml\cube-js_cube\examples\recipes\joining-multiple-datasources-data\docker-compose.yml | docker-compose.yml | YAML | 540 | 0.7 | 0 | 0 | react-lib | 355 | 2024-10-21T14:50:41.540996 | Apache-2.0 | false | 440e032e5633ceaeb69fb3f568eea803 |
version: '2.2'\n\nservices:\n cube:\n image: cubejs/cube:latest\n ports:\n - 4000:4000\n - 3000:3000\n env_file: .env\n volumes:\n - .:/cube/conf\n\n query:\n image: cfmanteiga/alpine-bash-curl-jq\n depends_on:\n - cube\n volumes:\n - .:/query\n entrypoint: ["sh", "query/queries/run.sh"]\n | dataset_sample\yaml\cube-js_cube\examples\recipes\mandatory-filters\docker-compose.yml | docker-compose.yml | YAML | 322 | 0.7 | 0 | 0 | node-utils | 924 | 2024-08-14T09:25:55.708603 | GPL-3.0 | false | f98aec9b754e9c666ab8f6dc73d094bc |
version: '2.2'\n\nservices:\n cube:\n image: cubejs/cube:latest\n ports:\n - 4000:4000\n - 3000:3000\n env_file: .env\n volumes:\n - .:/cube/conf\n\n postgres:\n image: postgres\n restart: always\n ports:\n - 5432:5432\n environment:\n POSTGRES_PASSWORD: example\n POSTGRES_DB: localDB\n volumes:\n - ./scripts:/docker-entrypoint-initdb.d\n\n query:\n image: cfmanteiga/alpine-bash-curl-jq\n depends_on:\n - cube\n volumes:\n - .:/query\n entrypoint: ["bash", "query/queries/run.sh"]\n | dataset_sample\yaml\cube-js_cube\examples\recipes\multiple-data-sources\docker-compose.yml | docker-compose.yml | YAML | 542 | 0.7 | 0 | 0 | python-kit | 842 | 2024-04-01T05:31:48.241489 | MIT | false | d89e0d6a132ec58e712cb75309ff114b |
version: '2.2'\n\nservices:\n cube:\n image: cubejs/cube:latest\n ports:\n - 4000:4000\n - 3000:3000\n env_file: .env\n volumes:\n - .:/cube/conf\n\n query:\n image: cfmanteiga/alpine-bash-curl-jq\n depends_on:\n - cube\n volumes:\n - .:/query\n entrypoint: ["sh", "query/queries/run.sh"]\n | dataset_sample\yaml\cube-js_cube\examples\recipes\non-additivity\docker-compose.yml | docker-compose.yml | YAML | 322 | 0.7 | 0 | 0 | vue-tools | 54 | 2024-02-25T13:33:59.324749 | MIT | false | f98aec9b754e9c666ab8f6dc73d094bc |
version: '2.2'\n\nservices:\n cube:\n image: cubejs/cube:latest\n ports:\n - 4000:4000\n - 3000:3000\n env_file: .env\n volumes:\n - .:/cube/conf\n\n query:\n image: cfmanteiga/alpine-bash-curl-jq\n depends_on:\n - cube\n volumes:\n - .:/query\n entrypoint: ["sh", "query/queries/run.sh"]\n | dataset_sample\yaml\cube-js_cube\examples\recipes\pagination\docker-compose.yml | docker-compose.yml | YAML | 322 | 0.7 | 0 | 0 | react-lib | 187 | 2024-02-08T05:23:41.848677 | BSD-3-Clause | false | f98aec9b754e9c666ab8f6dc73d094bc |
version: '2.2'\n\nservices:\n cube:\n image: cubejs/cube:latest\n ports:\n - 4000:4000\n - 3000:3000\n env_file: .env\n volumes:\n - .:/cube/conf\n\n query:\n image: cfmanteiga/alpine-bash-curl-jq\n depends_on:\n - cube\n volumes:\n - .:/query\n entrypoint: ["sh", "query/queries/run.sh"]\n | dataset_sample\yaml\cube-js_cube\examples\recipes\passing-dynamic-parameters-in-query\docker-compose.yml | docker-compose.yml | YAML | 322 | 0.7 | 0 | 0 | vue-tools | 983 | 2024-12-26T07:41:45.805774 | BSD-3-Clause | false | f98aec9b754e9c666ab8f6dc73d094bc |
version: '2.2'\n\nservices:\n cube:\n image: cubejs/cube:latest\n ports:\n - 4000:4000\n - 3000:3000\n env_file: .env\n volumes:\n - .:/cube/conf\n\n query:\n image: cfmanteiga/alpine-bash-curl-jq\n depends_on:\n - cube\n volumes:\n - .:/query\n entrypoint: ["sh", "query/queries/run.sh"] | dataset_sample\yaml\cube-js_cube\examples\recipes\percentiles\docker-compose.yml | docker-compose.yml | YAML | 321 | 0.7 | 0 | 0 | react-lib | 832 | 2023-07-15T20:38:54.907544 | BSD-3-Clause | false | 8d5ed66009e6d6730fefc0ec64613385 |
version: '2.2'\n\nservices:\n cube:\n image: cubejs/cube:latest\n ports:\n - 4000:4000\n - 3000:3000\n env_file: .env\n volumes:\n - .:/cube/conf\n\n query:\n image: cfmanteiga/alpine-bash-curl-jq\n depends_on:\n - cube\n volumes:\n - .:/query\n entrypoint: ["sh", "query/queries/run.sh"]\n | dataset_sample\yaml\cube-js_cube\examples\recipes\referencing-dynamic-measures\docker-compose.yml | docker-compose.yml | YAML | 322 | 0.7 | 0 | 0 | react-lib | 781 | 2024-12-02T14:06:37.238401 | Apache-2.0 | false | f98aec9b754e9c666ab8f6dc73d094bc |
version: '2.2'\n\nservices:\n cubestore:\n image: cubejs/cubestore:arm64-experimental-v2021-07-29\n environment:\n - CUBESTORE_REMOTE_DIR=/cube/data\n volumes:\n - .cubestore:/cube/data\n\n cube:\n image: cubejs/cube:latest\n ports:\n - 4000:4000\n - 3000:3000\n env_file: .env\n volumes:\n - .:/cube/conf \n depends_on:\n - cubestore\n links:\n - cubestore\n\n postgres:\n image: postgres\n restart: always\n ports:\n - 5432:5432\n environment:\n POSTGRES_PASSWORD: example\n POSTGRES_DB: localDB\n volumes:\n - ./db-scripts:/docker-entrypoint-initdb.d\n\n query:\n image: cfmanteiga/alpine-bash-curl-jq\n depends_on:\n - cube\n volumes:\n - .:/query\n entrypoint: ["sh", "query/queries/run.sh"]\n\n node:\n build:\n context: .\n dockerfile: Dockerfile\n volumes:\n - .:/node\n entrypoint: ["sh", "node/data-updater/data-updater.sh"]\n | dataset_sample\yaml\cube-js_cube\examples\recipes\refreshing-select-partitions\docker-compose.yml | docker-compose.yml | YAML | 936 | 0.7 | 0 | 0 | vue-tools | 104 | 2024-01-11T20:06:55.515916 | Apache-2.0 | false | c4a60787515bcb25776a869e4d032b1b |
version: '2.2'\n\nservices:\n cube:\n image: cubejs/cube:latest\n ports:\n - 4000:4000\n - 3000:3000\n env_file: .env\n volumes:\n - .:/cube/conf\n\n query:\n image: cfmanteiga/alpine-bash-curl-jq\n depends_on:\n - cube\n volumes:\n - .:/query\n entrypoint: ["sh", "query/queries/run.sh"]\n | dataset_sample\yaml\cube-js_cube\examples\recipes\role-based-access\docker-compose.yml | docker-compose.yml | YAML | 322 | 0.7 | 0 | 0 | awesome-app | 414 | 2024-02-01T00:07:19.758064 | Apache-2.0 | false | f98aec9b754e9c666ab8f6dc73d094bc |
version: '2.2'\n\nservices:\n cube:\n image: cubejs/cube:latest\n ports:\n - 4000:4000\n - 3000:3000\n env_file: .env\n volumes:\n - .:/cube/conf\n\n query:\n image: cfmanteiga/alpine-bash-curl-jq\n depends_on:\n - cube\n volumes:\n - .:/query\n entrypoint: ["sh", "query/queries/run.sh"] | dataset_sample\yaml\cube-js_cube\examples\recipes\snapshots\docker-compose.yml | docker-compose.yml | YAML | 321 | 0.7 | 0 | 0 | awesome-app | 604 | 2025-04-11T02:09:37.934208 | GPL-3.0 | false | 8d5ed66009e6d6730fefc0ec64613385 |
version: '2.2'\n\nservices:\n cube:\n image: cubejs/cube:latest\n ports:\n - 4000:4000\n - 3000:3000\n env_file: .env\n volumes:\n - .:/cube/conf\n\n query:\n image: cfmanteiga/alpine-bash-curl-jq\n depends_on:\n - cube\n volumes:\n - .:/query\n entrypoint: ["sh", "query/queries/run.sh"]\n | dataset_sample\yaml\cube-js_cube\examples\recipes\using-different-schemas-for-tenants\docker-compose.yml | docker-compose.yml | YAML | 322 | 0.7 | 0 | 0 | node-utils | 194 | 2025-06-04T23:04:24.091863 | GPL-3.0 | false | f98aec9b754e9c666ab8f6dc73d094bc |
version: '2.2'\n\nservices:\n cube:\n image: cubejs/cube:latest\n ports:\n - 4000:4000\n - 3000:3000\n env_file: .env\n volumes:\n - .:/cube/conf | dataset_sample\yaml\cube-js_cube\examples\supabase\docker-compose.yml | docker-compose.yml | YAML | 163 | 0.7 | 0 | 0 | react-lib | 755 | 2025-06-12T18:00:00.796527 | MIT | false | 967a6cf3f6bc1593c2cf0a295a994639 |
version: '2.2'\nservices:\n cube:\n image: cubejs/cube:latest\n ports:\n - 4000:4000\n - 3306:3306\n env_file: .env\n volumes:\n - .:/cube/conf\n | dataset_sample\yaml\cube-js_cube\examples\superset\docker-compose.yml | docker-compose.yml | YAML | 163 | 0.7 | 0 | 0 | node-utils | 172 | 2025-06-29T13:31:23.896785 | Apache-2.0 | false | 44c753e8d6e566ae27ce38101f2c4309 |
version: '2.2'\n\nservices:\n cube:\n image: cubejs/cube:latest\n ports:\n - 4000:4000\n environment:\n - CUBEJS_DEV_MODE=true\n - CUBEJS_API_SECRET=SECRET\n env_file:\n - .env\n volumes:\n - .:/cube/conf | dataset_sample\yaml\cube-js_cube\examples\twitter-stream\docker-compose.yml | docker-compose.yml | YAML | 230 | 0.7 | 0 | 0 | react-lib | 721 | 2023-11-05T10:03:58.580488 | Apache-2.0 | false | dcb33fddad95872881750d2b34a0a2e1 |
openapi: "3.0.0"\ninfo:\n description: "Cube.js Swagger Schema"\n version: "1.0.0"\n title: "Cube.js"\npaths:\n "/v1/meta":\n get:\n summary: "Load Metadata"\n # TODO: Doesn't work in rust code-gen\n # parameters:\n # - in: query\n # name: extended\n # required: true\n # schema:\n # type: boolean\n # description: You will receive extended response if this parameter is true\n description: ""\n operationId: "metaV1"\n responses:\n "200":\n description: "successful operation"\n content:\n application/json:\n schema:\n $ref: "#/components/schemas/V1MetaResponse"\n "4XX":\n description: "Request could not be completed"\n content:\n application/json:\n schema:\n $ref: "#/components/schemas/V1Error"\n "5XX":\n description: "Internal Server Error"\n content:\n application/json:\n schema:\n $ref: "#/components/schemas/V1Error"\n "/v1/load":\n post:\n summary: "Load data via Cube JSON Query"\n operationId: "loadV1"\n requestBody:\n content:\n 'application/json':\n schema:\n $ref: '#/components/schemas/V1LoadRequest'\n responses:\n "200":\n description: "successful operation"\n content:\n application/json:\n schema:\n $ref: "#/components/schemas/V1LoadResponse"\n "4XX":\n description: "Request could not be completed"\n content:\n application/json:\n schema:\n $ref: "#/components/schemas/V1Error"\n "5XX":\n description: "Internal Server Error"\n content:\n application/json:\n schema:\n $ref: "#/components/schemas/V1Error"\nsecurity:\n - bearerAuth: []\ncomponents:\n securitySchemes:\n bearerAuth:\n type: http\n scheme: bearer\n bearerFormat: JWT\n schemas:\n V1CubeMetaJoin:\n type: "object"\n required:\n - name\n - relationship\n properties:\n name:\n type: "string"\n relationship:\n type: "string"\n V1CubeMetaSegment:\n type: "object"\n required:\n - name\n - title\n - shortTitle\n properties:\n name:\n type: "string"\n title:\n type: "string"\n description:\n type: "string"\n shortTitle:\n type: "string"\n meta:\n type: "object"\n V1CubeMetaDimensionGranularity:\n type: "object"\n required:\n - name\n - title\n - interval\n properties:\n name:\n type: "string"\n title:\n type: "string"\n interval:\n type: "string"\n offset:\n type: "string"\n origin:\n type: "string"\n V1CubeMetaDimension:\n type: "object"\n required:\n - name\n - type\n properties:\n name:\n type: "string"\n title:\n type: "string"\n shortTitle:\n type: "string"\n description:\n type: "string"\n type:\n type: "string"\n aliasMember:\n description: "When dimension is defined in View, it keeps the original path: Cube.dimension"\n type: "string"\n granularities:\n type: array\n items:\n $ref: "#/components/schemas/V1CubeMetaDimensionGranularity"\n meta:\n type: "object"\n V1CubeMetaMeasure:\n type: "object"\n required:\n - name\n - type\n properties:\n name:\n type: "string"\n title:\n type: "string"\n shortTitle:\n type: "string"\n description:\n type: "string"\n type:\n type: "string"\n aggType:\n type: "string"\n meta:\n type: "object"\n V1CubeMetaFolder:\n type: "object"\n required:\n - name\n - members\n properties:\n name:\n type: "string"\n members:\n type: array\n items:\n type: "string"\n V1CubeMetaHierarchy:\n type: "object"\n required:\n - name\n - levels\n properties:\n name:\n type: "string"\n title:\n type: "string"\n levels:\n type: "array"\n items:\n type: "string"\n V1CubeMeta:\n type: "object"\n required:\n - name\n - type\n - measures\n - dimensions\n - segments\n properties:\n name:\n type: "string"\n title:\n type: "string"\n type:\n $ref: "#/components/schemas/V1CubeMetaType"\n meta:\n type: "object"\n description:\n type: "string"\n measures:\n type: "array"\n items:\n $ref: "#/components/schemas/V1CubeMetaMeasure"\n dimensions:\n type: "array"\n items:\n $ref: "#/components/schemas/V1CubeMetaDimension"\n segments:\n type: "array"\n items:\n $ref: "#/components/schemas/V1CubeMetaSegment"\n joins:\n type: "array"\n items:\n $ref: "#/components/schemas/V1CubeMetaJoin"\n folders:\n type: "array"\n items:\n $ref: "#/components/schemas/V1CubeMetaFolder"\n hierarchies:\n type: "array"\n items:\n $ref: "#/components/schemas/V1CubeMetaHierarchy"\n V1CubeMetaType:\n type: "string"\n description: Type of cube\n enum:\n - "cube"\n - "view"\n V1MetaResponse:\n type: "object"\n properties:\n cubes:\n type: "array"\n items:\n $ref: "#/components/schemas/V1CubeMeta"\n compilerId:\n type: "string"\n V1LoadResultAnnotation:\n type: "object"\n required:\n - measures\n - dimensions\n - segments\n - timeDimensions\n properties:\n measures:\n type: "object"\n dimensions:\n type: "object"\n segments:\n type: "object"\n timeDimensions:\n type: "object"\n V1LoadResultData:\n type: "object"\n V1LoadResult:\n type: "object"\n required:\n - annotation\n - data\n properties:\n dataSource:\n type: "string"\n annotation:\n $ref: "#/components/schemas/V1LoadResultAnnotation"\n data:\n type: "array"\n items:\n $ref: "#/components/schemas/V1LoadResultData"\n refreshKeyValues:\n type: "array"\n items:\n type: "object"\n V1Error:\n type: "object"\n required:\n - "error"\n properties:\n error:\n type: "string"\n V1LoadResponse:\n type: "object"\n required:\n - results\n properties:\n pivotQuery:\n type: "object"\n slowQuery:\n type: "boolean"\n queryType:\n type: "string"\n results:\n type: "array"\n items:\n $ref: "#/components/schemas/V1LoadResult"\n V1LoadRequestQueryFilterLogicalAnd:\n type: "object"\n properties:\n and:\n type: "array"\n items:\n type: "object"\n V1LoadRequestQueryFilterLogicalOr:\n type: "object"\n properties:\n or:\n type: "array"\n items:\n type: "object"\n V1LoadRequestQueryFilterBase:\n type: "object"\n properties:\n member:\n type: "string"\n operator:\n type: "string"\n values:\n type: "array"\n items:\n type: "string"\n V1LoadRequestQueryTimeDimension:\n type: "object"\n required:\n - "dimension"\n properties:\n dimension:\n type: "string"\n granularity:\n type: "string"\n dateRange:\n type: "object"\n # Don't work in rust code-gen\n # oneOf:\n # - type: "string"\n # - type: "array"\n # items:\n # type: "string"\n V1LoadRequestQueryFilterItem:\n oneOf:\n - $ref: "#/components/schemas/V1LoadRequestQueryFilterBase"\n - $ref: "#/components/schemas/V1LoadRequestQueryFilterLogicalOr"\n - $ref: "#/components/schemas/V1LoadRequestQueryFilterLogicalAnd"\n V1LoadRequestQueryJoinSubquery:\n type: "object"\n properties:\n sql:\n type: "string"\n # TODO This is _always_ a member expression, maybe pass as parsed, without intermediate string?\n "on":\n type: "string"\n # TODO why string? it's enum\n joinType:\n type: "string"\n alias:\n type: "string"\n required:\n - sql\n - "on"\n - joinType\n - alias\n V1LoadRequestQuery:\n type: "object"\n properties:\n measures:\n type: "array"\n items:\n type: "string"\n dimensions:\n type: "array"\n items:\n type: "string"\n segments:\n type: "array"\n items:\n type: "string"\n timeDimensions:\n type: "array"\n items:\n $ref: "#/components/schemas/V1LoadRequestQueryTimeDimension"\n order:\n type: "array"\n items:\n type: "array"\n items:\n type: "string"\n limit:\n type: "integer"\n format: "int32"\n offset:\n type: "integer"\n format: "int32"\n filters:\n type: "array"\n items:\n $ref: "#/components/schemas/V1LoadRequestQueryFilterItem"\n ungrouped:\n type: "boolean"\n # vector of (subquery sql: string, join condition: member expression, join type: enum)\n # they will be added to end of joinQuery in BaseQuery, in same order as here\n subqueryJoins:\n type: "array"\n items:\n $ref: "#/components/schemas/V1LoadRequestQueryJoinSubquery"\n V1LoadRequest:\n type: "object"\n properties:\n queryType:\n type: "string"\n query:\n type: "object"\n $ref: "#/components/schemas/V1LoadRequestQuery"\n | dataset_sample\yaml\cube-js_cube\packages\cubejs-api-gateway\openspec.yml | openspec.yml | YAML | 10,332 | 0.95 | 0.00237 | 0.042654 | node-utils | 672 | 2024-02-18T04:07:24.008380 | GPL-3.0 | false | 29d0b1c546c0294fb583fe0c5b031058 |
{%- set unsafe_string = '"unsafe string" <>' -%}\n{%- set safe_string = new_safe_string() -%}\n{%- set dict_as_obj = new_object_from_dict() -%}\n\ntest:\n unsafe_string: "{{ unsafe_string }}"\n safe_string: "{{ safe_string }}"\n\ndump:\n dict_as_obj: {{ debug(dict_as_obj) }}\n | dataset_sample\yaml\cube-js_cube\packages\cubejs-backend-native\test\templates\python.yml | python.yml | YAML | 270 | 0.7 | 0 | 0 | node-utils | 65 | 2023-09-29T10:11:41.002017 | GPL-3.0 | true | 9d885100857d6846924fdfd64a2e5e0c |
version: "3"\n\nvolumes:\n metadata_data: {}\n middle_var: {}\n historical_var: {}\n broker_var: {}\n coordinator_var: {}\n router_var: {}\n\nservices:\n postgres:\n image: postgres:${TEST_POSTGRES_VERSION:-13}\n container_name: postgres\n volumes:\n - metadata_data:/var/lib/postgresql/data\n environment:\n - POSTGRES_PASSWORD=FoolishPassword\n - POSTGRES_USER=druid\n - POSTGRES_DB=druid\n restart: always\n healthcheck:\n test: ["CMD-SHELL", "pg_isready -U druid"]\n interval: 10s\n timeout: 5s\n retries: 5\n\n zookeeper:\n image: zookeeper:${TEST_ZOOKEEPER_VERSION:-3.5}\n container_name: zookeeper\n environment:\n - ZOO_MY_ID=1\n restart: always\n\n coordinator:\n image: apache/druid:${TEST_DRUID_VERSION:-27.0.0}\n container_name: coordinator\n volumes:\n - ./storage:/opt/data\n - coordinator_var:/opt/druid/var\n depends_on:\n - zookeeper\n - postgres\n ports:\n - "8081:8081"\n command:\n - coordinator\n env_file:\n - environment\n restart: always\n healthcheck:\n test: "wget --header 'Authorization: Basic YWRtaW46cGFzc3dvcmQx' --no-verbose --tries=1 --spider http://localhost:8081/status/health || exit 1"\n interval: 5s\n timeout: 5s\n retries: 5\n start_period: 60s\n\n broker:\n image: apache/druid:${TEST_DRUID_VERSION:-27.0.0}\n container_name: broker\n volumes:\n - broker_var:/opt/druid/var\n depends_on:\n - zookeeper\n - postgres\n - coordinator\n ports:\n - "8082:8082"\n command:\n - broker\n env_file:\n - environment\n restart: always\n healthcheck:\n test: "wget --header 'Authorization: Basic YWRtaW46cGFzc3dvcmQx' --no-verbose --tries=1 --spider http://localhost:8082/status/health || exit 1"\n interval: 5s\n timeout: 5s\n retries: 5\n start_period: 60s\n\n historical:\n image: apache/druid:${TEST_DRUID_VERSION:-27.0.0}\n container_name: historical\n volumes:\n - ./storage:/opt/data\n - historical_var:/opt/druid/var\n depends_on:\n - zookeeper\n - postgres\n - coordinator\n ports:\n - "8083:8083"\n command:\n - historical\n env_file:\n - environment\n restart: always\n healthcheck:\n test: "wget --header 'Authorization: Basic YWRtaW46cGFzc3dvcmQx' --no-verbose --tries=1 --spider http://localhost:8083/druid/historical/v1/readiness || exit 1"\n interval: 5s\n timeout: 5s\n retries: 5\n start_period: 60s\n\n middlemanager:\n image: apache/druid:${TEST_DRUID_VERSION:-27.0.0}\n container_name: middlemanager\n volumes:\n - ./storage:/opt/data\n - middle_var:/opt/druid/var\n depends_on:\n - zookeeper\n - postgres\n - coordinator\n ports:\n - "8091:8091"\n command:\n - middleManager\n env_file:\n - environment\n restart: always\n healthcheck:\n test: "wget --header 'Authorization: Basic YWRtaW46cGFzc3dvcmQx' --no-verbose --tries=1 --spider http://localhost:8091/status/health || exit 1"\n interval: 5s\n timeout: 5s\n retries: 5\n start_period: 60s\n\n router:\n image: apache/druid:${TEST_DRUID_VERSION:-27.0.0}\n container_name: router\n volumes:\n - router_var:/opt/druid/var\n depends_on:\n - zookeeper\n - postgres\n - coordinator\n ports:\n - "8888:8888"\n command:\n - router\n env_file:\n - environment\n restart: always\n healthcheck:\n test: "wget --header 'Authorization: Basic YWRtaW46cGFzc3dvcmQx' --no-verbose --tries=1 --spider http://localhost:8888/status/health || exit 1"\n interval: 5s\n timeout: 5s\n retries: 5\n start_period: 60s\n | dataset_sample\yaml\cube-js_cube\packages\cubejs-druid-driver\docker-compose.yml | docker-compose.yml | YAML | 3,682 | 0.8 | 0 | 0 | node-utils | 415 | 2025-01-19T14:32:31.526257 | GPL-3.0 | false | 058528bd46d6aef00078b1feb293edb9 |
version: '3'\nservices:\n es-distro:\n image: amazon/opendistro-for-elasticsearch:1.13.1\n environment:\n - discovery.type=single-node\n - bootstrap.memory_lock=true # along with the memlock settings below, disables swapping\n - ES_JAVA_OPTS=-Xms512m -Xmx512m\n ports:\n - 9200:9200\n healthcheck:\n test: curl -k -u admin:admin --silent --fail https://localhost:9200/_cluster/health || exit 1\n timeout: 5s\n interval: 2s\n retries: 3\n start_period: 10s\n | dataset_sample\yaml\cube-js_cube\packages\cubejs-elasticsearch-driver\docker-compose.yml | docker-compose.yml | YAML | 499 | 0.8 | 0.0625 | 0 | react-lib | 465 | 2025-05-22T09:32:20.759302 | BSD-3-Clause | false | 14fcf35d9fa81f19ec1e1463fdda011d |
version: '3'\n\nservices:\n mongo:\n image: mongo:${MONGO_TAG}\n ports:\n - 27017:27017\n volumes:\n - ./mongo-init.js:/docker-entrypoint-initdb.d/mongo-init.js:ro\n \n mongosqld:\n container_name: mongosqld\n build: \n context: ./\n args:\n MONGOBI: ${MONGOBI_VERSION}\n ports:\n - 3307:3307\n depends_on:\n - mongo\n | dataset_sample\yaml\cube-js_cube\packages\cubejs-mongobi-driver\test\docker-compose.yml | docker-compose.yml | YAML | 367 | 0.7 | 0 | 0 | vue-tools | 384 | 2024-03-18T20:32:01.135933 | Apache-2.0 | true | c68356f82bf752ac91b7a07c2c7e930e |
version: "3"\n\nservices:\n mysql:\n image: mysql:${TEST_MYSQL_VERSION:-5.6.50}\n container_name: mysql\n environment:\n MYSQL_ROOT_PASSWORD: Test1test\n ports:\n - 3306\n restart: always\n healthcheck:\n test: ["CMD", "mysqladmin" ,"ping", "-h", "localhost"]\n timeout: 20s\n interval: 15s\n retries: 5\n\n router:\n image: koxudaxi/local-data-api:${TEST_LOCAL_DATA_API_VERSION:-0.6.4}\n container_name: router\n environment:\n MYSQL_HOST: mysql\n MYSQL_PORT: 3306\n MYSQL_USER: root\n MYSQL_PASSWORD: Test1test\n SECRET_ARN: arn:aws:secretsmanager:us-east-1:123456789012:secret:dummy\n RESOURCE_ARN: arn:aws:rds:us-east-1:123456789012:cluster:dummy\n ports:\n - 80\n depends_on:\n - mysql\n links:\n - mysql\n restart: always\n | dataset_sample\yaml\cube-js_cube\packages\cubejs-mysql-aurora-serverless-driver\docker-compose.yml | docker-compose.yml | YAML | 811 | 0.7 | 0 | 0 | react-lib | 112 | 2025-03-06T15:14:12.247215 | BSD-3-Clause | false | 97ec39dda6f642279f11d618240bff15 |
services:\n pinot-controller:\n image: apachepinot/pinot:1.1.0\n hostname: pinot-controller\n restart: unless-stopped\n command: StartController -configFileName /tmp/data/test-resources/controller.conf\n container_name: pinot-controller-cube-tests\n ports:\n - 9000:9000\n depends_on:\n zookeeper:\n condition: service_healthy\n healthcheck:\n test: curl -f "http://localhost:9000/health"\n interval: 5s\n timeout: 5s\n retries: 3\n start_period: 10s\n volumes:\n - ./pinot-resources:/tmp/data/test-resources:ro\n\n pinot-broker:\n image: apachepinot/pinot:1.1.0\n hostname: pinot-broker\n restart: unless-stopped\n command: StartBroker -zkAddress zookeeper:2181 -configFileName /tmp/data/test-resources/broker.conf\n container_name: pinot-broker-cube-tests\n volumes:\n - ./pinot-resources:/tmp/data/test-resources:ro\n ports:\n - 8099:8099\n depends_on:\n pinot-controller:\n condition: service_healthy\n healthcheck:\n test: curl -f "http://localhost:8099/health"\n interval: 5s\n timeout: 5s\n retries: 3\n start_period: 10s\n\n pinot-server:\n image: apachepinot/pinot:1.1.0\n hostname: pinot-server\n restart: unless-stopped\n container_name: pinot-server-cube-tests\n command: StartServer -zkAddress zookeeper:2181\n ports:\n - 8098:8098\n - 8097:8097\n depends_on:\n pinot-broker:\n condition: service_healthy\n healthcheck:\n test: curl -f "http://localhost:8097/health"\n interval: 5s\n timeout: 5s\n retries: 3\n start_period: 10s\n\n ###\n # Utils (Zookeeper, Kafka, Spark, ...)\n ###\n zookeeper:\n image: zookeeper:latest\n hostname: zookeeper\n ports:\n - 2181:2181\n environment:\n ZOO_MY_ID: 1\n ZOO_PORT: 2181\n ZOO_SERVERS: server.1=zookeeper:2888:3888;2181\n healthcheck:\n test: nc -z localhost 2181 || exit -1\n interval: 10s\n timeout: 5s\n retries: 3\n start_period: 2s\n | dataset_sample\yaml\cube-js_cube\packages\cubejs-pinot-driver\docker-compose.yml | docker-compose.yml | YAML | 2,003 | 0.8 | 0 | 0.04 | react-lib | 449 | 2024-06-16T22:10:48.057003 | BSD-3-Clause | false | 789967c8c32d2da657c845575e12c5b4 |
executionFrameworkSpec:\n name: 'standalone'\n segmentGenerationJobRunnerClassName: 'org.apache.pinot.plugin.ingestion.batch.standalone.SegmentGenerationJobRunner'\n segmentTarPushJobRunnerClassName: 'org.apache.pinot.plugin.ingestion.batch.standalone.SegmentTarPushJobRunner'\n segmentUriPushJobRunnerClassName: 'org.apache.pinot.plugin.ingestion.batch.standalone.SegmentUriPushJobRunner'\njobType: SegmentCreationAndTarPush\ninputDirURI: '/tmp/data/test-resources/rawdata/scores/'\nincludeFileNamePattern: 'glob:**/*.csv'\noutputDirURI: '/tmp/data/'\noverwriteOutput: true\npinotFSSpecs:\n - scheme: file\n className: org.apache.pinot.spi.filesystem.LocalPinotFS\nrecordReaderSpec:\n dataFormat: 'csv'\n className: 'org.apache.pinot.plugin.inputformat.csv.CSVRecordReader'\n configClassName: 'org.apache.pinot.plugin.inputformat.csv.CSVRecordReaderConfig'\ntableSpec:\n tableName: 'scores'\npinotClusterSpecs:\n - controllerURI: 'http://localhost:9000'\npushJobSpec:\n pushAttempts: 1 | dataset_sample\yaml\cube-js_cube\packages\cubejs-pinot-driver\pinot-resources\scores.jobspec.yml | scores.jobspec.yml | YAML | 978 | 0.8 | 0 | 0 | awesome-app | 82 | 2024-12-22T15:43:05.277602 | GPL-3.0 | false | 541f0b9fa859b59766b194da8ae9b5b0 |
executionFrameworkSpec:\n name: 'standalone'\n segmentGenerationJobRunnerClassName: 'org.apache.pinot.plugin.ingestion.batch.standalone.SegmentGenerationJobRunner'\n segmentTarPushJobRunnerClassName: 'org.apache.pinot.plugin.ingestion.batch.standalone.SegmentTarPushJobRunner'\n segmentUriPushJobRunnerClassName: 'org.apache.pinot.plugin.ingestion.batch.standalone.SegmentUriPushJobRunner'\njobType: SegmentCreationAndTarPush\ninputDirURI: '/tmp/data/test-resources/rawdata/students/'\nincludeFileNamePattern: 'glob:**/*.csv'\noutputDirURI: '/tmp/data/'\noverwriteOutput: true\npinotFSSpecs:\n - scheme: file\n className: org.apache.pinot.spi.filesystem.LocalPinotFS\nrecordReaderSpec:\n dataFormat: 'csv'\n className: 'org.apache.pinot.plugin.inputformat.csv.CSVRecordReader'\n configClassName: 'org.apache.pinot.plugin.inputformat.csv.CSVRecordReaderConfig'\ntableSpec:\n tableName: 'students'\npinotClusterSpecs:\n - controllerURI: 'http://localhost:9000'\npushJobSpec:\n pushAttempts: 1 | dataset_sample\yaml\cube-js_cube\packages\cubejs-pinot-driver\pinot-resources\students.jobspec.yml | students.jobspec.yml | YAML | 982 | 0.8 | 0 | 0 | react-lib | 784 | 2024-02-24T09:41:10.486022 | BSD-3-Clause | false | d0a5298797f57158da24bc3b5e2df042 |
version: '2.2'\n\nservices:\n coordinator:\n image: trinodb/trino\n ports:\n - "8080:8080"\n container_name: "coordinator"\n healthcheck:\n test: "trino --execute 'SELECT 1' || exit 1"\n interval: 10s\n timeout: 5s\n retries: 5\n | dataset_sample\yaml\cube-js_cube\packages\cubejs-prestodb-driver\docker-compose.yml | docker-compose.yml | YAML | 254 | 0.7 | 0 | 0 | react-lib | 165 | 2023-08-23T15:34:55.511380 | BSD-3-Clause | false | 70643a4579c53bc4ae16f3cc4b454ef4 |
cubes:\n - name: base_orders\n sql_table: orders\n measures:\n - name: count\n sql: id\n type: count\n dimensions:\n - name: id\n sql: id\n type: number\n primary_key: true\n\n - name: number\n sql: number\n type: number\n\n - name: status\n sql: status\n type: string\n\n - name: city\n sql: city\n type: string\n hierarchies:\n - name: base_orders_hierarchy\n title: Hello Hierarchy\n levels:\n - "{CUBE}.status"\n - number\n #\n - name: orders\n extends: base_orders\n hierarchies:\n - name: orders_hierarchy\n levels:\n - state\n - city\n dimensions:\n - name: state\n sql: state\n type: string\n - name: city\n sql: city\n type: string\n\nviews:\n - name: test_view\n cubes:\n - join_path: orders\n includes: "*"\n\n\n\n | dataset_sample\yaml\cube-js_cube\packages\cubejs-schema-compiler\test\unit\fixtures\hierarchies-extended-cubes.yml | hierarchies-extended-cubes.yml | YAML | 918 | 0.8 | 0 | 0.021277 | node-utils | 584 | 2023-12-14T12:47:49.816871 | BSD-3-Clause | true | 8006604e4a6abe80e0afdf50cadbe788 |
cubes:\n - name: orders\n sql: SELECT * FROM orders\n joins:\n - name: users\n sql: "{CUBE}.order_id = {orders}.id"\n relationship: many_to_one\n measures:\n - name: count\n sql: id\n type: count\n dimensions:\n - name: id\n sql: id\n type: number\n primary_key: true\n\n - name: number\n sql: number\n type: number\n\n - name: status\n sql: status\n type: string\n\n - name: city\n sql: city\n type: string\n hierarchies:\n - name: orders_hierarchy\n title: Hello Hierarchy\n levels:\n - "{CUBE}.status"\n - number\n - users.city\n - name: some_other_hierarchy\n title: Some other hierarchy\n levels:\n - users.state\n - users.city\n #\n - name: users\n sql: SELECT * FROM users\n hierarchies:\n - name: users_hierarchy\n levels:\n - users.age\n - city\n dimensions:\n - name: age\n sql: age\n type: number\n - name: state\n sql: state\n type: string\n - name: city\n sql: city\n type: string\n\nviews:\n - name: orders_users_view\n cubes:\n - join_path: orders\n includes: "*"\n - join_path: users\n includes:\n - age\n - state\n - name: city\n alias: user_city\n - name: orders_includes_excludes_view\n cubes:\n - join_path: orders\n includes: "*"\n excludes:\n - some_other_hierarchy\n - name: empty_view\n cubes:\n - join_path: orders\n includes:\n - count\n - status\n - name: all_hierarchy_view\n cubes:\n - join_path: orders\n includes: "*"\n - join_path: users\n prefix: true\n includes: "*"\n - name: only_hierarchy_included_view\n cubes:\n - join_path: orders\n includes:\n - orders_hierarchy\n - join_path: users\n includes:\n - city\n - name: auto_include_view\n cubes:\n - join_path: orders\n includes:\n - orders_hierarchy\n - some_other_hierarchy\n | dataset_sample\yaml\cube-js_cube\packages\cubejs-schema-compiler\test\unit\fixtures\hierarchies.yml | hierarchies.yml | YAML | 2,126 | 0.8 | 0 | 0.010101 | vue-tools | 831 | 2024-06-22T10:58:05.032417 | BSD-3-Clause | true | 88126b2a9123e87c9ffb9a812bcb32ef |
cubes:\n - name: orders\n sql: SELECT * FROM orders\n measures:\n - name: count\n sql: id\n type: count\n dimensions:\n - name: id\n sql: id\n type: number\n primary_key: true\n\n - name: status\n sql: status\n type: string\n\n - name: city\n sql: city\n type: string\n hierarchies:\n - name: correct_hierarchy\n levels:\n - status\n - city\n - name: orders_hierarchy\n levels:\n - "{CUBE}.status"\n - count\n\nviews:\n - name: test_view\n cubes:\n - join_path: orders\n includes: "*"\n | dataset_sample\yaml\cube-js_cube\packages\cubejs-schema-compiler\test\unit\fixtures\hierarchy-with-measure.yml | hierarchy-with-measure.yml | YAML | 623 | 0.7 | 0 | 0 | vue-tools | 502 | 2024-07-15T04:58:13.724226 | GPL-3.0 | true | 53c5fe9a7847ef813fad73a7673cd04b |
cubes:\n - name: orders\n sql_table: orders\n\n dimensions:\n - name: id\n sql: id\n type: number\n primary_key: true\n\n - name: user_id\n sql: user_id\n type: number\n\n - name: status\n sql: status\n type: string\n\n - name: created_at\n sql: created_at\n type: time\n\n - name: completed_at\n sql: completed_at\n type: time\n\n measures:\n - name: count\n sql: id\n type: count\n\n joins:\n - name: order_users\n relationship: many_to_one\n sql: "${CUBE}.user_id = ${order_users.id}"\n\n segments:\n - name: sfUsers\n description: SF users segment from createCubeSchema\n sql: "${CUBE}.location = 'San Francisco'"\n\n hierarchies:\n - name: hello\n title: World\n levels: [status]\n\n pre_aggregations:\n - name: countCreatedAt\n type: rollup\n measures:\n - CUBE.count\n time_dimension: created_at\n granularity: day\n partition_granularity: month\n refresh_key:\n every: 1 hour\n scheduled_refresh: true\n\n accessPolicy:\n - role: common\n rowLevel:\n allowAll: true\n - role: admin\n conditions:\n - if: "{ !security_context.isBlocked }"\n rowLevel:\n filters:\n - member: status\n operator: equals\n values: ["completed"]\n - or:\n - member: "{CUBE}.created_at"\n operator: notInDateRange\n values:\n - 2022-01-01\n - "{ security_context.currentDate }"\n - member: "created_at"\n operator: equals\n values:\n - "{ securityContext.currentDate }"\n - and:\n - member: "{CUBE}.completed_at"\n operator: notInDateRange\n values:\n - 2022-01-01\n - "{ security_context.currentDate }"\n - member: "completed_at"\n operator: equals\n values:\n - "{ securityContext.currentDate }"\n memberLevel:\n includes:\n - status\n | dataset_sample\yaml\cube-js_cube\packages\cubejs-schema-compiler\test\unit\fixtures\orders_big.yml | orders_big.yml | YAML | 2,220 | 0.7 | 0.010753 | 0 | python-kit | 557 | 2025-06-10T11:11:08.213287 | Apache-2.0 | true | 16143a045dbf13085670455257d8afdb |
cubes:\n - name: ordersExt\n extends: orders\n\n dimensions:\n - name: city\n sql: city\n type: string\n\n measures:\n - name: count_distinct\n sql: status\n type: count_distinct\n\n segments:\n - name: anotherStatus\n description: Just another one\n sql: ${CUBE}.status = 'Rock and Roll'\n\n hierarchies:\n - name: ehlo\n title: UnderGround\n levels: [status, city]\n\n pre_aggregations:\n - name: mainPreAggs\n type: rollup\n measures: [count_distinct]\n dimensions: [city]\n\n accessPolicy:\n - role: manager\n memberLevel:\n excludes:\n - status\n | dataset_sample\yaml\cube-js_cube\packages\cubejs-schema-compiler\test\unit\fixtures\orders_ext.yml | orders_ext.yml | YAML | 672 | 0.7 | 0 | 0 | node-utils | 708 | 2024-03-25T01:08:49.546486 | GPL-3.0 | true | cdbc3df7ffe4fc6f8722cbad49394c48 |
cubes:\n - name: orders\n sql_table: orders\n\n dimensions:\n - name: id\n sql: id\n type: number\n primary_key: true\n\n - name: user_id\n sql: user_id\n type: number\n\n - name: status\n sql: status\n type: string\n\n - name: created_at\n sql: created_at\n type: time\n\n - name: completed_at\n sql: completed_at\n type: time\n\n measures:\n - name: count\n sql: id\n type: count\n\n joins:\n - name: order_users\n relationship: many_to_one\n sql: "${CUBE}.user_id = ${order_users.id}"\n\n segments:\n - name: sfUsers\n description: SF users segment from createCubeSchema\n sql: "${CUBE}.location = 'San Francisco'"\n\n hierarchies:\n - name: hello\n title: World\n levels: [status]\n\n pre_aggregations:\n - name: countCreatedAt\n type: rollup\n measures:\n - CUBE.count\n time_dimension: created_at\n granularity: day\n partition_granularity: month\n refresh_key:\n every: 1 hour\n scheduled_refresh: true\n\n accessPolicy:\n - role: common\n rowLevel:\n allowAll: true\n - role: admin\n conditions:\n - if: "{ !security_context.isBlocked }"\n rowLevel:\n filters:\n - member: "{CUBE}.order_users.name"\n operator: equals\n values: ["completed"]\n - or:\n - member: "{CUBE}.created_at"\n operator: notInDateRange\n values:\n - 2022-01-01\n - "{ security_context.currentDate }"\n - member: "created_at"\n operator: equals\n values:\n - "{ securityContext.currentDate }"\n memberLevel:\n includes:\n - status\n | dataset_sample\yaml\cube-js_cube\packages\cubejs-schema-compiler\test\unit\fixtures\orders_incorrect_acl.yml | orders_incorrect_acl.yml | YAML | 1,873 | 0.7 | 0.012048 | 0 | node-utils | 645 | 2025-01-21T09:31:23.994877 | GPL-3.0 | true | 2ab7b97ebe6e784b5a4d2d526be7fbd1 |
cubes:\n - name: orders\n sql_table: orders\n\n dimensions:\n - name: id\n sql: id\n type: number\n primary_key: true\n\n - name: user_id\n sql: user_id\n type: number\n\n - name: status\n sql: status\n type: string\n\n - name: created_at\n sql: created_at\n type: time\n\n - name: completed_at\n sql: completed_at\n type: time\n\n measures:\n - name: count\n sql: id\n type: count\n\n joins:\n - name: order_users\n relationship: many_to_one\n sql: "${CUBE}.user_id = ${order_users.id}"\n\n segments:\n - name: sfUsers\n description: SF users segment from createCubeSchema\n sql: "${CUBE}.location = 'San Francisco'"\n\n hierarchies:\n - name: hello\n title: World\n levels: [status]\n\n pre_aggregations:\n - name: countCreatedAt\n type: rollup\n measures:\n - CUBE.count\n time_dimension: created_at\n granularity: day\n partition_granularity: month\n refresh_key:\n every: 1 hour\n scheduled_refresh: true\n\n accessPolicy:\n - role: common\n rowLevel:\n allowAll: true\n - role: admin\n conditions:\n - if: "{ !security_context.isBlocked }"\n rowLevel:\n filters:\n - member: "{CUBE}.other.path.created_at"\n operator: equals\n values: ["completed"]\n - or:\n - member: "{CUBE}.created_at"\n operator: notInDateRange\n values:\n - 2022-01-01\n - "{ security_context.currentDate }"\n - member: "created_at"\n operator: equals\n values:\n - "{ securityContext.currentDate }"\n memberLevel:\n includes:\n - status\n | dataset_sample\yaml\cube-js_cube\packages\cubejs-schema-compiler\test\unit\fixtures\orders_nonexist_acl.yml | orders_nonexist_acl.yml | YAML | 1,878 | 0.7 | 0.012048 | 0 | awesome-app | 855 | 2024-01-03T01:25:59.227053 | Apache-2.0 | true | fb51ab21ec9d917f8a8acd5f8b9569e2 |
cubes:\n - name: order_users\n sql: SELECT * FROM order_users;\n\n measures:\n - name: count\n sql: id\n type: count\n dimensions:\n - name: id\n sql: id\n type: number\n primary_key: true\n\n - name: name\n sql: name\n type: string\n | dataset_sample\yaml\cube-js_cube\packages\cubejs-schema-compiler\test\unit\fixtures\order_users.yml | order_users.yml | YAML | 291 | 0.7 | 0 | 0 | awesome-app | 443 | 2024-07-23T00:28:57.999090 | BSD-3-Clause | true | 39137845579ebb901c89edef4c25cc49 |
cubes:\n - name: validate_preaggs\n sql: SELECT * FROM order_users;\n\n measures:\n - name: count\n sql: id\n type: count\n dimensions:\n - name: id\n sql: id\n type: number\n primary_key: true\n\n - name: name\n sql: name\n type: string\n\n - name: created_at\n sql: created_at\n type: time\n\n pre_aggregations:\n - name: autoRollupFail\n type: autoRollup\n maxPreAggregations: string_instead_of_number\n\n - name: originalSqlFail\n type: originalSql\n partition_granularity: invalid_partition_granularity\n\n - name: originalSqlFail2\n type: originalSql\n partition_granularity: month\n uniqueKeyColumns: not_an_array\n\n - name: rollupJoinFail\n type: rollupJoin\n partition_granularity: month\n # no rollups\n\n - name: rollupLambdaFail\n type: rollupLambda\n partition_granularity: month\n granularity: day\n time_dimension: created_at\n rollups: not_a_func\n\n - name: rollupFail\n type: rollup\n measures:\n - CUBE.count\n time_dimension: # as array\n - CUBE.created_at\n granularity: day\n partition_granularity: month\n refresh_key:\n every: 1 hour\n scheduled_refresh: true\n\n - name: rollupFail2\n type: rollup\n measures:\n - CUBE.count\n time_dimensions: CUBE.created_at # not an array\n | dataset_sample\yaml\cube-js_cube\packages\cubejs-schema-compiler\test\unit\fixtures\validate_preaggs.yml | validate_preaggs.yml | YAML | 1,477 | 0.8 | 0 | 0.018182 | react-lib | 455 | 2024-11-20T12:54:36.566077 | MIT | true | a52a81480f607f009841b2a3c6a47b1c |
version: "2.2"\n\nservices:\n cube:\n container_name: birdbox-cube\n image: ${BIRDBOX_CUBEJS_REGISTRY_PATH}cubejs/cube:${BIRDBOX_CUBEJS_VERSION:-latest}\n environment:\n CUBEJS_DB_TYPE: athena\n\n # from env\n CUBEJS_AWS_KEY: ${CUBEJS_AWS_KEY}\n CUBEJS_AWS_SECRET: ${CUBEJS_AWS_SECRET}\n CUBEJS_AWS_REGION: ${CUBEJS_AWS_REGION}\n CUBEJS_AWS_S3_OUTPUT_LOCATION: ${CUBEJS_AWS_S3_OUTPUT_LOCATION}\n CUBEJS_DB_EXPORT_BUCKET: ${CUBEJS_DB_EXPORT_BUCKET}\n\n # from test\n CUBEJS_SCHEDULED_REFRESH_DEFAULT: ${CUBEJS_SCHEDULED_REFRESH_DEFAULT}\n CUBEJS_REFRESH_WORKER: ${CUBEJS_REFRESH_WORKER}\n CUBEJS_EXTERNAL_DEFAULT: ${CUBEJS_EXTERNAL_DEFAULT}\n CUBEJS_ROLLUP_ONLY: ${CUBEJS_ROLLUP_ONLY}\n\n # defaults\n CUBEJS_DEV_MODE: "true"\n CUBEJS_API_SECRET: mysupersecret\n CUBEJS_SCHEMA_PATH: schema\n volumes:\n - ./postgresql/single/cube.js:/cube/conf/cube.js\n - ./postgresql/schema/Orders.js:/cube/conf/schema/Orders.js\n - ./postgresql/schema/OrdersPA.js:/cube/conf/schema/OrdersPA.js\n - ./postgresql/schema/CAST.js:/cube/conf/schema/CAST.js\n - ./postgresql/schema/Customers.js:/cube/conf/schema/Customers.js\n - ./postgresql/schema/Customers.sql.js:/cube/conf/schema/Customers.sql.js\n - ./postgresql/schema/ECommerce.js:/cube/conf/schema/ECommerce.js\n - ./postgresql/schema/ECommerce.sql.js:/cube/conf/schema/ECommerce.sql.js\n - ./postgresql/schema/Products.js:/cube/conf/schema/Products.js\n - ./postgresql/schema/Products.sql.js:/cube/conf/schema/Products.sql.js\n ports:\n - "4000:4000"\n restart: always\n | dataset_sample\yaml\cube-js_cube\packages\cubejs-testing\birdbox-fixtures\athena.yml | athena.yml | YAML | 1,623 | 0.8 | 0 | 0.083333 | node-utils | 421 | 2025-04-25T10:39:51.016612 | MIT | true | ed37434aed0d4a6258e366489659afdf |
version: "2.2"\n\nservices:\n cube:\n container_name: birdbox-cube\n image: ${BIRDBOX_CUBEJS_REGISTRY_PATH}cubejs/cube:${BIRDBOX_CUBEJS_VERSION:-latest}\n environment:\n CUBEJS_DB_TYPE: bigquery\n\n # from env\n CUBEJS_DB_BQ_PROJECT_ID: ${CUBEJS_DB_BQ_PROJECT_ID}\n CUBEJS_DB_BQ_CREDENTIALS: ${CUBEJS_DB_BQ_CREDENTIALS}\n CUBEJS_DB_EXPORT_BUCKET: ${CUBEJS_DB_EXPORT_BUCKET}\n\n # from test\n CUBEJS_SCHEDULED_REFRESH_DEFAULT: ${CUBEJS_SCHEDULED_REFRESH_DEFAULT}\n CUBEJS_REFRESH_WORKER: ${CUBEJS_REFRESH_WORKER}\n CUBEJS_EXTERNAL_DEFAULT: ${CUBEJS_EXTERNAL_DEFAULT}\n CUBEJS_ROLLUP_ONLY: ${CUBEJS_ROLLUP_ONLY}\n\n # defaults\n CUBEJS_DEV_MODE: "true"\n CUBEJS_API_SECRET: mysupersecret\n CUBEJS_SCHEMA_PATH: schema\n volumes:\n - ./postgresql/single/cube.js:/cube/conf/cube.js\n - ./postgresql/schema/Orders.js:/cube/conf/schema/Orders.js\n - ./postgresql/schema/OrdersPA.js:/cube/conf/schema/OrdersPA.js\n - ./postgresql/schema/CAST.js:/cube/conf/schema/CAST.js\n - ./postgresql/schema/Customers.js:/cube/conf/schema/Customers.js\n - ./postgresql/schema/Customers.sql.js:/cube/conf/schema/Customers.sql.js\n - ./postgresql/schema/ECommerce.js:/cube/conf/schema/ECommerce.js\n - ./postgresql/schema/ECommerce.sql.js:/cube/conf/schema/ECommerce.sql.js\n - ./postgresql/schema/Products.js:/cube/conf/schema/Products.js\n - ./postgresql/schema/Products.sql.js:/cube/conf/schema/Products.sql.js\n ports:\n - "4000:4000"\n restart: always\n | dataset_sample\yaml\cube-js_cube\packages\cubejs-testing\birdbox-fixtures\bigquery.yml | bigquery.yml | YAML | 1,541 | 0.8 | 0 | 0.088235 | python-kit | 710 | 2023-11-01T17:06:31.473909 | Apache-2.0 | true | 8242580554054826c635a8be0cd49b57 |
version: "2.2"\n\nservices:\n cube:\n container_name: birdbox-cube\n image: ${BIRDBOX_CUBEJS_REGISTRY_PATH}cubejs/cube:${BIRDBOX_CUBEJS_VERSION:-latest}\n environment:\n CUBEJS_DB_TYPE: databricks-jdbc\n\n # from test\n CUBEJS_SCHEDULED_REFRESH_DEFAULT: ${CUBEJS_SCHEDULED_REFRESH_DEFAULT}\n CUBEJS_REFRESH_WORKER: ${CUBEJS_REFRESH_WORKER}\n CUBEJS_EXTERNAL_DEFAULT: ${CUBEJS_EXTERNAL_DEFAULT}\n CUBEJS_ROLLUP_ONLY: ${CUBEJS_ROLLUP_ONLY}\n CUBEJS_TELEMETRY: ${CUBEJS_TELEMETRY}\n\n # defaults\n CUBEJS_DEV_MODE: "true"\n CUBEJS_API_SECRET: mysupersecret\n CUBEJS_SCHEMA_PATH: schema\n volumes:\n - ./databricks-jdbc/cube.js:/cube/conf/cube.js\n ports:\n - "4000:4000"\n restart: always\n | dataset_sample\yaml\cube-js_cube\packages\cubejs-testing\birdbox-fixtures\databricks-jdbc.yml | databricks-jdbc.yml | YAML | 746 | 0.8 | 0 | 0.090909 | node-utils | 960 | 2025-03-21T02:29:12.076259 | GPL-3.0 | true | b43eb5bc92fb2f0159954a809613fda4 |
version: "2.2"\n\nservices:\n cube:\n container_name: birdbox-cube\n image: ${BIRDBOX_CUBEJS_REGISTRY_PATH}cubejs/cube:${BIRDBOX_CUBEJS_VERSION:-latest}\n environment:\n CUBEJS_DB_TYPE: materialize\n\n CUBEJS_DB_HOST: host.docker.internal\n CUBEJS_DB_PORT: ${CUBEJS_DB_PORT:-6875}\n CUBEJS_DB_NAME: ${CUBEJS_DB_NAME:-materialize}\n CUBEJS_DB_USER: ${CUBEJS_DB_USER:-materialize}\n CUBEJS_DB_PASS: ${CUBEJS_DB_PASS:-materialize}\n\n CUBEJS_DEV_MODE: "true"\n CUBEJS_WEB_SOCKETS: "true"\n CUBEJS_API_SECRET: mysupersecret\n CUBEJS_SCHEMA_PATH: schema\n volumes:\n - ./questdb/schema:/cube/conf/schema\n ports:\n - "4000"\n restart: always\n\n # Database instances, for troubleshooting.\n # NOT required.\n # Tests provide db instances via testcontainers, in both LOCAL and DOCKER modes.\n\n db:\n container_name: birdbox-db\n image: questdb/questdb:6.2.1\n ports:\n - "8812"\n - "9000"\n - "9003"\n restart: always\n volumes:\n - ./datasets:/data:ro\n - ./questdb/scripts:/scripts:ro\n healthcheck:\n test: ["CMD-SHELL", "curl -f http://127.0.0.1:9003 || exit 1"]\n interval: 10s\n timeout: 5s\n retries: 5\n | dataset_sample\yaml\cube-js_cube\packages\cubejs-testing\birdbox-fixtures\materialize.yml | materialize.yml | YAML | 1,207 | 0.95 | 0.022222 | 0.075 | react-lib | 224 | 2025-04-14T22:00:47.136421 | GPL-3.0 | true | 1d8bb3504f60a29f7d8faf2362aa488b |
version: "2.2"\n\nservices:\n cube:\n container_name: birdbox-cube\n image: ${BIRDBOX_CUBEJS_REGISTRY_PATH}cubejs/cube:${BIRDBOX_CUBEJS_VERSION:-latest}\n environment:\n CUBEJS_DB_TYPE: postgres\n\n CUBEJS_DB_HOST: host.docker.internal\n CUBEJS_DB_PORT: ${CUBEJS_DB_PORT:-5432}\n CUBEJS_DB_NAME: ${CUBEJS_DB_NAME:-test}\n CUBEJS_DB_USER: ${CUBEJS_DB_USER:-test}\n CUBEJS_DB_PASS: ${CUBEJS_DB_PASS:-test}\n\n CUBEJS_DB_HOST2: host.docker.internal\n CUBEJS_DB_PORT2: ${CUBEJS_DB_PORT2:-3306}\n CUBEJS_DB_NAME2: ${CUBEJS_DB_NAME2:-mysql}\n CUBEJS_DB_USER2: ${CUBEJS_DB_USER2:-root}\n CUBEJS_DB_PASS2: ${CUBEJS_DB_PASS2:-Test1test}\n\n CUBEJS_DEV_MODE: "true"\n CUBEJS_WEB_SOCKETS: "true"\n CUBEJS_API_SECRET: mysupersecret\n CUBEJS_SCHEMA_PATH: schema\n volumes:\n - ./multidb/cube.js:/cube/conf/cube.js\n - ./multidb/schema:/cube/conf/schema\n ports:\n - "4000:4000"\n restart: always\n\n # Database instances, for troubleshooting.\n # NOT required.\n # Tests provide db instances via testcontainers, in both LOCAL and DOCKER modes.\n\n db:\n container_name: birdbox-db\n image: postgres:13\n environment:\n - POSTGRES_PASSWORD=test\n - POSTGRES_USER=test\n - POSTGRES_DB=test\n ports:\n - "5432"\n restart: always\n healthcheck:\n test: ["CMD-SHELL", "pg_isready -U test"]\n interval: 10s\n timeout: 5s\n retries: 5\n\n db2:\n container_name: birdbox-db2\n image: mysql:8.0\n command: --default-authentication-plugin=mysql_native_password\n environment:\n - MYSQL_ROOT_PASSWORD=Test1test\n ports:\n - "3306:3306"\n restart: always\n healthcheck:\n test: [ "CMD", "mysqladmin" ,"ping", "-h", "localhost" ]\n interval: 10s\n timeout: 5s\n retries: 5 | dataset_sample\yaml\cube-js_cube\packages\cubejs-testing\birdbox-fixtures\multidb.yml | multidb.yml | YAML | 1,804 | 0.95 | 0.015385 | 0.050847 | python-kit | 868 | 2023-11-22T01:28:41.127201 | MIT | true | 53d73b7346d897873e57788f364b2151 |
version: "2.2"\n\nservices:\n cube:\n container_name: birdbox-cube\n image: ${BIRDBOX_CUBEJS_REGISTRY_PATH}cubejs/cube:${BIRDBOX_CUBEJS_VERSION:-latest}\n environment:\n CUBEJS_DB_TYPE: postgres\n CUBEJS_DB_HOST: db\n CUBEJS_DB_NAME: test\n CUBEJS_DB_USER: test\n CUBEJS_DB_PASS: test\n CUBEJS_DEV_MODE: ${CUBEJS_DEV_MODE:-true}\n CUBEJS_WEB_SOCKETS: "true"\n CUBEJS_API_SECRET: mysupersecret\n CUBEJS_CUBESTORE_HOST: cubestore\n CUBEJS_PLAYGROUND_AUTH_SECRET: mysupersecret\n CUBEJS_CACHE_AND_QUEUE_DRIVER: memory\n CUBEJS_SCHEMA_PATH: schema\n depends_on:\n - db\n - cubestore\n links:\n - db\n - cubestore\n volumes:\n - ./postgresql:/cube/conf\n - ./postgresql/cubestore/cube.js:/cube/conf/cube.js\n ports:\n - "4000"\n restart: always\n\n db:\n container_name: birdbox-db\n image: hbontempo/postgres-hll:16-v2.18\n environment:\n - POSTGRES_PASSWORD=test\n - POSTGRES_USER=test\n - POSTGRES_DB=test\n restart: always\n ports:\n - "5432"\n volumes:\n - ./datasets:/data:ro\n - ./postgresql/scripts:/scripts:ro\n healthcheck:\n test: ["CMD-SHELL", "pg_isready -U test"]\n interval: 10s\n timeout: 5s\n retries: 5\n\n cubestore:\n container_name: birdbox-cubestore\n image: cubejs/cubestore:${BIRDBOX_CUBESTORE_VERSION:-latest}\n# Possible workaround for next version of testcontainers\n# environment:\n# # Workaround for Error during processing MySQL connection: peer terminated connection\n# # Because testcontainers know about exposed 3306 from Dockerfile & EXPOSE\n# CUBESTORE_PORT: "13306"\n | dataset_sample\yaml\cube-js_cube\packages\cubejs-testing\birdbox-fixtures\postgresql-cubestore.yml | postgresql-cubestore.yml | YAML | 1,656 | 0.8 | 0.033898 | 0.089286 | node-utils | 560 | 2024-06-25T12:34:09.866349 | Apache-2.0 | true | 688afcfd87022e76cf675de1a647e0b1 |
version: "2.2"\n\nservices:\n cube:\n container_name: birdbox-cube\n image: ${BIRDBOX_CUBEJS_REGISTRY_PATH}cubejs/cube:${BIRDBOX_CUBEJS_VERSION:-latest}\n environment:\n CUBEJS_DB_TYPE: postgres\n CUBEJS_DB_HOST: db\n CUBEJS_DB_NAME: test\n CUBEJS_DB_USER: test\n CUBEJS_DB_PASS: test\n CUBEJS_DEV_MODE: ${CUBEJS_DEV_MODE:-true}\n CUBEJS_WEB_SOCKETS: "true"\n CUBEJS_API_SECRET: mysupersecret\n CUBEJS_CACHE_AND_QUEUE_DRIVER: memory\n CUBEJS_SCHEMA_PATH: schema\n depends_on:\n - db\n links:\n - db\n volumes:\n - ./postgresql:/cube/conf\n - ./postgresql/single/cube.js:/cube/conf/cube.js\n - ./postgresql/schema/CAST.js:/cube/conf/schema/CAST.js\n - ./postgresql/schema/Customers.js:/cube/conf/schema/Customers.js\n - ./postgresql/schema/Customers.sql.js:/cube/conf/schema/Customers.sql.js\n - ./postgresql/schema/ECommerce.js:/cube/conf/schema/ECommerce.js\n - ./postgresql/schema/ECommerce.sql.js:/cube/conf/schema/ECommerce.sql.js\n - ./postgresql/schema/Products.js:/cube/conf/schema/Products.js\n - ./postgresql/schema/Products.sql.js:/cube/conf/schema/Products.sql.js\n ports:\n - "4000"\n restart: always\n\n db:\n container_name: birdbox-db\n image: hbontempo/postgres-hll:16-v2.18\n environment:\n - POSTGRES_PASSWORD=test\n - POSTGRES_USER=test\n - POSTGRES_DB=test\n ports:\n - "5432"\n restart: always\n volumes:\n - ./datasets:/data:ro\n - ./postgresql/scripts:/scripts:ro\n healthcheck:\n test: ["CMD-SHELL", "pg_isready -U test"]\n interval: 10s\n timeout: 5s\n retries: 5\n | dataset_sample\yaml\cube-js_cube\packages\cubejs-testing\birdbox-fixtures\postgresql.yml | postgresql.yml | YAML | 1,644 | 0.7 | 0 | 0 | awesome-app | 779 | 2023-09-20T14:20:59.559698 | Apache-2.0 | true | 40e2377802bdf03acd6af7d7a88c44ee |
version: "2.2"\n\nservices:\n cube:\n container_name: birdbox-cube\n image: ${BIRDBOX_CUBEJS_REGISTRY_PATH}cubejs/cube:${BIRDBOX_CUBEJS_VERSION:-latest}\n environment:\n CUBEJS_DB_TYPE: prestodb\n\n CUBEJS_DB_HOST: host.docker.internal\n CUBEJS_DB_PORT: ${CUBEJS_DB_PORT:-8080}\n CUBEJS_DB_PRESTO_CATALOG: ${CUBEJS_DB_NAME:-memory}\n CUBEJS_DB_USER: ${CUBEJS_DB_USER:-presto}\n\n CUBEJS_DEV_MODE: "true"\n CUBEJS_WEB_SOCKETS: "true"\n CUBEJS_API_SECRET: mysupersecret\n CUBEJS_SCHEMA_PATH: schema\n volumes:\n - ./prestodb/schema:/cube/conf/schema\n ports:\n - "4000"\n restart: always\n | dataset_sample\yaml\cube-js_cube\packages\cubejs-testing\birdbox-fixtures\prestodb.yml | prestodb.yml | YAML | 638 | 0.7 | 0 | 0 | awesome-app | 355 | 2024-04-20T15:39:37.037921 | MIT | true | 8ab8bb0918a9a09203c3d2f2a006885c |
version: "2.2"\n\nservices:\n cube:\n container_name: birdbox-cube\n image: ${BIRDBOX_CUBEJS_REGISTRY_PATH}cubejs/cube:${BIRDBOX_CUBEJS_VERSION:-latest}\n environment:\n CUBEJS_DB_TYPE: questdb\n\n CUBEJS_DB_HOST: host.docker.internal\n CUBEJS_DB_PORT: ${CUBEJS_DB_PORT:-8812}\n CUBEJS_DB_NAME: ${CUBEJS_DB_NAME:-qdb}\n CUBEJS_DB_USER: ${CUBEJS_DB_USER:-admin}\n CUBEJS_DB_PASS: ${CUBEJS_DB_PASS:-quest}\n\n CUBEJS_DEV_MODE: "true"\n CUBEJS_WEB_SOCKETS: "true"\n CUBEJS_API_SECRET: mysupersecret\n CUBEJS_SCHEMA_PATH: schema\n volumes:\n - ./questdb/schema:/cube/conf/schema\n ports:\n - "4000"\n restart: always\n\n # Database instances, for troubleshooting.\n # NOT required.\n # Tests provide db instances via testcontainers, in both LOCAL and DOCKER modes.\n\n db:\n container_name: birdbox-db\n image: questdb/questdb:6.2.1\n ports:\n - "8812"\n - "9000"\n - "9003"\n restart: always\n volumes:\n - ./datasets:/data:ro\n - ./questdb/scripts:/scripts:ro\n healthcheck:\n test: ["CMD-SHELL", "curl -f http://127.0.0.1:9003 || exit 1"]\n interval: 10s\n timeout: 5s\n retries: 5\n | dataset_sample\yaml\cube-js_cube\packages\cubejs-testing\birdbox-fixtures\questdb.yml | questdb.yml | YAML | 1,183 | 0.95 | 0.022222 | 0.075 | python-kit | 752 | 2024-06-25T14:56:22.330940 | MIT | true | de2a9d74400e14091258e40116d4a9cf |
version: "2.2"\n\nservices:\n cube:\n container_name: birdbox-cube\n image: ${BIRDBOX_CUBEJS_REGISTRY_PATH}cubejs/cube:${BIRDBOX_CUBEJS_VERSION:-latest}\n environment:\n CUBEJS_DB_TYPE: vertica\n\n CUBEJS_DB_HOST: db\n CUBEJS_DB_PORT: 5433\n CUBEJS_DB_NAME: vmart\n CUBEJS_DB_USER: dbadmin\n CUBEJS_DB_PASS:\n\n CUBEJS_DEV_MODE: "true"\n CUBEJS_WEB_SOCKETS: "true"\n CUBEJS_API_SECRET: mysupersecret\n volumes:\n - ./vertica/schema:/cube/conf/schema\n ports:\n - "4000"\n depends_on:\n db:\n condition: service_healthy\n links:\n - db\n restart: always\n\n # Database instances, for troubleshooting.\n # NOT required.\n # Tests provide db instances via testcontainers, in both LOCAL and DOCKER modes.\n\n db:\n container_name: birdbox-db\n image: vertica/vertica-ce:12.0.4-0\n ports:\n - "5433"\n restart: always\n healthcheck:\n test: ["CMD-SHELL", "/opt/vertica/bin/vsql -c 'select 1;' || exit 1"]\n interval: 10s\n timeout: 5s\n retries: 5\n | dataset_sample\yaml\cube-js_cube\packages\cubejs-testing\birdbox-fixtures\vertica.yml | vertica.yml | YAML | 1,041 | 0.95 | 0.022727 | 0.076923 | python-kit | 14 | 2024-10-26T22:28:05.470706 | Apache-2.0 | true | 40cedac9aa99ee9638c0d36747355f20 |
\n# Name your project! Project names should contain only lowercase characters\n# and underscores. A good package name should reflect your organization's\n# name or the intended use of these models\nname: 'my_new_project'\nversion: '1.0.0'\nconfig-version: 2\n\n# This setting configures which "profile" dbt uses for this project.\nprofile: 'default'\n\n# These configurations specify where dbt should look for different types of files.\n# The `source-paths` config, for example, states that models in this project can be\n# found in the "models/" directory. You probably won't need to change these!\nsource-paths: ["models"]\nanalysis-paths: ["analysis"]\ntest-paths: ["tests"]\ndata-paths: ["data"]\nmacro-paths: ["macros"]\nsnapshot-paths: ["snapshots"]\n\ntarget-path: "target" # directory which will store compiled SQL files\nclean-targets: # directories to be removed by `dbt clean`\n - "target"\n - "dbt_modules"\n\n\n# Configuring models\n# Full documentation: https://docs.getdbt.com/docs/configuring-models\n\n# In this example config, we tell dbt to build all models in the example/ directory\n# as tables. These settings can be overridden in the individual model files\n# using the `{{ config(...) }}` macro.\nmodels:\n my_new_project:\n # Config indicated by + and applies to all files under models/example/\n example:\n +materialized: view\n | dataset_sample\yaml\cube-js_cube\packages\cubejs-testing\birdbox-fixtures\postgresql\dbt-project\dbt_project.yml | dbt_project.yml | YAML | 1,339 | 0.8 | 0.078947 | 0.419355 | react-lib | 56 | 2025-06-12T21:50:04.275673 | GPL-3.0 | true | f81c1a43db734c717b224d2f734467d0 |
version: "2.2"\n\nservices:\n cube:\n image: cubejs/cube:latest\n ports:\n - 4000:4000\n - 15432:15432\n environment:\n - CUBEJS_DB_TYPE=vertica\n - CUBEJS_DB_HOST= #host\n - CUBEJS_DB_NAME= #database name \n - CUBEJS_DB_PORT=5433\n - CUBEJS_DB_USER= #database user\n - CUBEJS_DB_PASS= #database password\n - CUBEJS_DEV_MODE=true #if running locally\n volumes:\n - .:/cube/conf\n\n | dataset_sample\yaml\cube-js_cube\packages\cubejs-vertica-driver\docker-compose.yml | docker-compose.yml | YAML | 425 | 0.8 | 0.052632 | 0 | awesome-app | 598 | 2025-06-10T09:23:37.630700 | GPL-3.0 | false | c51dc7dc81fd8dd6aac7dd14e24ef44c |
language: php\nphp:\n - 5.4\n - 5.5\n - 5.6\n - 7.0\n - 7.1\n - hhvm\n\ninstall:\n - travis_retry composer install --no-interaction\n - composer info -i\n\nscript: vendor/bin/phpunit --bootstrap tests/bootstrap.php --configuration tests/phpunit.xml tests\n | dataset_sample\yaml\cytopia_devilbox\.devilbox\www\htdocs\vendor\phpmyadmin-4.9.7\vendor\bacon\bacon-qr-code\.travis.yml | .travis.yml | YAML | 250 | 0.7 | 0 | 0 | node-utils | 686 | 2025-02-27T07:34:47.005576 | Apache-2.0 | false | ba3ba9106c3263ffa94ccdc7077f0c44 |
dist: trusty\n\nlanguage: php\n\nsudo: false\n\nphp:\n - '5.5'\n - '5.6'\n - '7.0'\n - '7.1'\n - '7.2'\n - '7.3'\n\nbefore_script:\n - composer install\n - phpenv version-name | grep ^5.[34] && echo "extension=apc.so" >> ~/.phpenv/versions/$(phpenv version-name)/etc/php.ini ; true\n - phpenv version-name | grep ^5.[34] && echo "apc.enable_cli=1" >> ~/.phpenv/versions/$(phpenv version-name)/etc/php.ini ; true\n\nscript:\n - mkdir -p build/logs\n - composer run-script lint\n - composer run-script test\n\nafter_success:\n - travis_retry php vendor/bin/php-coveralls\n\ncache:\n directories:\n - "$HOME/.composer/cache/files"\n\ngit:\n depth: 5\n | dataset_sample\yaml\cytopia_devilbox\.devilbox\www\htdocs\vendor\phpmyadmin-4.9.7\vendor\google\recaptcha\.travis.yml | .travis.yml | YAML | 633 | 0.7 | 0 | 0 | awesome-app | 151 | 2024-08-16T07:13:21.996145 | BSD-3-Clause | false | 64a8d9bc341eb662a27a10a25012444f |
language: php\nsudo: false\n\nphp:\n - "7.2"\n - "7.1"\n - "7.0"\n - "5.6"\n - "hhvm"\n\nmatrix:\n fast_finish: true\n\ninstall:\n - composer self-update\n - composer update\n\nscript:\n - vendor/bin/phpunit\n - vendor/bin/psalm\n | dataset_sample\yaml\cytopia_devilbox\.devilbox\www\htdocs\vendor\phpmyadmin-4.9.7\vendor\paragonie\constant_time_encoding\.travis.yml | .travis.yml | YAML | 240 | 0.7 | 0 | 0 | node-utils | 396 | 2024-05-28T05:43:34.154658 | BSD-3-Clause | false | e76307148e14f04456a13917e4127399 |
# Configuration for probot-stale - https://github.com/probot/stale\n\n# Number of days of inactivity before an Issue or Pull Request becomes stale\ndaysUntilStale: 60\n# Number of days of inactivity before a stale Issue or Pull Request is closed\ndaysUntilClose: 7\n# Issues or Pull Requests with these labels will never be considered stale. Set to `[]` to disable\nexemptLabels:\n - pinned\n - security\n# Label to use when marking as stale\nstaleLabel: wontfix\n# Comment to post when marking as stale. Set to `false` to disable\nmarkComment: >\n This issue has been automatically marked as stale because it has not had\n recent activity. It will be closed if no further activity occurs. Thank you\n for your contributions.\n# Comment to post when removing the stale label. Set to `false` to disable\nunmarkComment: false\n# Comment to post when closing a stale Issue or Pull Request. Set to `false` to disable\ncloseComment: false\n# Limit to only `issues` or `pulls`\nonly: pulls\n | dataset_sample\yaml\cytopia_devilbox\.devilbox\www\htdocs\vendor\phpmyadmin-4.9.7\vendor\phpmyadmin\motranslator\.github\stale.yml | stale.yml | YAML | 967 | 0.8 | 0.130435 | 0.409091 | awesome-app | 816 | 2025-01-23T02:20:13.044585 | BSD-3-Clause | false | 90983e0d8bacbadf842fdcc08115b5f9 |
build: false\nshallow_clone: false\nplatform:\n - x86\n - x64\nclone_folder: C:\projects\phpseclib\n\ninstall:\n - cinst -y OpenSSL.Light\n - SET PATH=C:\Program Files\OpenSSL;%PATH%\n - sc config wuauserv start= auto\n - net start wuauserv\n - cinst -y php --version 5.6.30\n - cd c:\tools\php56\n - copy php.ini-production php.ini\n - echo date.timezone="UTC" >> php.ini\n - echo extension_dir=ext >> php.ini\n - echo extension=php_openssl.dll >> php.ini\n - echo extension=php_gmp.dll >> php.ini\n - cd C:\projects\phpseclib\n - SET PATH=C:\tools\php56;%PATH%\n - php.exe -r "readfile('http://getcomposer.org/installer');" | php.exe\n - php.exe composer.phar install --prefer-source --no-interaction\n\ntest_script:\n - cd C:\projects\phpseclib\n - vendor\bin\phpunit.bat tests/Windows32Test.php | dataset_sample\yaml\cytopia_devilbox\.devilbox\www\htdocs\vendor\phpmyadmin-4.9.7\vendor\phpseclib\phpseclib\appveyor.yml | appveyor.yml | YAML | 791 | 0.8 | 0 | 0 | react-lib | 350 | 2023-09-26T13:17:38.803329 | Apache-2.0 | false | 474ccecb63ad22a214483411eb6aa793 |
checks:\n php:\n remove_extra_empty_lines: true\n remove_php_closing_tag: true\n remove_trailing_whitespace: true\n fix_use_statements:\n remove_unused: true\n preserve_multiple: false\n preserve_blanklines: true\n order_alphabetically: true\n fix_php_opening_tag: true\n fix_linefeed: true\n fix_line_ending: true\n fix_identation_4spaces: true\n fix_doc_comments: true\n\nfilter:\n paths: [src/*]\n excluded_paths: [tests/*]\n\ncoding_style:\n php: { }\n\nbuild:\n tests:\n override:\n -\n command: 'vendor/bin/phpunit -c phpunit.xml'\n coverage:\n file: 'coverage/coverage-clover.xml'\n format: 'clover'\n nodes:\n analysis:\n tests:\n override:\n - php-scrutinizer-run\n tests: true\n\ntools:\n external_code_coverage: true\n | dataset_sample\yaml\cytopia_devilbox\.devilbox\www\htdocs\vendor\phpmyadmin-4.9.7\vendor\pragmarx\google2fa\.scrutinizer.yml | .scrutinizer.yml | YAML | 946 | 0.8 | 0 | 0 | awesome-app | 971 | 2024-12-27T02:45:00.238969 | Apache-2.0 | false | 5d36cd85a01f491e590ce998add650a1 |
dist: trusty\nlanguage: php\n\nphp:\n - 5.4\n - 5.5\n - 5.6\n - 7.0\n - 7.1\n - 7.2\n - nightly\n\n# This triggers builds to run on the new TravisCI infrastructure.\n# See: http://docs.travis-ci.com/user/workers/container-based-infrastructure/\nsudo: false\n\n## Cache composer\ncache:\n directories:\n - $HOME/.composer/cache\n\nbefore_script:\n - travis_retry composer update ${COMPOSER_FLAGS} --no-interaction --prefer-dist\n\nscript:\n - vendor/bin/phpunit --coverage-text --coverage-clover=coverage.clover\n\nafter_script:\n - |\n if [[ "$TRAVIS_PHP_VERSION" == '7.2' ]]; then\n wget https://scrutinizer-ci.com/ocular.phar\n php ocular.phar code-coverage:upload --format=php-clover coverage.clover\n fi\n | dataset_sample\yaml\cytopia_devilbox\.devilbox\www\htdocs\vendor\phpmyadmin-4.9.7\vendor\pragmarx\google2fa\.travis.yml | .travis.yml | YAML | 706 | 0.8 | 0.030303 | 0.111111 | python-kit | 269 | 2023-10-14T08:55:27.259220 | BSD-3-Clause | false | b9c4166b1c7cb75b8606a29498299cb6 |
custom: ['https://www.paypal.com/cgi-bin/webscr?cmd=_donations¤cy_code=GBP&business=paypal@tecnick.com&item_name=donation%20for%20tcpdf%20project']\n | dataset_sample\yaml\cytopia_devilbox\.devilbox\www\htdocs\vendor\phpmyadmin-4.9.7\vendor\tecnickcom\tcpdf\.github\FUNDING.yml | FUNDING.yml | YAML | 154 | 0.8 | 0 | 0 | vue-tools | 994 | 2025-07-01T06:36:39.505547 | Apache-2.0 | false | 1380e6f3e59ef381152c697b92ee1232 |
language: php\n\nsudo: false\n\ncache:\n directories:\n - vendor\n - $HOME/.composer/cache/files\n\nenv:\n - DEPS=no\n\nbefore_install:\n - phpenv config-rm xdebug.ini\n\nbefore_script:\n - if [ "$DEPS" == "low" ]; then composer --prefer-lowest --prefer-stable update; fi;\n - if [ "$DEPS" == "no" ]; then composer install; fi;\n\nscript: |\n ./vendor/bin/simple-phpunit\n\nmatrix:\n include:\n - php: 5.3\n dist: precise\n - php: 5.4\n - php: 5.5\n - php: 5.6\n env: DEPS=low\n - php: 7.0\n - php: 7.1\n - php: 7.2\n fast_finish: true\n | dataset_sample\yaml\cytopia_devilbox\.devilbox\www\htdocs\vendor\phpmyadmin-4.9.7\vendor\twig\extensions\.travis.yml | .travis.yml | YAML | 608 | 0.7 | 0.058824 | 0 | react-lib | 760 | 2023-09-26T08:02:21.030428 | Apache-2.0 | false | 0dd5240f444ad066e2796d5d3cfcd6af |
language: php\n\ndist: trusty\n\nsudo: false\n\ncache:\n directories:\n - vendor\n - $HOME/.composer/cache/files\n\n\nenv:\n global:\n - TWIG_EXT=no\n - SYMFONY_PHPUNIT_REMOVE_RETURN_TYPEHINT=1\n\nbefore_install:\n - phpenv config-rm xdebug.ini || return 0\n\ninstall:\n - travis_retry composer install\n\nbefore_script:\n - if [ "$TWIG_EXT" == "yes" ]; then sh -c "cd ext/twig && phpize && ./configure --enable-twig && make && make install"; fi\n - if [ "$TWIG_EXT" == "yes" ]; then echo "extension=twig.so" >> `php --ini | grep "Loaded Configuration" | sed -e "s|.*:\s*||"`; fi\n\nscript: ./vendor/bin/simple-phpunit\n\njobs:\n fast_finish: true\n include:\n - php: 5.5\n - php: 5.5\n env: TWIG_EXT=yes\n - php: 5.6\n - php: 5.6\n env: TWIG_EXT=yes\n - php: 7.0\n - php: 7.1\n - php: 7.2\n - php: 7.3\n - php: 7.4snapshot\n - stage: integration tests\n php: 7.3\n script: ./drupal_test.sh\n | dataset_sample\yaml\cytopia_devilbox\.devilbox\www\htdocs\vendor\phpmyadmin-4.9.7\vendor\twig\twig\.travis.yml | .travis.yml | YAML | 1,004 | 0.7 | 0.043478 | 0 | python-kit | 407 | 2024-05-12T09:02:19.134196 | MIT | false | 1e7c0c1a85ea7b4efe909fa64374dcb4 |
# These are supported funding model platforms\n\ngithub: [d2phap]\npatreon: d2phap\nopen_collective: # Replace with a single Open Collective username\nko_fi: # Replace with a single Ko-fi username\ntidelift: # Replace with a single Tidelift platform-name/package-name e.g., npm/babel\ncustom: ["https://donate.stripe.com/6oE15Kab3740du828a", "https://www.paypal.me/d2phap"]\n | dataset_sample\yaml\d2phap_ImageGlass\.github\FUNDING.yml | FUNDING.yml | YAML | 367 | 0.8 | 0 | 0.142857 | awesome-app | 806 | 2023-09-10T07:28:09.304502 | Apache-2.0 | false | 684359257bace3ec2488eee3d4ab9aa8 |
blank_issues_enabled: false\ncontact_links:\n - name: Ask question\n url: https://github.com/d2phap/ImageGlass/discussions/new/choose\n about: Ask a question about ImageGlass features.\n \n - name: Ask Discord community\n url: https://discord.gg/tWjbynH2X8\n about: Join in official Discord server of ImageGlass.\n | dataset_sample\yaml\d2phap_ImageGlass\.github\ISSUE_TEMPLATE\config.yml | config.yml | YAML | 322 | 0.8 | 0 | 0 | vue-tools | 520 | 2024-06-30T01:50:24.236279 | Apache-2.0 | false | c9eaf73334408ed96202314a78d2a341 |
name: Deploy\n\non:\n workflow_dispatch: {}\n push:\n branches: [main]\n\njobs:\n deploy:\n runs-on: ubuntu-latest\n permissions:\n pages: write\n id-token: write\n environment:\n name: github-pages\n url: ${{ steps.deployment.outputs.page_url }}\n steps:\n - uses: actions/checkout@v4\n - uses: actions/checkout@v4\n with:\n repository: d3/d3.github.com\n path: build/d3.github.com\n - uses: actions/setup-node@v4\n with:\n node-version: 20\n cache: 'yarn'\n - run: yarn --frozen-lockfile\n - run: yarn prepublishOnly\n - run: yarn docs:build\n - uses: actions/configure-pages@v4\n - uses: actions/upload-pages-artifact@v3\n with:\n path: docs/.vitepress/dist\n - name: Deploy\n id: deployment\n uses: actions/deploy-pages@v4\n | dataset_sample\yaml\d3_d3\.github\workflows\deploy.yml | deploy.yml | YAML | 855 | 0.7 | 0 | 0 | python-kit | 465 | 2024-04-20T14:58:16.088929 | MIT | false | e169f773821486452cb45264cfb7f49e |
version: 2\n\nproject_name: dagger\n\nbuilds:\n - builder: prebuilt\n binary: ./dagger\n goos:\n - linux\n - windows\n - darwin\n goarch:\n - amd64\n - arm64\n - arm\n goarm:\n - "7"\n\n prebuilt:\n path: build/dagger_{{ .Env.ENGINE_VERSION }}_{{ .Os }}_{{ .Arch }}{{ with .Arm }}_v{{ . }}{{ end }}/dagger\n\nchecksum:\n name_template: "checksums.txt"\n | dataset_sample\yaml\dagger_dagger\.goreleaser.common.yml | .goreleaser.common.yml | YAML | 387 | 0.7 | 0 | 0 | react-lib | 718 | 2023-11-24T07:49:51.299171 | GPL-3.0 | false | 1af3d4425f03dd36841f7988f50f0208 |
version: 2\n\nincludes:\n - from_file:\n path: ./.goreleaser.common.yml\n\nnightly:\n # version_template will override .Version for nightly builds:\n # https://goreleaser.com/customization/nightlies/#how-it-works\n version_template: "{{ .FullCommit }}"\n\narchives:\n - name_template: "{{ .ProjectName }}_{{ .Version }}_{{ .Os }}_{{ .Arch }}{{ if .Arm }}v{{ .Arm }}{{ end }}{{ if .Mips }}_{{ .Mips }}{{ end }}"\n id: sha\n files:\n - LICENSE\n format_overrides:\n - goos: windows\n formats: [zip]\n - name_template: "{{ .ProjectName }}_head_{{ .Os }}_{{ .Arch }}{{ if .Arm }}v{{ .Arm }}{{ end }}{{ if .Mips }}_{{ .Mips }}{{ end }}"\n id: head\n files:\n - LICENSE\n format_overrides:\n - goos: windows\n formats: [zip]\n\nblobs:\n - provider: s3\n ids:\n - sha\n region: "{{ .Env.AWS_REGION }}"\n bucket: "{{ .Env.AWS_BUCKET }}"\n directory: "dagger/main/{{ .Version }}"\n - provider: s3\n ids:\n - head\n region: "{{ .Env.AWS_REGION }}"\n bucket: "{{ .Env.AWS_BUCKET }}"\n directory: "dagger/main/head"\n | dataset_sample\yaml\dagger_dagger\.goreleaser.nightly.yml | .goreleaser.nightly.yml | YAML | 1,064 | 0.8 | 0.125 | 0.055556 | awesome-app | 533 | 2024-06-09T22:10:53.034773 | MIT | false | 04fc446628aeb7a70a43341817e8ed54 |
version: 2\n\nincludes:\n - from_file:\n path: ./.goreleaser.common.yml\n\nnightly:\n # version_template will override .Version for nightly builds:\n # https://goreleaser.com/customization/nightlies/#how-it-works\n # version should *not* have a v prefix\n version_template: "{{ trimprefix .Tag \"v\" }}"\n\narchives:\n - name_template: "{{ .ProjectName }}_v{{ .Version }}_{{ .Os }}_{{ .Arch }}{{ if .Arm }}v{{ .Arm }}{{ end }}{{ if .Mips }}_{{ .Mips }}{{ end }}"\n id: sha\n files:\n - LICENSE\n format_overrides:\n - goos: windows\n formats: [zip]\n\nblobs:\n - provider: s3\n region: "{{ .Env.AWS_REGION }}"\n bucket: "{{ .Env.AWS_BUCKET }}"\n directory: "dagger/releases/{{ .Version }}"\n | dataset_sample\yaml\dagger_dagger\.goreleaser.prerelease.yml | .goreleaser.prerelease.yml | YAML | 712 | 0.8 | 0.115385 | 0.136364 | awesome-app | 5 | 2024-08-23T02:45:31.388754 | MIT | false | 08d23202fa2620ee56b950b04a114976 |
# yaml-language-server: $schema=https://raw.githubusercontent.com/goreleaser/goreleaser/v2.7.0/www/docs/static/schema-pro.json\n\nversion: 2\n\nincludes:\n - from_file:\n path: ./.goreleaser.common.yml\n\nsnapshot:\n version_template: "{{ .Tag }}-next"\n\narchives:\n - name_template: "{{ .ProjectName }}_{{ .Tag }}_{{ .Os }}_{{ .Arch }}{{ if .Arm }}v{{ .Arm }}{{ end }}{{ if .Mips }}_{{ .Mips }}{{ end }}"\n files:\n - LICENSE\n format_overrides:\n - goos: windows\n formats: [zip]\n\n# https://goreleaser.com/customization/homebrew/\nbrews:\n - repository:\n owner: "{{ .Env.GH_ORG_NAME }}"\n name: homebrew-tap\n name: dagger\n commit_author:\n name: dagger-bot\n email: noreply@dagger.io\n url_template: "https://{{ .Env.ARTEFACTS_FQDN }}/dagger/releases/{{ .Version }}/{{ .ArtifactName }}"\n homepage: "https://dagger.io"\n description: "Dagger is an integrated platform to orchestrate the delivery of applications"\n test: |\n system "#{bin}/dagger version"\n\n# https://goreleaser.com/customization/nix/\nnix:\n - name: dagger\n repository:\n owner: "{{ .Env.GH_ORG_NAME }}"\n name: nix\n commit_author:\n name: dagger-bot\n email: noreply@dagger.io\n url_template: "https://{{ .Env.ARTEFACTS_FQDN }}/dagger/releases/{{ .Version }}/{{ .ArtifactName }}"\n homepage: "https://dagger.io"\n description: "Dagger is an integrated platform to orchestrate the delivery of applications"\n license: "asl20"\n post_install: |\n installShellCompletion --cmd dagger \\n --bash <($out/bin/dagger completion bash) \\n --fish <($out/bin/dagger completion fish) \\n --zsh <($out/bin/dagger completion zsh)\n\n# https://goreleaser.com/customization/winget/\nwinget:\n - name: dagger\n repository:\n owner: "{{ .Env.GH_ORG_NAME }}"\n name: winget-pkgs\n branch: "{{ .ProjectName }}-{{ .Version }}"\n pull_request:\n enabled: true\n check_boxes: true\n base:\n owner: microsoft\n name: winget-pkgs\n branch: master\n commit_author:\n name: dagger-bot\n email: noreply@dagger.io\n url_template: "https://{{ .Env.ARTEFACTS_FQDN }}/dagger/releases/{{ .Version }}/{{ .ArtifactName }}"\n homepage: "https://dagger.io"\n publisher: Dagger\n short_description: "Dagger is an integrated platform to orchestrate the delivery of applications"\n license: "asl20"\n publisher_url: "https://dagger.io"\n publisher_support_url: https://github.com/dagger/dagger/issues/new/choose\n path: manifests/d/Dagger/Cli/{{ .Version }}\n package_identifier: Dagger.Cli\n use: archive\n skip_upload: false\n tags:\n - dagger\n - cli\n - cicd\n - workflows\n - sandbox\n - containers\n - devops\n - llm\n\nblobs:\n - provider: s3\n region: "{{ .Env.AWS_REGION }}"\n bucket: "{{ .Env.AWS_BUCKET }}"\n directory: "dagger/releases/{{ .Version }}"\n | dataset_sample\yaml\dagger_dagger\.goreleaser.yml | .goreleaser.yml | YAML | 2,924 | 0.8 | 0.020833 | 0.045455 | vue-tools | 828 | 2024-06-04T20:22:12.961057 | Apache-2.0 | false | dc24996fb26ef957e5c3c2de23f35526 |
version: '3.8'\n\nservices:\n fabric-api:\n build: .\n ports:\n - "8080:8080"\n volumes:\n - ./ENV:/root/.config/fabric/.env:ro\n environment:\n - GIN_MODE=release | dataset_sample\yaml\danielmiessler_fabric\docker-compose.yml | docker-compose.yml | YAML | 182 | 0.7 | 0 | 0 | python-kit | 307 | 2024-07-30T13:55:23.018204 | GPL-3.0 | false | 90b125e8a8fbf58ce984dc1d04f94a4f |
image: Visual Studio 2022\n\nskip_branch_with_pr: true\nskip_tags: true\nskip_commits:\n files:\n - '**/*.md'\n\ninstall:\n - choco install dotnet-sdk --version 9.0.101\n\nenvironment:\n Appveyor: true\n # Postgres\n POSTGRES_PATH: C:\Program Files\PostgreSQL\16\n PGUSER: postgres\n PGPASSWORD: Password12!\n POSTGRES_ENV_POSTGRES_USER: postgres\n POSTGRES_ENV_POSTGRES_PASSWORD: Password12!\n POSTGRES_ENV_POSTGRES_DB: test\n # MySQL\n MYSQL_PATH: C:\Program Files\MySQL\MySQL Server 8.0\n MYSQL_PWD: Password12!\n MYSQL_ENV_MYSQL_USER: root\n MYSQL_ENV_MYSQL_PASSWORD: Password12!\n MYSQL_ENV_MYSQL_DATABASE: test\n # Connection strings for tests:\n MySqlConnectionString: Server=localhost;Database=test;Uid=root;Pwd=Password12!\n OLEDBConnectionString: Provider=SQLOLEDB;Data Source=(local)\SQL2019;Initial Catalog=tempdb;User Id=sa;Password=Password12!\n PostgesConnectionString: Server=localhost;Port=5432;User Id=postgres;Password=Password12!;Database=test\n SqlServerConnectionString: Server=(local)\SQL2019;Database=tempdb;User ID=sa;Password=Password12!\n\ninit:\n - git config --global core.autocrlf input\n - SET PATH=%POSTGRES_PATH%\bin;%MYSQL_PATH%\bin;%PATH%\n - net start MSSQL$SQL2019\n - net start postgresql-x64-16\n - ps: Start-Service MySQL80\n \nnuget:\n disable_publish_on_pr: true\n\nbefore_build:\n # Postgres\n - createdb test\n # MySQL\n - '"C:\Program Files\MySQL\MySQL Server 8.0\bin\mysql" -e "create database test;" --user=root'\n\nbuild_script:\n # Our stuff\n - ps: .\build.ps1 -PullRequestNumber "$env:APPVEYOR_PULL_REQUEST_NUMBER" -CreatePackages $true\n\ntest: off\nartifacts:\n - path: .\.nupkgs\*.nupkg\n \ndeploy:\n- provider: NuGet\n server: https://www.myget.org/F/stackoverflow/api/v2\n on:\n branch: main \n api_key:\n secure: P/UHxq2DEs0GI1SoDXDesHjRVsSVgdywz5vmsnhFQQY5aJgO3kP+QfhwfhXz19Rw\n symbol_server: https://www.myget.org/F/stackoverflow/symbols/api/v2/package\n- provider: NuGet\n server: https://www.myget.org/F/dapper/api/v2\n on:\n branch: main \n api_key:\n secure: PV7ERAltWWLhy7AT2h+Vb5c1BM9/WFgvggb+rKyQ8hDg3fYqpZauYdidOOgt2lp4\n symbol_server: https://www.myget.org/F/dapper/api/v2/package | dataset_sample\yaml\DapperLib_Dapper\appveyor.yml | appveyor.yml | YAML | 2,143 | 0.8 | 0.014286 | 0.096774 | react-lib | 386 | 2023-12-19T07:22:47.676696 | GPL-3.0 | false | 16ab51415b21fb9f8f66c5a3a2dcdc94 |
# options for analysis running\nrun:\n # default concurrency is a available CPU number\n concurrency: 4\n\n # timeout for analysis, e.g. 30s, 5m, default is 1m\n timeout: 15m\n\n # exit code when at least one issue was found, default is 1\n issues-exit-code: 1\n\n # include test files or not, default is true\n tests: true\n\n # list of build tags, all linters use it. Default is empty list.\n build-tags:\n - unit\n - allcomponents\n - subtlecrypto\n\n # which files to skip: they will be analyzed, but issues from them\n # won't be reported. Default value is empty list, but there is\n # no need to include all autogenerated files, we confidently recognize\n # autogenerated files. If it's not please let us know.\n # skip-files:\n # - ".*\\.my\\.go$"\n # - lib/bad.go\n\nissues:\n # which dirs to skip: they won't be analyzed;\n # can use regexp here: generated.*, regexp is applied on full path;\n # default value is empty list, but next dirs are always skipped independently\n # from this option's value:\n # third_party$, testdata$, examples$, Godeps$, builtin$\n exclude-dirs:\n - ^pkg.*client.*clientset.*versioned.*\n - ^pkg.*client.*informers.*externalversions.*\n - ^pkg.*proto.*\n - pkg/proto\n\n# output configuration options\noutput:\n # colored-line-number|line-number|json|tab|checkstyle, default is "colored-line-number"\n formats:\n - format: tab\n\n # print lines of code with issue, default is true\n print-issued-lines: true\n\n # print linter name in the end of issue text, default is true\n print-linter-name: true\n\n\n# all available settings of specific linters\nlinters-settings:\n errcheck:\n # report about not checking of errors in type assetions: `a := b.(MyStruct)`;\n # default is false: such cases aren't reported by default.\n check-type-assertions: false\n\n # report about assignment of errors to blank identifier: `num, _ := strconv.Atoi(numStr)`;\n # default is false: such cases aren't reported by default.\n check-blank: false\n\n exclude-functions:\n - fmt:.*\n - io/ioutil:^Read.*\n\n # path to a file containing a list of functions to exclude from checking\n # see https://github.com/kisielk/errcheck#excluding-functions for details\n # exclude:\n\n funlen:\n lines: 60\n statements: 40\n\n govet:\n # settings per analyzer\n settings:\n printf: # analyzer name, run `go tool vet help` to see all analyzers\n funcs: # run `go tool vet help printf` to see available settings for `printf` analyzer\n - (github.com/golangci/golangci-lint/pkg/logutils.Log).Infof\n - (github.com/golangci/golangci-lint/pkg/logutils.Log).Warnf\n - (github.com/golangci/golangci-lint/pkg/logutils.Log).Errorf\n - (github.com/golangci/golangci-lint/pkg/logutils.Log).Fatalf\n\n # enable or disable analyzers by name\n enable:\n - atomicalign\n - shadow\n enable-all: false\n disable-all: false\n revive:\n # minimal confidence for issues, default is 0.8\n confidence: 0.8\n gofmt:\n # simplify code: gofmt with `-s` option, true by default\n simplify: true\n goimports:\n # put imports beginning with prefix after 3rd-party packages;\n # it's a comma-separated list of prefixes\n local-prefixes: github.com/dapr/\n gocyclo:\n # minimal code complexity to report, 30 by default (but we recommend 10-20)\n min-complexity: 10\n gocognit:\n # minimal code complexity to report, 30 by default (but we recommend 10-20)\n min-complexity: 10\n dupl:\n # tokens count to trigger issue, 150 by default\n threshold: 100\n goconst:\n # minimal length of string constant, 3 by default\n min-len: 3\n # minimal occurrences count to trigger, 3 by default\n min-occurrences: 5\n depguard:\n rules:\n master:\n deny:\n - pkg: "github.com/Sirupsen/logrus"\n desc: "must use github.com/dapr/kit/logger"\n - pkg: "github.com/agrea/ptr"\n desc: "must use github.com/dapr/kit/ptr"\n - pkg: "go.uber.org/atomic"\n desc: "must use sync/atomic"\n - pkg: "golang.org/x/net/context"\n desc: "must use context"\n - pkg: "github.com/pkg/errors"\n desc: "must use standard library (errors package and/or fmt.Errorf)"\n - pkg: "github.com/go-chi/chi$"\n desc: "must use github.com/go-chi/chi/v5"\n - pkg: "github.com/cenkalti/backoff$"\n desc: "must use github.com/cenkalti/backoff/v4"\n - pkg: "github.com/cenkalti/backoff/v2"\n desc: "must use github.com/cenkalti/backoff/v4"\n - pkg: "github.com/cenkalti/backoff/v3"\n desc: "must use github.com/cenkalti/backoff/v4"\n - pkg: "github.com/benbjohnson/clock"\n desc: "must use k8s.io/utils/clock"\n - pkg: "github.com/ghodss/yaml"\n desc: "must use sigs.k8s.io/yaml"\n - pkg: "gopkg.in/yaml.v2"\n desc: "must use gopkg.in/yaml.v3"\n - pkg: "github.com/golang-jwt/jwt"\n desc: "must use github.com/lestrrat-go/jwx/v2"\n - pkg: "github.com/golang-jwt/jwt/v2"\n desc: "must use github.com/lestrrat-go/jwx/v2"\n - pkg: "github.com/golang-jwt/jwt/v3"\n desc: "must use github.com/lestrrat-go/jwx/v2"\n - pkg: "github.com/golang-jwt/jwt/v4"\n desc: "must use github.com/lestrrat-go/jwx/v2"\n # pkg: Commonly auto-completed by gopls\n - pkg: "github.com/gogo/status"\n desc: "must use google.golang.org/grpc/status"\n - pkg: "github.com/gogo/protobuf"\n desc: "must use google.golang.org/protobuf"\n - pkg: "github.com/lestrrat-go/jwx/jwa"\n desc: "must use github.com/lestrrat-go/jwx/v2"\n - pkg: "github.com/lestrrat-go/jwx/jwt"\n desc: "must use github.com/lestrrat-go/jwx/v2"\n - pkg: "github.com/labstack/gommon/log"\n desc: "must use github.com/dapr/kit/logger"\n - pkg: "github.com/gobuffalo/logger"\n desc: "must use github.com/dapr/kit/logger"\n - pkg: "k8s.io/utils/pointer"\n desc: "must use github.com/dapr/kit/ptr"\n - pkg: "k8s.io/utils/ptr"\n desc: "must use github.com/dapr/kit/ptr"\n misspell:\n # Correct spellings using locale preferences for US or UK.\n # Default is to use a neutral variety of English.\n # Setting locale to US will correct the British spelling of 'colour' to 'color'.\n # locale: default\n ignore-words:\n - someword\n lll:\n # max line length, lines longer will be reported. Default is 120.\n # '\t' is counted as 1 character by default, and can be changed with the tab-width option\n line-length: 120\n # tab width in spaces. Default to 1.\n tab-width: 1\n nakedret:\n # make an issue if func has more lines of code than this setting and it has naked returns; default is 30\n max-func-lines: 30\n prealloc:\n # XXX: we don't recommend using this linter before doing performance profiling.\n # For most programs usage of prealloc will be a premature optimization.\n\n # Report preallocation suggestions only on simple loops that have no returns/breaks/continues/gotos in them.\n # True by default.\n simple: true\n range-loops: true # Report preallocation suggestions on range loops, true by default\n for-loops: false # Report preallocation suggestions on for loops, false by default\n gocritic:\n # Which checks should be enabled; can't be combined with 'disabled-checks';\n # See https://go-critic.github.io/overview#checks-overview\n # To check which checks are enabled run `GL_DEBUG=gocritic golangci-lint run`\n # By default list of stable checks is used.\n # enabled-checks:\n\n # Which checks should be disabled; can't be combined with 'enabled-checks'; default is empty\n disabled-checks:\n - regexpMust\n - rangeValCopy\n - hugeParam\n - ifElseChain\n - singleCaseSwitch\n - exitAfterDefer\n\n # Enable multiple checks by tags, run `GL_DEBUG=gocritic golangci-lint run` to see all tags and checks.\n # Empty list by default. See https://github.com/go-critic/go-critic#usage -> section "Tags".\n enabled-tags:\n - performance\n\n settings: # settings passed to gocritic\n captLocal: # must be valid enabled check name\n paramsOnly: true\n godox:\n # report any comments starting with keywords, this is useful for TODO or FIXME comments that\n # might be left in the code accidentally and should be resolved before merging\n keywords: # default keywords are TODO, BUG, and FIXME, these can be overwritten by this setting\n - NOTE\n - OPTIMIZE # marks code that should be optimized before merging\n - HACK # marks hack-arounds that should be removed before merging\n godot:\n exclude:\n - 'nosec'\n capital: false\n scope: all\n dogsled:\n # checks assignments with too many blank identifiers; default is 2\n max-blank-identifiers: 2\n\n whitespace:\n multi-if: false # Enforces newlines (or comments) after every multi-line if statement\n multi-func: false # Enforces newlines (or comments) after every multi-line function signature\n\n wsl:\n # If true append is only allowed to be cuddled if appending value is\n # matching variables, fields or types on line above. Default is true.\n strict-append: true\n # Allow calls and assignments to be cuddled as long as the lines have any\n # matching variables, fields or types. Default is true.\n allow-assign-and-call: true\n # Allow multiline assignments to be cuddled. Default is true.\n allow-multiline-assign: true\n # Allow declarations (var) to be cuddled.\n allow-cuddle-declarations: false\n # If the number of lines in a case block is equal to or lager than this number,\n # the case *must* end white a newline.\n # https://github.com/bombsimon/wsl/blob/master/doc/configuration.md#force-case-trailing-whitespace\n # Default: 0\n force-case-trailing-whitespace: 1\n\nlinters:\n fast: false\n enable-all: true\n disable:\n # TODO Enforce the below linters later\n - musttag\n - dupl\n - errcheck\n - funlen\n - gochecknoglobals\n - gochecknoinits\n - gocyclo\n - gocognit\n - godox\n - lll\n - unparam\n - wsl\n - testpackage\n - nestif\n - nlreturn\n - exhaustive\n - exhaustruct\n - noctx\n - gci\n - tparallel\n - paralleltest\n - wrapcheck\n - tagliatelle\n - ireturn\n - errchkjson\n - contextcheck\n - gomoddirectives\n - godot\n - cyclop\n - varnamelen\n - errorlint\n - forcetypeassert\n - maintidx\n - nilnil\n - predeclared\n - tenv\n - thelper\n - wastedassign\n - containedctx\n - gosimple\n - nonamedreturns\n - asasalint\n - rowserrcheck\n - sqlclosecheck\n - inamedparam\n - tagalign\n - mnd\n - canonicalheader\n - err113\n - fatcontext\n | dataset_sample\yaml\dapr_dapr\.golangci.yml | .golangci.yml | YAML | 10,729 | 0.95 | 0.045161 | 0.265734 | react-lib | 976 | 2024-04-26T20:35:05.635477 | Apache-2.0 | false | 87a73d2432168bb08f04849fe1266579 |
\nname: "CodeQL"\n\non: push\n # schedule:\n # - cron: '0 15 * * *' \n\njobs:\n\n analyze:\n name: Analyze\n \n runs-on: ubuntu-latest\n\n permissions:\n actions: read\n contents: read\n security-events: write\n\n steps:\n - name: Checkout repository\n uses: actions/checkout@v2\n with:\n fetch-depth: 2\n\n - uses: dbeaver/dbeaver/.github/actions/java-codeql@devel\n | dataset_sample\yaml\dbeaver_dbeaver\.github\codeql-analysis.yml | codeql-analysis.yml | YAML | 399 | 0.8 | 0 | 0.111111 | python-kit | 871 | 2024-08-24T16:34:21.530828 | Apache-2.0 | false | eabcbfd05851f4d6767f486754598d7d |
blank_issues_enabled: false\ncontact_links:\n - name: How to extract log files from DBeaver\n url: https://github.com/dbeaver/dbeaver/wiki/Log-files\n about: Common things you should know before creating a bug report.\n - name: Try advanced database management features in DBeaver PRO\n url: https://dbeaver.com/download/?utm_source=github&utm_medium=social&utm_campaign=issue\n about: Download a 14-day free trial on dbeaver.com. No credit card is needed.\n - name: Learn how to use DBeaver effectively with our tutorial videos\n url: https://www.youtube.com/channel/UC-yOjsQLSaJVEghg6UB3N7A\n about: Subscribe to our youtube channel and watch new tutorials every week\n - name: Follow us on X (ex. Twitter)\n url: https://x.com/dbeaver_news\n about: Get everyday tips, fresh news, and DBeaver feature overviews.\n | dataset_sample\yaml\dbeaver_dbeaver\.github\ISSUE_TEMPLATE\config.yml | config.yml | YAML | 829 | 0.8 | 0 | 0 | python-kit | 813 | 2024-01-14T11:49:26.429376 | MIT | false | 4b1018385724a17b5d36585a69c31f29 |
name: PR Check\n\non:\n pull_request:\n types:\n - opened\n - synchronize\n - reopened\n push:\n branches: [devel]\n\nconcurrency:\n group: ${{ github.workflow }}-${{ github.event_name == 'pull_request' && github.event.pull_request.number || github.ref_name }}\n cancel-in-progress: true\n\njobs:\n build-java:\n uses: dbeaver/dbeaver-common/.github/workflows/mvn-package.yml@devel\n name: Build\n secrets: inherit\n with:\n project-directory: ./dbeaver/product/aggregate\n project-deps: ./dbeaver/project.deps\n timeout-minutes: 5\n\n lint-java:\n name: Check style\n uses: dbeaver/dbeaver-common/.github/workflows/java-checkstyle.yml@devel\n secrets: inherit\n | dataset_sample\yaml\dbeaver_dbeaver\.github\workflows\push-pr-devel.yml | push-pr-devel.yml | YAML | 696 | 0.7 | 0 | 0 | awesome-app | 389 | 2024-02-07T03:20:29.657799 | GPL-3.0 | false | 9efa9329c9dfff374ee8d7dcc0795eb1 |
blank_issues_enabled: true\ncontact_links:\n - name: 🦕 Deploy Feedback\n url: https://github.com/denoland/deploy_feedback/\n about: Provide feature requests or bug reports for the Deno Deploy edge runtime.\n - name: 💬 Discord\n url: https://discord.gg/deno\n about: Join our community on Discord.\n | dataset_sample\yaml\denoland_deno\.github\ISSUE_TEMPLATE\config.yml | config.yml | YAML | 310 | 0.8 | 0.125 | 0 | vue-tools | 25 | 2023-09-01T14:07:35.263119 | Apache-2.0 | false | 9413d743e6da74b6da47b0a4c3d0e93d |
name: mtime cache\ndescription:\n Preserve last-modified timestamps by storing them in the Github Actions cache\ninputs:\n cache-path:\n description: Path where the mtime cache database should be located\n required: true\nruns:\n main: action.js\n using: node20\n | dataset_sample\yaml\denoland_deno\.github\mtime_cache\action.yml | action.yml | YAML | 263 | 0.85 | 0 | 0 | awesome-app | 105 | 2024-02-16T08:32:06.609949 | Apache-2.0 | false | cec9ac9f2da43836f09f586ee2629c71 |
name: cargo_publish\n\non: workflow_dispatch\n\n# Ensures only one publish is running at a time\nconcurrency:\n group: ${{ github.workflow }}\n cancel-in-progress: true\n\njobs:\n build:\n name: cargo publish\n runs-on: ubuntu-24.04-xl\n timeout-minutes: 90\n\n env:\n CARGO_TERM_COLOR: always\n RUST_BACKTRACE: full\n RUSTC_FORCE_INCREMENTAL: 1\n\n steps:\n - name: Configure git\n run: |\n git config --global core.symlinks true\n git config --global fetch.parallel 32\n\n - name: Clone repository\n uses: actions/checkout@v4\n with:\n fetch-depth: 0\n token: ${{ secrets.DENOBOT_PAT }}\n submodules: recursive\n\n - uses: dsherret/rust-toolchain-file@v1\n\n - name: Install deno\n uses: denoland/setup-deno@v2\n with:\n deno-version: v2.x\n\n - name: Publish\n env:\n CARGO_REGISTRY_TOKEN: ${{ secrets.CARGO_REGISTRY_TOKEN }}\n run: ./tools/release/03_publish_crates.ts\n\n - name: Create release tag and check forward commit to main\n env:\n # the default secrets.GITHUB_TOKEN won't trigger a workflow run\n # when tagging, but it will if we provide a custom PAT\n GITHUB_TOKEN: ${{ secrets.DENOBOT_PAT }}\n GH_WORKFLOW_ACTOR: ${{ github.actor }}\n run: |\n git config user.email "${{ github.actor }}@users.noreply.github.com"\n git config user.name "${{ github.actor }}"\n ./tools/release/04_post_publish.ts\n | dataset_sample\yaml\denoland_deno\.github\workflows\cargo_publish.yml | cargo_publish.yml | YAML | 1,511 | 0.8 | 0.018182 | 0.066667 | react-lib | 129 | 2023-10-07T09:17:17.198258 | GPL-3.0 | false | a8b3e6d709c4037870c24620e8d5a442 |
# GENERATED BY ./ci.generate.ts -- DO NOT DIRECTLY EDIT\n\nname: ci\npermissions:\n contents: write\non:\n push:\n branches:\n - main\n tags:\n - '*'\n pull_request:\n types:\n - opened\n - reopened\n - synchronize\n - ready_for_review\nconcurrency:\n group: '${{ github.workflow }}-${{ !contains(github.event.pull_request.labels.*.name, ''ci-test-flaky'') && github.head_ref || github.run_id }}'\n cancel-in-progress: true\njobs:\n pre_build:\n name: pre-build\n runs-on: ubuntu-latest\n outputs:\n skip_build: '${{ steps.check.outputs.skip_build }}'\n steps:\n - name: Configure git\n run: |-\n git config --global core.symlinks true\n git config --global fetch.parallel 32\n if: github.event.pull_request.draft == true\n - name: Clone repository\n uses: actions/checkout@v4\n with:\n fetch-depth: 5\n submodules: false\n if: github.event.pull_request.draft == true\n - id: check\n if: 'github.event.pull_request.draft == true && (!contains(github.event.pull_request.labels.*.name, ''ci-draft''))'\n run: |-\n GIT_MESSAGE=$(git log --format=%s -n 1 ${{github.event.after}})\n echo Commit message: $GIT_MESSAGE\n echo $GIT_MESSAGE | grep '\[ci\]' || (echo 'Exiting due to draft PR. Commit with [ci] to bypass or add the ci-draft label.' ; echo 'skip_build=true' >> $GITHUB_OUTPUT)\n build:\n name: '${{ matrix.job }} ${{ matrix.profile }} ${{ matrix.os }}-${{ matrix.arch }}'\n needs:\n - pre_build\n if: '${{ needs.pre_build.outputs.skip_build != ''true'' }}'\n runs-on: '${{ matrix.runner }}'\n timeout-minutes: 240\n defaults:\n run:\n shell: bash\n strategy:\n matrix:\n include:\n - os: macos\n arch: x86_64\n runner: macos-13\n job: test\n profile: debug\n - os: macos\n arch: x86_64\n runner: '${{ (!contains(github.event.pull_request.labels.*.name, ''ci-full'') && (github.event_name == ''pull_request'')) && ''ubuntu-24.04'' || ''macos-13'' }}'\n job: test\n profile: release\n skip: '${{ !contains(github.event.pull_request.labels.*.name, ''ci-full'') && (github.event_name == ''pull_request'') }}'\n - os: macos\n arch: aarch64\n runner: macos-14\n job: test\n profile: debug\n - os: macos\n arch: aarch64\n runner: '${{ (!contains(github.event.pull_request.labels.*.name, ''ci-full'') && (github.event_name == ''pull_request'')) && ''ubuntu-24.04'' || github.repository == ''denoland/deno'' && (github.ref == ''refs/heads/main'' || startsWith(github.ref, ''refs/tags/'')) && ''ghcr.io/cirruslabs/macos-runner:sonoma'' || ''macos-14'' }}'\n job: test\n profile: release\n skip: '${{ !contains(github.event.pull_request.labels.*.name, ''ci-full'') && (github.event_name == ''pull_request'') }}'\n - os: windows\n arch: x86_64\n runner: windows-2022\n job: test\n profile: debug\n - os: windows\n arch: x86_64\n runner: '${{ (!contains(github.event.pull_request.labels.*.name, ''ci-full'') && (github.event_name == ''pull_request'')) && ''ubuntu-24.04'' || github.repository == ''denoland/deno'' && ''windows-2022-xl'' || ''windows-2022'' }}'\n job: test\n profile: release\n skip: '${{ !contains(github.event.pull_request.labels.*.name, ''ci-full'') && (github.event_name == ''pull_request'') }}'\n - os: linux\n arch: x86_64\n runner: '${{ github.repository == ''denoland/deno'' && ''ubuntu-24.04-xl'' || ''ubuntu-24.04'' }}'\n job: test\n profile: release\n use_sysroot: true\n wpt: '${{ !startsWith(github.ref, ''refs/tags/'') }}'\n - os: linux\n arch: x86_64\n runner: '${{ (!contains(github.event.pull_request.labels.*.name, ''ci-full'') && (github.event_name == ''pull_request'' && !contains(github.event.pull_request.labels.*.name, ''ci-bench''))) && ''ubuntu-24.04'' || github.repository == ''denoland/deno'' && ''ubuntu-24.04-xl'' || ''ubuntu-24.04'' }}'\n job: bench\n profile: release\n use_sysroot: true\n skip: '${{ !contains(github.event.pull_request.labels.*.name, ''ci-full'') && (github.event_name == ''pull_request'' && !contains(github.event.pull_request.labels.*.name, ''ci-bench'')) }}'\n - os: linux\n arch: x86_64\n runner: ubuntu-24.04\n job: test\n profile: debug\n use_sysroot: true\n - os: linux\n arch: x86_64\n runner: ubuntu-24.04\n job: lint\n profile: debug\n - os: linux\n arch: aarch64\n runner: ubicloud-standard-16-arm\n job: test\n profile: debug\n - os: linux\n arch: aarch64\n runner: ubicloud-standard-16-arm\n job: test\n profile: release\n use_sysroot: true\n - os: macos\n arch: x86_64\n runner: macos-13\n job: lint\n profile: debug\n - os: windows\n arch: x86_64\n runner: windows-2022\n job: lint\n profile: debug\n fail-fast: '${{ github.event_name == ''pull_request'' || (github.ref != ''refs/heads/main'' && !startsWith(github.ref, ''refs/tags/'')) }}'\n env:\n CARGO_TERM_COLOR: always\n RUST_BACKTRACE: full\n RUST_LIB_BACKTRACE: 0\n steps:\n - name: Configure git\n run: |-\n git config --global core.symlinks true\n git config --global fetch.parallel 32\n if: '!(matrix.skip)'\n - name: Clone repository\n uses: actions/checkout@v4\n with:\n fetch-depth: 5\n submodules: false\n if: '!(matrix.skip)'\n - name: Clone submodule ./tests/util/std\n run: git submodule update --init --recursive --depth=1 -- ./tests/util/std\n if: '!(matrix.skip)'\n - name: Clone submodule ./tests/wpt/suite\n run: git submodule update --init --recursive --depth=1 -- ./tests/wpt/suite\n if: '!(matrix.skip) && (matrix.wpt)'\n - name: Clone submodule ./tests/node_compat/runner/suite\n run: git submodule update --init --recursive --depth=1 -- ./tests/node_compat/runner/suite\n if: '!(matrix.skip) && (matrix.job == ''lint'' && matrix.os == ''linux'')'\n - name: Clone submodule ./cli/bench/testdata/lsp_benchdata\n run: git submodule update --init --recursive --depth=1 -- ./cli/bench/testdata/lsp_benchdata\n if: '!(matrix.skip) && (matrix.job == ''bench'')'\n - name: 'Create source tarballs (release, linux)'\n if: |-\n !(matrix.skip) && (matrix.os == 'linux' &&\n matrix.profile == 'release' &&\n matrix.job == 'test' &&\n github.repository == 'denoland/deno' &&\n startsWith(github.ref, 'refs/tags/'))\n run: |-\n mkdir -p target/release\n tar --exclude=".git*" --exclude=target --exclude=third_party/prebuilt \\n -czvf target/release/deno_src.tar.gz -C .. deno\n - name: Cache Cargo home\n uses: cirruslabs/cache@v4\n with:\n path: |-\n ~/.cargo/.crates.toml\n ~/.cargo/.crates2.json\n ~/.cargo/bin\n ~/.cargo/registry/index\n ~/.cargo/registry/cache\n ~/.cargo/git/db\n key: '53-cargo-home-${{ matrix.os }}-${{ matrix.arch }}-${{ hashFiles(''Cargo.lock'') }}'\n restore-keys: '53-cargo-home-${{ matrix.os }}-${{ matrix.arch }}-'\n if: '!(matrix.skip)'\n - uses: dsherret/rust-toolchain-file@v1\n if: '!(matrix.skip)'\n - if: '!(matrix.skip) && (matrix.job == ''lint'' || matrix.job == ''test'' || matrix.job == ''bench'')'\n name: Install Deno\n uses: denoland/setup-deno@v2\n with:\n deno-version: v2.x\n - name: Install Python\n uses: actions/setup-python@v5\n with:\n python-version: 3.11\n if: '!(matrix.skip) && (matrix.job != ''lint'' && (matrix.os != ''linux'' || matrix.arch != ''aarch64''))'\n - name: Remove unused versions of Python\n if: '!(matrix.skip) && (matrix.job != ''lint'' && (matrix.os != ''linux'' || matrix.arch != ''aarch64'') && (matrix.os == ''windows''))'\n shell: pwsh\n run: |-\n $env:PATH -split ";" |\n Where-Object { Test-Path "$_\python.exe" } |\n Select-Object -Skip 1 |\n ForEach-Object { Move-Item "$_" "$_.disabled" }\n - if: '!(matrix.skip) && (matrix.job == ''bench'' || matrix.job == ''test'')'\n name: Install Node\n uses: actions/setup-node@v4\n with:\n node-version: 18\n - if: |-\n !(matrix.skip) && (matrix.profile == 'release' &&\n matrix.job == 'test' &&\n github.repository == 'denoland/deno' &&\n (github.ref == 'refs/heads/main' ||\n startsWith(github.ref, 'refs/tags/')))\n name: Authenticate with Google Cloud\n uses: google-github-actions/auth@v2\n with:\n project_id: denoland\n credentials_json: '${{ secrets.GCP_SA_KEY }}'\n export_environment_variables: true\n create_credentials_file: true\n - name: Setup gcloud (unix)\n if: |-\n !(matrix.skip) && (matrix.os != 'windows' &&\n matrix.profile == 'release' &&\n matrix.job == 'test' &&\n github.repository == 'denoland/deno' &&\n (github.ref == 'refs/heads/main' ||\n startsWith(github.ref, 'refs/tags/')))\n uses: google-github-actions/setup-gcloud@v2\n with:\n project_id: denoland\n - name: Setup gcloud (windows)\n if: |-\n !(matrix.skip) && (matrix.os == 'windows' &&\n matrix.profile == 'release' &&\n matrix.job == 'test' &&\n github.repository == 'denoland/deno' &&\n (github.ref == 'refs/heads/main' ||\n startsWith(github.ref, 'refs/tags/')))\n uses: google-github-actions/setup-gcloud@v2\n env:\n CLOUDSDK_PYTHON: '${{env.pythonLocation}}\python.exe'\n with:\n project_id: denoland\n - name: Configure canary build\n if: |-\n !(matrix.skip) && (matrix.job == 'test' &&\n matrix.profile == 'release' &&\n github.repository == 'denoland/deno' &&\n github.ref == 'refs/heads/main')\n run: echo "DENO_CANARY=true" >> $GITHUB_ENV\n - if: '!(matrix.skip) && (matrix.use_sysroot)'\n name: Set up incremental LTO and sysroot build\n run: |-\n # Setting up sysroot\n export DEBIAN_FRONTEND=noninteractive\n # Avoid running man-db triggers, which sometimes takes several minutes\n # to complete.\n sudo apt-get -qq remove --purge -y man-db > /dev/null 2> /dev/null\n # Remove older clang before we install\n sudo apt-get -qq remove 'clang-12*' 'clang-13*' 'clang-14*' 'clang-15*' 'clang-16*' 'clang-17*' 'clang-18*' 'llvm-12*' 'llvm-13*' 'llvm-14*' 'llvm-15*' 'llvm-16*' 'lld-12*' 'lld-13*' 'lld-14*' 'lld-15*' 'lld-16*' 'lld-17*' 'lld-18*' > /dev/null 2> /dev/null\n\n # Install clang-XXX, lld-XXX, and debootstrap.\n echo "deb http://apt.llvm.org/jammy/ llvm-toolchain-jammy-19 main" |\n sudo dd of=/etc/apt/sources.list.d/llvm-toolchain-jammy-19.list\n curl https://apt.llvm.org/llvm-snapshot.gpg.key |\n gpg --dearmor |\n sudo dd of=/etc/apt/trusted.gpg.d/llvm-snapshot.gpg\n sudo apt-get update\n # this was unreliable sometimes, so try again if it fails\n sudo apt-get install --no-install-recommends clang-19 lld-19 clang-tools-19 clang-format-19 clang-tidy-19 || echo 'Failed. Trying again.' && sudo apt-get clean && sudo apt-get update && sudo apt-get install --no-install-recommends clang-19 lld-19 clang-tools-19 clang-format-19 clang-tidy-19\n # Fix alternatives\n (yes '' | sudo update-alternatives --force --all) > /dev/null 2> /dev/null || true\n\n clang-19 -c -o /tmp/memfd_create_shim.o tools/memfd_create_shim.c -fPIC\n\n echo "Decompressing sysroot..."\n wget -q https://github.com/denoland/deno_sysroot_build/releases/download/sysroot-20241030/sysroot-`uname -m`.tar.xz -O /tmp/sysroot.tar.xz\n cd /\n xzcat /tmp/sysroot.tar.xz | sudo tar -x\n sudo mount --rbind /dev /sysroot/dev\n sudo mount --rbind /sys /sysroot/sys\n sudo mount --rbind /home /sysroot/home\n sudo mount -t proc /proc /sysroot/proc\n cd\n\n echo "Done."\n\n # Configure the build environment. Both Rust and Clang will produce\n # llvm bitcode only, so we can use lld's incremental LTO support.\n\n # Load the sysroot's env vars\n echo "sysroot env:"\n cat /sysroot/.env\n . /sysroot/.env\n\n # Important notes:\n # 1. -ldl seems to be required to avoid a failure in FFI tests. This flag seems\n # to be in the Rust default flags in the smoketest, so uncertain why we need\n # to be explicit here.\n # 2. RUSTFLAGS and RUSTDOCFLAGS must be specified, otherwise the doctests fail\n # to build because the object formats are not compatible.\n echo "\n CARGO_PROFILE_BENCH_INCREMENTAL=false\n CARGO_PROFILE_RELEASE_INCREMENTAL=false\n RUSTFLAGS<<__1\n -C linker-plugin-lto=true\n -C linker=clang-19\n -C link-arg=-fuse-ld=lld-19\n -C link-arg=-ldl\n -C link-arg=-Wl,--allow-shlib-undefined\n -C link-arg=-Wl,--thinlto-cache-dir=$(pwd)/target/release/lto-cache\n -C link-arg=-Wl,--thinlto-cache-policy,cache_size_bytes=700m\n -C link-arg=/tmp/memfd_create_shim.o\n --cfg tokio_unstable\n $RUSTFLAGS\n __1\n RUSTDOCFLAGS<<__1\n -C linker-plugin-lto=true\n -C linker=clang-19\n -C link-arg=-fuse-ld=lld-19\n -C link-arg=-ldl\n -C link-arg=-Wl,--allow-shlib-undefined\n -C link-arg=-Wl,--thinlto-cache-dir=$(pwd)/target/release/lto-cache\n -C link-arg=-Wl,--thinlto-cache-policy,cache_size_bytes=700m\n -C link-arg=/tmp/memfd_create_shim.o\n --cfg tokio_unstable\n $RUSTFLAGS\n __1\n CC=/usr/bin/clang-19\n CFLAGS=$CFLAGS\n " > $GITHUB_ENV\n - name: Remove macOS cURL --ipv4 flag\n run: |-\n curl --version\n which curl\n cat /etc/hosts\n rm ~/.curlrc || true\n if: '!(matrix.skip) && (matrix.os == ''macos'')'\n - name: Install macOS aarch64 lld\n run: ./tools/install_prebuilt.js ld64.lld\n if: '!(matrix.skip) && (matrix.os == ''macos'' && matrix.arch == ''aarch64'')'\n - name: Install rust-codesign\n run: |-\n ./tools/install_prebuilt.js rcodesign\n echo $GITHUB_WORKSPACE/third_party/prebuilt/mac >> $GITHUB_PATH\n if: '!(matrix.skip) && (matrix.os == ''macos'')'\n - name: Log versions\n run: |-\n echo '*** Python'\n command -v python && python --version || echo 'No python found or bad executable'\n echo '*** Rust'\n command -v rustc && rustc --version || echo 'No rustc found or bad executable'\n echo '*** Cargo'\n command -v cargo && cargo --version || echo 'No cargo found or bad executable'\n echo '*** Deno'\n command -v deno && deno --version || echo 'No deno found or bad executable'\n echo '*** Node'\n command -v node && node --version || echo 'No node found or bad executable'\n echo '*** Installed packages'\n command -v dpkg && dpkg -l || echo 'No dpkg found or bad executable'\n if: '!(matrix.skip)'\n - name: Install benchmark tools\n if: '!(matrix.skip) && (matrix.job == ''bench'')'\n run: ./tools/install_prebuilt.js wrk hyperfine\n - name: Restore cache build output (PR)\n uses: actions/cache/restore@v4\n if: '!(matrix.skip) && (github.ref != ''refs/heads/main'' && !startsWith(github.ref, ''refs/tags/''))'\n with:\n path: |-\n ./target\n !./target/*/gn_out\n !./target/*/gn_root\n !./target/*/*.zip\n !./target/*/*.tar.gz\n key: never_saved\n restore-keys: '53-cargo-target-${{ matrix.os }}-${{ matrix.arch }}-${{ matrix.profile }}-${{ matrix.job }}-'\n - name: Apply and update mtime cache\n if: '!(matrix.skip) && (!startsWith(github.ref, ''refs/tags/''))'\n uses: ./.github/mtime_cache\n with:\n cache-path: ./target\n - name: Set up playwright cache\n uses: actions/cache@v4\n with:\n path: ./.ms-playwright\n key: 'playwright-${{ runner.os }}-${{ runner.arch }}'\n if: '!(matrix.skip)'\n - name: test_format.js\n if: '!(matrix.skip) && (matrix.job == ''lint'' && matrix.os == ''linux'')'\n run: deno run --allow-write --allow-read --allow-run --allow-net ./tools/format.js --check\n - name: Lint PR title\n if: '!(matrix.skip) && (matrix.job == ''lint'' && github.event_name == ''pull_request'' && matrix.os == ''linux'')'\n env:\n PR_TITLE: '${{ github.event.pull_request.title }}'\n run: deno run ./tools/verify_pr_title.js "$PR_TITLE"\n - name: lint.js\n if: '!(matrix.skip) && (matrix.job == ''lint'')'\n run: deno run --allow-write --allow-read --allow-run --allow-net ./tools/lint.js\n - name: jsdoc_checker.js\n if: '!(matrix.skip) && (matrix.job == ''lint'')'\n run: deno run --allow-read --allow-env --allow-sys ./tools/jsdoc_checker.js\n - name: node_compat/setup.ts --check\n if: '!(matrix.skip) && (matrix.job == ''lint'' && matrix.os == ''linux'')'\n run: deno run --allow-write --allow-read --allow-run=git ./tests/node_compat/runner/setup.ts --check\n - name: Check tracing build\n if: '!(matrix.skip) && (matrix.job == ''test'' && matrix.profile == ''debug'' && matrix.os == ''linux'' && matrix.arch == ''x86_64'')'\n run: cargo check -p deno --features=lsp-tracing\n env:\n CARGO_PROFILE_DEV_DEBUG: 0\n - name: Build debug\n if: '!(matrix.skip) && (matrix.job == ''test'' && matrix.profile == ''debug'')'\n run: cargo build --locked --all-targets --features=panic-trace\n env:\n CARGO_PROFILE_DEV_DEBUG: 0\n - name: Build release\n if: |-\n !(matrix.skip) && ((matrix.job == 'test' || matrix.job == 'bench') &&\n matrix.profile == 'release' && (matrix.use_sysroot ||\n github.repository == 'denoland/deno'))\n run: |-\n df -h\n cargo build --release --locked --all-targets --features=panic-trace\n df -h\n - name: Check deno binary\n if: '!(matrix.skip) && (matrix.job == ''test'')'\n run: 'target/${{ matrix.profile }}/deno eval "console.log(1+2)" | grep 3'\n env:\n NO_COLOR: 1\n - name: Check deno binary (in sysroot)\n if: '!(matrix.skip) && (matrix.job == ''test'' && matrix.use_sysroot)'\n run: 'sudo chroot /sysroot "$(pwd)/target/${{ matrix.profile }}/deno" --version'\n - name: Generate symcache\n if: |-\n !(matrix.skip) && ((matrix.job == 'test' || matrix.job == 'bench') &&\n matrix.profile == 'release' && (matrix.use_sysroot ||\n github.repository == 'denoland/deno'))\n run: |-\n target/release/deno -A tools/release/create_symcache.ts ./deno.symcache\n du -h deno.symcache\n du -h target/release/deno\n env:\n NO_COLOR: 1\n - name: Upload PR artifact (linux)\n if: |-\n !(matrix.skip) && (matrix.job == 'test' &&\n matrix.profile == 'release' && (matrix.use_sysroot ||\n (github.repository == 'denoland/deno' &&\n (github.ref == 'refs/heads/main' ||\n startsWith(github.ref, 'refs/tags/')))))\n uses: actions/upload-artifact@v4\n with:\n name: 'deno-${{ matrix.os }}-${{ matrix.arch }}-${{ github.event.number }}'\n path: target/release/deno\n - name: Pre-release (linux)\n if: |-\n !(matrix.skip) && (matrix.os == 'linux' &&\n (matrix.job == 'test' || matrix.job == 'bench') &&\n matrix.profile == 'release' &&\n github.repository == 'denoland/deno')\n run: |-\n cd target/release\n ./deno -A ../../tools/release/create_symcache.ts deno-${{ matrix.arch }}-unknown-linux-gnu.symcache\n strip ./deno\n zip -r deno-${{ matrix.arch }}-unknown-linux-gnu.zip deno\n shasum -a 256 deno-${{ matrix.arch }}-unknown-linux-gnu.zip > deno-${{ matrix.arch }}-unknown-linux-gnu.zip.sha256sum\n strip ./denort\n zip -r denort-${{ matrix.arch }}-unknown-linux-gnu.zip denort\n shasum -a 256 denort-${{ matrix.arch }}-unknown-linux-gnu.zip > denort-${{ matrix.arch }}-unknown-linux-gnu.zip.sha256sum\n ./deno types > lib.deno.d.ts\n - name: Pre-release (mac)\n if: |-\n !(matrix.skip) && (matrix.os == 'macos' &&\n matrix.job == 'test' &&\n matrix.profile == 'release' &&\n github.repository == 'denoland/deno')\n env:\n APPLE_CODESIGN_KEY: '${{ secrets.APPLE_CODESIGN_KEY }}'\n APPLE_CODESIGN_PASSWORD: '${{ secrets.APPLE_CODESIGN_PASSWORD }}'\n run: |-\n target/release/deno -A tools/release/create_symcache.ts target/release/deno-${{ matrix.arch }}-apple-darwin.symcache\n strip -x -S target/release/deno\n echo "Key is $(echo $APPLE_CODESIGN_KEY | base64 -d | wc -c) bytes"\n rcodesign sign target/release/deno --code-signature-flags=runtime --p12-password="$APPLE_CODESIGN_PASSWORD" --p12-file=<(echo $APPLE_CODESIGN_KEY | base64 -d) --entitlements-xml-file=cli/entitlements.plist\n cd target/release\n zip -r deno-${{ matrix.arch }}-apple-darwin.zip deno\n shasum -a 256 deno-${{ matrix.arch }}-apple-darwin.zip > deno-${{ matrix.arch }}-apple-darwin.zip.sha256sum\n strip -x -S ./denort\n zip -r denort-${{ matrix.arch }}-apple-darwin.zip denort\n shasum -a 256 denort-${{ matrix.arch }}-apple-darwin.zip > denort-${{ matrix.arch }}-apple-darwin.zip.sha256sum\n - name: Pre-release (windows)\n if: |-\n !(matrix.skip) && (matrix.os == 'windows' &&\n matrix.job == 'test' &&\n matrix.profile == 'release' &&\n github.repository == 'denoland/deno')\n shell: pwsh\n run: |-\n Compress-Archive -CompressionLevel Optimal -Force -Path target/release/deno.exe -DestinationPath target/release/deno-${{ matrix.arch }}-pc-windows-msvc.zip\n Get-FileHash target/release/deno-${{ matrix.arch }}-pc-windows-msvc.zip -Algorithm SHA256 | Format-List > target/release/deno-${{ matrix.arch }}-pc-windows-msvc.zip.sha256sum\n Compress-Archive -CompressionLevel Optimal -Force -Path target/release/denort.exe -DestinationPath target/release/denort-${{ matrix.arch }}-pc-windows-msvc.zip\n Get-FileHash target/release/denort-${{ matrix.arch }}-pc-windows-msvc.zip -Algorithm SHA256 | Format-List > target/release/denort-${{ matrix.arch }}-pc-windows-msvc.zip.sha256sum\n target/release/deno.exe -A tools/release/create_symcache.ts target/release/deno-${{ matrix.arch }}-pc-windows-msvc.symcache\n - name: Upload canary to dl.deno.land\n if: |-\n !(matrix.skip) && (matrix.job == 'test' &&\n matrix.profile == 'release' &&\n github.repository == 'denoland/deno' &&\n github.ref == 'refs/heads/main')\n run: |-\n gsutil -h "Cache-Control: public, max-age=3600" cp ./target/release/*.zip gs://dl.deno.land/canary/$(git rev-parse HEAD)/\n gsutil -h "Cache-Control: public, max-age=3600" cp ./target/release/*.sha256sum gs://dl.deno.land/canary/$(git rev-parse HEAD)/\n gsutil -h "Cache-Control: public, max-age=3600" cp ./target/release/*.symcache gs://dl.deno.land/canary/$(git rev-parse HEAD)/\n echo ${{ github.sha }} > canary-latest.txt\n gsutil -h "Cache-Control: no-cache" cp canary-latest.txt gs://dl.deno.land/canary-$(rustc -vV | sed -n "s|host: ||p")-latest.txt\n - name: Autobahn testsuite\n if: |-\n !(matrix.skip) && ((matrix.os == 'linux' && matrix.arch != 'aarch64') &&\n matrix.job == 'test' &&\n matrix.profile == 'release' &&\n !startsWith(github.ref, 'refs/tags/'))\n run: target/release/deno run -A --config tests/config/deno.json ext/websocket/autobahn/fuzzingclient.js\n - name: 'Test (full, debug)'\n if: |-\n !(matrix.skip) && (matrix.job == 'test' &&\n matrix.profile == 'debug' &&\n !startsWith(github.ref, 'refs/tags/') &&\n matrix.os == 'linux')\n run: cargo test --locked --features=panic-trace\n env:\n CARGO_PROFILE_DEV_DEBUG: 0\n - name: 'Test (fast, debug)'\n if: |-\n !(matrix.skip) && (matrix.job == 'test' &&\n matrix.profile == 'debug' &&\n (startsWith(github.ref, 'refs/tags/') || matrix.os != 'linux'))\n run: |-\n cargo test --locked --lib --features=panic-trace\n cargo test --locked --tests --features=panic-trace\n env:\n CARGO_PROFILE_DEV_DEBUG: 0\n - name: Test (release)\n if: |-\n !(matrix.skip) && (matrix.job == 'test' &&\n matrix.profile == 'release' &&\n (matrix.use_sysroot || (\n github.repository == 'denoland/deno' &&\n !startsWith(github.ref, 'refs/tags/'))))\n run: cargo test --release --locked --features=panic-trace\n - name: Configure hosts file for WPT\n if: '!(matrix.skip) && (matrix.wpt)'\n run: ./wpt make-hosts-file | sudo tee -a /etc/hosts\n working-directory: tests/wpt/suite/\n - name: Run web platform tests (debug)\n if: '!(matrix.skip) && (matrix.wpt && matrix.profile == ''debug'')'\n env:\n DENO_BIN: ./target/debug/deno\n run: |-\n deno run -A --lock=tools/deno.lock.json --config tests/config/deno.json\\n ./tests/wpt/wpt.ts setup\n deno run -A --lock=tools/deno.lock.json --config tests/config/deno.json\\n ./tests/wpt/wpt.ts run --quiet --binary="$DENO_BIN"\n - name: Run web platform tests (release)\n if: '!(matrix.skip) && (matrix.wpt && matrix.profile == ''release'')'\n env:\n DENO_BIN: ./target/release/deno\n run: |-\n deno run -A --lock=tools/deno.lock.json --config tests/config/deno.json\\n ./tests/wpt/wpt.ts setup\n deno run -A --lock=tools/deno.lock.json --config tests/config/deno.json\\n ./tests/wpt/wpt.ts run --quiet --release \\n --binary="$DENO_BIN" \\n --json=wpt.json \\n --wptreport=wptreport.json\n - name: Upload wpt results to dl.deno.land\n continue-on-error: true\n if: |-\n !(matrix.skip) && (matrix.wpt &&\n matrix.os == 'linux' &&\n matrix.profile == 'release' &&\n github.repository == 'denoland/deno' &&\n github.ref == 'refs/heads/main' && !startsWith(github.ref, 'refs/tags/'))\n run: |-\n gzip ./wptreport.json\n gsutil -h "Cache-Control: public, max-age=3600" cp ./wpt.json gs://dl.deno.land/wpt/$(git rev-parse HEAD).json\n gsutil -h "Cache-Control: public, max-age=3600" cp ./wptreport.json.gz gs://dl.deno.land/wpt/$(git rev-parse HEAD)-wptreport.json.gz\n echo $(git rev-parse HEAD) > wpt-latest.txt\n gsutil -h "Cache-Control: no-cache" cp wpt-latest.txt gs://dl.deno.land/wpt-latest.txt\n - name: Upload wpt results to wpt.fyi\n continue-on-error: true\n if: |-\n !(matrix.skip) && (matrix.wpt &&\n matrix.os == 'linux' &&\n matrix.profile == 'release' &&\n github.repository == 'denoland/deno' &&\n github.ref == 'refs/heads/main' && !startsWith(github.ref, 'refs/tags/'))\n env:\n WPT_FYI_USER: deno\n WPT_FYI_PW: '${{ secrets.WPT_FYI_PW }}'\n GITHUB_TOKEN: '${{ secrets.DENOBOT_PAT }}'\n run: |-\n ./target/release/deno run --allow-all --lock=tools/deno.lock.json \\n ./tools/upload_wptfyi.js $(git rev-parse HEAD) --ghstatus\n - name: Run benchmarks\n if: '!(matrix.skip) && (matrix.job == ''bench'' && !startsWith(github.ref, ''refs/tags/''))'\n run: cargo bench --locked\n - name: Post Benchmarks\n if: |-\n !(matrix.skip) && (matrix.job == 'bench' &&\n github.repository == 'denoland/deno' &&\n github.ref == 'refs/heads/main' && !startsWith(github.ref, 'refs/tags/'))\n env:\n DENOBOT_PAT: '${{ secrets.DENOBOT_PAT }}'\n run: |-\n git clone --depth 1 --branch gh-pages \\n https://${DENOBOT_PAT}@github.com/denoland/benchmark_data.git \\n gh-pages\n ./target/release/deno run --allow-all ./tools/build_benchmark_jsons.js --release\n cd gh-pages\n git config user.email "propelml@gmail.com"\n git config user.name "denobot"\n git add .\n git commit --message "Update benchmarks"\n git push origin gh-pages\n - name: Build product size info\n if: '!(matrix.skip) && (matrix.job != ''lint'' && matrix.profile != ''debug'' && github.repository == ''denoland/deno'' && (github.ref == ''refs/heads/main'' || startsWith(github.ref, ''refs/tags/'')))'\n run: |-\n du -hd1 "./target/${{ matrix.profile }}"\n du -ha "./target/${{ matrix.profile }}/deno"\n du -ha "./target/${{ matrix.profile }}/denort"\n - name: Worker info\n if: '!(matrix.skip) && (matrix.job == ''bench'')'\n run: |-\n cat /proc/cpuinfo\n cat /proc/meminfo\n - name: Upload release to dl.deno.land (unix)\n if: |-\n !(matrix.skip) && (matrix.os != 'windows' &&\n matrix.job == 'test' &&\n matrix.profile == 'release' &&\n github.repository == 'denoland/deno' &&\n startsWith(github.ref, 'refs/tags/'))\n run: |-\n gsutil -h "Cache-Control: public, max-age=3600" cp ./target/release/*.zip gs://dl.deno.land/release/${GITHUB_REF#refs/*/}/\n gsutil -h "Cache-Control: public, max-age=3600" cp ./target/release/*.sha256sum gs://dl.deno.land/release/${GITHUB_REF#refs/*/}/\n gsutil -h "Cache-Control: public, max-age=3600" cp ./target/release/*.symcache gs://dl.deno.land/release/${GITHUB_REF#refs/*/}/\n - name: Upload release to dl.deno.land (windows)\n if: |-\n !(matrix.skip) && (matrix.os == 'windows' &&\n matrix.job == 'test' &&\n matrix.profile == 'release' &&\n github.repository == 'denoland/deno' &&\n startsWith(github.ref, 'refs/tags/'))\n env:\n CLOUDSDK_PYTHON: '${{env.pythonLocation}}\python.exe'\n run: |-\n gsutil -h "Cache-Control: public, max-age=3600" cp ./target/release/*.zip gs://dl.deno.land/release/${GITHUB_REF#refs/*/}/\n gsutil -h "Cache-Control: public, max-age=3600" cp ./target/release/*.sha256sum gs://dl.deno.land/release/${GITHUB_REF#refs/*/}/\n gsutil -h "Cache-Control: public, max-age=3600" cp ./target/release/*.symcache gs://dl.deno.land/release/${GITHUB_REF#refs/*/}/\n - name: Create release notes\n if: |-\n !(matrix.skip) && (matrix.job == 'test' &&\n matrix.profile == 'release' &&\n github.repository == 'denoland/deno' &&\n startsWith(github.ref, 'refs/tags/'))\n run: |-\n export PATH=$PATH:$(pwd)/target/release\n ./tools/release/05_create_release_notes.ts\n - name: Upload release to GitHub\n uses: softprops/action-gh-release@v0.1.15\n if: |-\n !(matrix.skip) && (matrix.job == 'test' &&\n matrix.profile == 'release' &&\n github.repository == 'denoland/deno' &&\n startsWith(github.ref, 'refs/tags/'))\n env:\n GITHUB_TOKEN: '${{ secrets.GITHUB_TOKEN }}'\n with:\n files: |-\n target/release/deno-x86_64-pc-windows-msvc.zip\n target/release/deno-x86_64-pc-windows-msvc.zip.sha256sum\n target/release/denort-x86_64-pc-windows-msvc.zip\n target/release/denort-x86_64-pc-windows-msvc.zip.sha256sum\n target/release/deno-x86_64-unknown-linux-gnu.zip\n target/release/deno-x86_64-unknown-linux-gnu.zip.sha256sum\n target/release/denort-x86_64-unknown-linux-gnu.zip\n target/release/denort-x86_64-unknown-linux-gnu.zip.sha256sum\n target/release/deno-x86_64-apple-darwin.zip\n target/release/deno-x86_64-apple-darwin.zip.sha256sum\n target/release/denort-x86_64-apple-darwin.zip\n target/release/denort-x86_64-apple-darwin.zip.sha256sum\n target/release/deno-aarch64-unknown-linux-gnu.zip\n target/release/deno-aarch64-unknown-linux-gnu.zip.sha256sum\n target/release/denort-aarch64-unknown-linux-gnu.zip\n target/release/denort-aarch64-unknown-linux-gnu.zip.sha256sum\n target/release/deno-aarch64-apple-darwin.zip\n target/release/deno-aarch64-apple-darwin.zip.sha256sum\n target/release/denort-aarch64-apple-darwin.zip\n target/release/denort-aarch64-apple-darwin.zip.sha256sum\n target/release/deno_src.tar.gz\n target/release/lib.deno.d.ts\n body_path: target/release/release-notes.md\n draft: true\n - name: Save cache build output (main)\n uses: actions/cache/save@v4\n if: '!(matrix.skip) && ((matrix.job == ''test'' || matrix.job == ''lint'') && github.ref == ''refs/heads/main'')'\n with:\n path: |-\n ./target\n !./target/*/gn_out\n !./target/*/gn_root\n !./target/*/*.zip\n !./target/*/*.tar.gz\n key: '53-cargo-target-${{ matrix.os }}-${{ matrix.arch }}-${{ matrix.profile }}-${{ matrix.job }}-${{ github.sha }}'\n wasm:\n name: build wasm32\n needs:\n - pre_build\n if: '${{ needs.pre_build.outputs.skip_build != ''true'' }}'\n runs-on: ubuntu-24.04\n timeout-minutes: 30\n steps:\n - name: Configure git\n run: |-\n git config --global core.symlinks true\n git config --global fetch.parallel 32\n if: '!(matrix.skip)'\n - name: Clone repository\n uses: actions/checkout@v4\n with:\n fetch-depth: 5\n submodules: false\n if: '!(matrix.skip)'\n - uses: dsherret/rust-toolchain-file@v1\n if: '!(matrix.skip)'\n - name: Install wasm target\n run: rustup target add wasm32-unknown-unknown\n if: '!(matrix.skip)'\n - name: Cargo build\n run: cargo build --target wasm32-unknown-unknown -p deno_resolver\n if: '!(matrix.skip)'\n publish-canary:\n name: publish canary\n runs-on: ubuntu-24.04\n needs:\n - build\n if: github.repository == 'denoland/deno' && github.ref == 'refs/heads/main'\n steps:\n - name: Authenticate with Google Cloud\n uses: google-github-actions/auth@v2\n with:\n project_id: denoland\n credentials_json: '${{ secrets.GCP_SA_KEY }}'\n export_environment_variables: true\n create_credentials_file: true\n - name: Setup gcloud\n uses: google-github-actions/setup-gcloud@v2\n with:\n project_id: denoland\n - name: Upload canary version file to dl.deno.land\n run: |-\n echo ${{ github.sha }} > canary-latest.txt\n gsutil -h "Cache-Control: no-cache" cp canary-latest.txt gs://dl.deno.land/canary-latest.txt\n | dataset_sample\yaml\denoland_deno\.github\workflows\ci.yml | ci.yml | YAML | 36,353 | 0.95 | 0.096104 | 0.02231 | vue-tools | 965 | 2023-09-27T02:37:07.515840 | MIT | false | 6693c80530b2770ffabf17cd370e2c72 |
name: npm_publish\n\non:\n workflow_dispatch:\n inputs:\n version:\n description: 'Version'\n type: string\n release:\n types: [published]\n\npermissions:\n id-token: write\n\njobs:\n build:\n name: npm publish\n runs-on: ubuntu-latest\n timeout-minutes: 30\n\n steps:\n - name: Configure git\n run: |\n git config --global core.symlinks true\n git config --global fetch.parallel 32\n\n - name: Clone repository\n uses: actions/checkout@v4\n with:\n submodules: recursive\n\n - name: Install Deno\n uses: denoland/setup-deno@v2\n with:\n deno-version: v2.x\n - name: Install Node\n uses: actions/setup-node@v4\n with:\n node-version: '22.x'\n registry-url: 'https://registry.npmjs.org'\n\n - name: Publish\n env:\n NODE_AUTH_TOKEN: ${{ secrets.NPM_TOKEN }}\n run: ./tools/release/npm/build.ts ${{ github.event.inputs.version }} --publish\n | dataset_sample\yaml\denoland_deno\.github\workflows\npm_publish.yml | npm_publish.yml | YAML | 986 | 0.8 | 0 | 0 | python-kit | 610 | 2025-03-23T18:34:33.335647 | BSD-3-Clause | false | 0d6d28afe318b28a9ec9681bc7bc18a0 |
name: post_publish\n\non:\n release:\n types: [published]\n\njobs:\n update-dl-version:\n name: update dl.deno.land version\n runs-on: ubuntu-latest\n if: github.repository == 'denoland/deno'\n steps:\n - name: Clone repository\n uses: actions/checkout@v4\n - uses: denoland/setup-deno@v2\n with:\n deno-version: v2.x\n - name: Authenticate with Google Cloud\n uses: google-github-actions/auth@v1\n with:\n project_id: denoland\n credentials_json: ${{ secrets.GCP_SA_KEY }}\n export_environment_variables: true\n create_credentials_file: true\n\n - name: Setup gcloud\n uses: google-github-actions/setup-gcloud@v1\n with:\n project_id: denoland\n\n - name: Upload version file to dl.deno.land\n run: |\n echo ${GITHUB_REF#refs/*/} > release-latest.txt\n deno run --allow-net tools/release/version_greater_latest.ts ${GITHUB_REF#refs/*/} && gsutil -h "Cache-Control: no-cache" cp release-latest.txt gs://dl.deno.land/release-latest.txt\n | dataset_sample\yaml\denoland_deno\.github\workflows\post_publish.yml | post_publish.yml | YAML | 1,065 | 0.8 | 0.029412 | 0 | python-kit | 65 | 2024-02-06T10:37:43.274464 | Apache-2.0 | false | cd87f65db7e9d501b766c85c5a55ebe9 |
name: promote_to_release\n\non:\n workflow_dispatch:\n inputs:\n releaseKind:\n description: 'Kind of release'\n type: choice\n options:\n - rc\n - lts\n required: true\n commitHash:\n description: Commit to promote to release\n required: true\n\njobs:\n promote-to-release:\n name: Promote to Release\n runs-on: macOS-latest\n if: github.repository == 'denoland/deno'\n steps:\n - name: Clone repository\n uses: actions/checkout@v4\n with:\n token: ${{ secrets.DENOBOT_PAT }}\n submodules: recursive\n\n - name: Authenticate with Google Cloud\n uses: google-github-actions/auth@v1\n with:\n project_id: denoland\n credentials_json: ${{ secrets.GCP_SA_KEY }}\n export_environment_variables: true\n create_credentials_file: true\n\n - name: Setup gcloud\n uses: google-github-actions/setup-gcloud@v1\n with:\n project_id: denoland\n\n - name: Install deno\n uses: denoland/setup-deno@v2\n with:\n deno-version: v2.x\n\n - name: Install rust-codesign\n run: |-\n ./tools/install_prebuilt.js rcodesign\n echo $GITHUB_WORKSPACE/third_party/prebuilt/mac >> $GITHUB_PATH\n\n - name: Promote to Release\n env:\n APPLE_CODESIGN_KEY: '${{ secrets.APPLE_CODESIGN_KEY }}'\n APPLE_CODESIGN_PASSWORD: '${{ secrets.APPLE_CODESIGN_PASSWORD }}'\n run: |\n deno run -A ./tools/release/promote_to_release.ts ${{github.event.inputs.releaseKind}} ${{github.event.inputs.commitHash}}\n\n - name: Upload archives to dl.deno.land\n run: |\n gsutil -h "Cache-Control: public, max-age=3600" cp ./*.zip gs://dl.deno.land/release/$(cat release-${{github.event.inputs.releaseKind}}-latest.txt)/\n gsutil -h "Cache-Control: no-cache" cp release-${{github.event.inputs.releaseKind}}-latest.txt gs://dl.deno.land/release-${{github.event.inputs.releaseKind}}-latest.txt\n | dataset_sample\yaml\denoland_deno\.github\workflows\promote_to_release.yml | promote_to_release.yml | YAML | 2,018 | 0.95 | 0.016129 | 0 | react-lib | 454 | 2024-08-21T06:44:15.658120 | MIT | false | cd22aca446ed889c2ca4364b42236af4 |
name: start_release\n\non:\n workflow_dispatch:\n inputs:\n releaseKind:\n description: 'Kind of release'\n default: 'patch'\n type: choice\n options:\n - patch\n - minor\n - major\n required: true\n\njobs:\n build:\n name: start release\n runs-on: ubuntu-24.04\n timeout-minutes: 30\n\n env:\n CARGO_TERM_COLOR: always\n RUST_BACKTRACE: full\n RUSTC_FORCE_INCREMENTAL: 1\n\n steps:\n - name: Configure git\n run: |\n git config --global core.symlinks true\n git config --global fetch.parallel 32\n\n - name: Clone repository\n uses: actions/checkout@v4\n\n - name: Install deno\n uses: denoland/setup-deno@v2\n with:\n deno-version: v2.x\n\n - name: Create Gist URL\n env:\n GITHUB_TOKEN: ${{ secrets.DENOBOT_GIST_PAT }}\n GH_WORKFLOW_ACTOR: ${{ github.actor }}\n run: ./tools/release/00_start_release.ts --${{github.event.inputs.releaseKind}}\n | dataset_sample\yaml\denoland_deno\.github\workflows\start_release.yml | start_release.yml | YAML | 1,010 | 0.85 | 0 | 0 | node-utils | 571 | 2025-01-16T01:08:10.436733 | GPL-3.0 | false | e57f16d8153590f198a84c93f85690f1 |
name: version_bump\n\non:\n workflow_dispatch:\n inputs:\n releaseKind:\n description: 'Kind of version bump'\n default: 'patch'\n type: choice\n options:\n - patch\n - minor\n - major\n required: true\n\njobs:\n build:\n name: version bump\n runs-on: ubuntu-24.04\n timeout-minutes: 90\n\n env:\n CARGO_TERM_COLOR: always\n RUST_BACKTRACE: full\n RUSTC_FORCE_INCREMENTAL: 1\n\n steps:\n - name: Configure git\n run: |\n git config --global core.symlinks true\n git config --global fetch.parallel 32\n\n - name: Clone repository\n uses: actions/checkout@v4\n with:\n token: ${{ secrets.DENOBOT_PAT }}\n submodules: recursive\n\n - uses: dsherret/rust-toolchain-file@v1\n\n - name: Install deno\n uses: denoland/setup-deno@v2\n with:\n deno-version: v2.x\n\n - name: Run version bump\n run: |\n git remote add upstream https://github.com/denoland/deno\n ./tools/release/01_bump_crate_versions.ts --${{github.event.inputs.releaseKind}}\n\n - name: Create PR\n env:\n GITHUB_TOKEN: ${{ secrets.DENOBOT_PAT }}\n GH_WORKFLOW_ACTOR: ${{ github.actor }}\n run: |\n git config user.email "${{ github.actor }}@users.noreply.github.com"\n git config user.name "${{ github.actor }}"\n ./tools/release/02_create_pr.ts\n | dataset_sample\yaml\denoland_deno\.github\workflows\version_bump.yml | version_bump.yml | YAML | 1,445 | 0.95 | 0 | 0 | node-utils | 965 | 2024-04-01T18:40:28.392578 | MIT | false | 26a2a628f1d7a5e655f7f462aed7f09e |
# This CI job runs every night and tests all versions of Deno (canary and latest\n# stable) across all OSes we support against the `epochs/daily` branch of WPT.\n\nname: wpt_epoch\n\non:\n schedule:\n # Every night at 0:30 UTC. This is 20 minutes after `epochs/daily` branch is\n # triggered to be created in WPT repo.\n - cron: 30 0 * * *\n workflow_dispatch:\n\njobs:\n wpt:\n name: wpt / ${{ matrix.os }} / ${{ matrix.deno-version }}\n if: github.repository == 'denoland/deno'\n runs-on: ${{ matrix.os }}\n timeout-minutes: 30\n strategy:\n fail-fast: false\n matrix:\n deno-version: [v1.x, canary]\n os: [ubuntu-24.04-xl]\n\n steps:\n - name: Clone repository\n uses: actions/checkout@v4\n with:\n submodules: true\n persist-credentials: false\n\n - name: Setup Deno\n uses: denoland/setup-deno@v2\n with:\n deno-version: ${{ matrix.deno-version }}\n\n - name: Install Python\n uses: actions/setup-python@v5\n with:\n python-version: '3.11'\n\n - name: Log versions\n run: |\n python --version\n deno --version\n\n - name: Switch WPT submodule to epochs/daily\n working-directory: tests/wpt/suite/\n shell: bash\n run: |\n git remote set-branches origin '*'\n git fetch origin\n git checkout $(./wpt rev-list --epoch 1d)\n git checkout -b epochs/daily\n\n - name: Configure hosts file for WPT (unix)\n if: runner.os != 'Windows'\n working-directory: tests/wpt/suite/\n run: ./wpt make-hosts-file | sudo tee -a /etc/hosts\n\n - name: Configure hosts file for WPT (windows)\n if: runner.os == 'Windows'\n working-directory: tests/wpt/suite/\n run: python wpt make-hosts-file | Out-File $env:SystemRoot\System32\drivers\etc\hosts -Encoding ascii -Append\n\n - name: Run web platform tests\n shell: bash\n run: |\n deno run -A --lock=tools/deno.lock.json --config=tests/config/deno.json \\n ./tests/wpt/wpt.ts setup\n deno run -A --lock=tools/deno.lock.json --config=tests/config/deno.json \\n ./tests/wpt/wpt.ts run \ \\n --binary=$(which deno) --quiet --release --no-ignore --json=wpt.json --wptreport=wptreport.json --exit-zero\n\n - name: Upload wpt results to wpt.fyi\n env:\n WPT_FYI_USER: deno\n WPT_FYI_PW: ${{ secrets.WPT_FYI_PW }}\n run: |\n deno run -A --lock=tools/deno.lock.json ./tools/upload_wptfyi.js wptreport.json --from-raw-file --daily-run\n | dataset_sample\yaml\denoland_deno\.github\workflows\wpt_epoch.yml | wpt_epoch.yml | YAML | 2,679 | 0.8 | 0.0625 | 0.058824 | python-kit | 524 | 2023-12-24T17:29:23.930232 | MIT | false | 46cb6aa2f36ea4d0fc2441c2d7a7e145 |
version: "2"\n\nrun:\n allow-parallel-runners: true\n\n # timeout for analysis, e.g. 30s, 5m, default is 1m\n timeout: 5m\n\n # exit code when at least one issue was found, default is 1\n issues-exit-code: 1\n tests: true\n\nlinters:\n enable:\n - sloglint\n - bodyclose\n - copyloopvar\n - depguard\n - errcheck\n - errorlint\n - gocheckcompilerdirectives\n - gocritic\n - godox\n - goprintffuncname\n - gosec\n - govet\n - intrange\n - ineffassign\n - misspell\n - noctx\n - nolintlint\n - revive\n - staticcheck\n - testifylint\n - unconvert\n - unparam\n - unused\n - whitespace\n - gocyclo\n - funlen\n - goconst\n - dogsled\n - lll\n\n settings:\n dogsled:\n max-blank-identifiers: 3\n\n gosec:\n excludes:\n - G109\n - G115\n - G204\n - G303\n\n sloglint:\n no-mixed-args: true\n kv-only: true\n attr-only: false\n no-global: ""\n context: ""\n static-msg: false\n no-raw-keys: true\n key-naming-case: camel\n forbidden-keys:\n - time\n - level\n - msg\n - source\n args-on-sep-lines: false\n\n depguard:\n rules:\n logger:\n deny:\n # logging is allowed only by logutils.Log,\n - pkg: "github.com/sirupsen/logrus"\n desc: logging is allowed only by logutils.Log.\n - pkg: "github.com/pkg/errors"\n desc: Should be replaced by standard lib errors package.\n - pkg: "github.com/instana/testify"\n desc: It's a fork of github.com/stretchr/testify.\n files:\n - "!**/pkg/logutils/**.go"\n\n dupl:\n threshold: 100\n\n funlen:\n lines: -1\n statements: 60\n\n goconst:\n min-len: 2\n min-occurrences: 3\n ignore-string-values:\n - blee\n - duh\n - cl-1\n - ct-1-1\n\n gocritic:\n enabled-tags:\n - diagnostic\n - experimental\n - opinionated\n - performance\n - style\n disabled-checks:\n - dupImport # https://github.com/go-critic/go-critic/issues/845\n - ifElseChain\n - octalLiteral\n - whyNoLint\n\n gocyclo:\n min-complexity: 35\n\n godox:\n keywords:\n - FIXME\n\n mnd:\n checks:\n - argument\n - case\n - condition\n - return\n ignored-numbers:\n - '0'\n - '1'\n - '2'\n - '3'\n ignored-functions:\n - strings.SplitN\n\n govet:\n settings:\n printf:\n funcs:\n - (github.com/golangci/golangci-lint/v2/pkg/logutils.Log).Infof\n - (github.com/golangci/golangci-lint/v2/pkg/logutils.Log).Warnf\n - (github.com/golangci/golangci-lint/v2/pkg/logutils.Log).Errorf\n - (github.com/golangci/golangci-lint/v2/pkg/logutils.Log).Fatalf\n enable:\n - nilness\n - shadow\n\n errorlint:\n asserts: false\n\n lll:\n line-length: 170\n\n misspell:\n locale: US\n ignore-rules:\n - "importas"\n\n nolintlint:\n allow-unused: false\n require-explanation: false\n require-specific: true\n\n revive:\n rules:\n - name: indent-error-flow\n - name: unexported-return\n disabled: true\n - name: unused-parameter\n - name: unused-receiver\n\n exclusions:\n presets:\n - comments\n - std-error-handling\n - common-false-positives\n - legacy\n paths:\n - test/testdata_etc # test files\n - internal/go # extracted from Go code\n - internal/x # extracted from x/tools code\n - pkg/goformatters/gci/internal # extracted from gci code\n - pkg/goanalysis/runner_checker.go # extracted from x/tools code\n rules:\n - path: (.+)_test\.go\n linters:\n - dupl\n - mnd\n - lll\n\n # Based on existing code, the modifications should be limited to make maintenance easier.\n - path: pkg/golinters/unused/unused.go\n linters: [gocritic]\n text: "rangeValCopy: each iteration copies 160 bytes \\(consider pointers or indexing\\)"\n\n # Related to the result of computation but divided multiple times by 1024.\n - path: test/bench/bench_test.go\n linters: [gosec]\n text: "G115: integer overflow conversion uint64 -> int"\n\n # The files created during the tests don't need to be secured.\n - path: scripts/website/expand_templates/linters_test.go\n linters: [gosec]\n text: "G306: Expect WriteFile permissions to be 0600 or less"\n\n # Related to migration command.\n - path: pkg/commands/internal/migrate/two/\n linters:\n - lll\n\n # Related to migration command.\n - path: pkg/commands/internal/migrate/\n linters:\n - gocritic\n text: "hugeParam:"\n\n # The codes are close but this is not duplication.\n - path: pkg/commands/(formatters|linters).go\n linters:\n - dupl\n\nformatters:\n enable:\n - gci\n - gofmt\n - goimports\n settings:\n gofmt:\n rewrite-rules:\n - pattern: 'interface{}'\n replacement: 'any'\n goimports:\n local-prefixes:\n - github.com/golangci/golangci-lint/v2\n exclusions:\n paths:\n - test/testdata_etc # test files\n - internal/go # extracted from Go code\n - internal/x # extracted from x/tools code\n - pkg/goformatters/gci/internal # extracted from gci code\n - pkg/goanalysis/runner_checker.go # extracted from x/tools code\n | dataset_sample\yaml\derailed_k9s\.golangci.yml | .golangci.yml | YAML | 5,501 | 0.95 | 0.004149 | 0.042453 | node-utils | 425 | 2025-04-26T12:56:30.675683 | GPL-3.0 | false | e9332209d42d63619ef6d0625707edae |
version: 2\n\nproject_name: k9s\n\nbefore:\n hooks:\n - go mod download\n - go generate ./...\n\nrelease:\n prerelease: false\n\nenv:\n - CGO_ENABLED=0\n\nbuilds:\n - id: linux\n goos:\n - linux\n - freebsd\n goarch:\n - amd64\n - arm64\n - arm\n - ppc64le\n - s390x\n goarm:\n - 7\n flags:\n - -trimpath\n ldflags:\n - -s -w -X github.com/derailed/k9s/cmd.version=v{{.Version}}\n - -s -w -X github.com/derailed/k9s/cmd.commit={{.Commit}}\n - -s -w -X github.com/derailed/k9s/cmd.date={{.Date}}\n\n - id: osx\n goos:\n - darwin\n goarch:\n - amd64\n - arm64\n flags:\n - -trimpath\n ldflags:\n - -s -w -X github.com/derailed/k9s/cmd.version=v{{.Version}}\n - -s -w -X github.com/derailed/k9s/cmd.commit={{.Commit}}\n - -s -w -X github.com/derailed/k9s/cmd.date={{.Date}}\n\n - id: windows\n goos:\n - windows\n goarch:\n - amd64\n - arm64\n flags:\n - -trimpath\n ldflags:\n - -s -w -X github.com/derailed/k9s/cmd.version=v{{.Version}}\n - -s -w -X github.com/derailed/k9s/cmd.commit={{.Commit}}\n - -s -w -X github.com/derailed/k9s/cmd.date={{.Date}}\n\narchives:\n - name_template: >-\n {{ .ProjectName }}_\n {{- title .Os }}_\n {{- if eq .Arch "amd64" }}amd64\n {{- else if eq .Arch "386" }}i386\n {{- else }}{{ .Arch }}{{ end }}\n {{- if .Arm }}v{{ .Arm }}{{ end }}\n format_overrides:\n - goos: windows\n format: zip\n\nchecksum:\n name_template: "checksums.sha256"\n\nsnapshot:\n name_template: "{{ .Tag }}-next"\n\nchangelog:\n sort: asc\n filters:\n exclude:\n - "^docs:"\n - "^test:"\n\nbrews:\n - name: k9s\n repository:\n owner: derailed\n name: homebrew-k9s\n commit_author:\n name: derailed\n email: fernand@imhotep.io\n directory: Formula\n homepage: https://k9scli.io/\n description: Kubernetes CLI To Manage Your Clusters In Style!\n test: |\n system "k9s version"\n\nnfpms:\n - file_name_template: '{{ .ProjectName }}_{{ .Os }}_{{ .Arch }}'\n maintainer: Fernand Galiana\n homepage: https://k9scli.io\n description: Kubernetes CLI To Manage Your Clusters In Style!\n license: "Apache-2.0"\n formats:\n - deb\n - rpm\n - apk\n bindir: /usr/bin\n section: utils\n contents:\n - src: ./LICENSE\n dst: /usr/share/doc/k9s/copyright\n file_info:\n mode: 0644\n\nsboms:\n - artifacts: archive | dataset_sample\yaml\derailed_k9s\.goreleaser.yml | .goreleaser.yml | YAML | 2,440 | 0.8 | 0.02521 | 0 | react-lib | 769 | 2023-10-24T12:17:56.784356 | BSD-3-Clause | false | 99d3d2c9d23b8d7d6f10c9855b5fe747 |
language: go\ngo_import_path: github.com/derailed/k9s\ngo:\n - "1.23"\n\njobs:\n include:\n - os: linux\n arch: amd64\n - os: linux\n arch: ppc64le\n - os: osx\n arch: amd64\n\ndist: trusty\nsudo: false\n\ninstall: true\n\nscript:\n - go build\n - go test ./...\n | dataset_sample\yaml\derailed_k9s\.travis.yml | .travis.yml | YAML | 257 | 0.85 | 0 | 0 | python-kit | 656 | 2024-07-09T18:26:22.477580 | BSD-3-Clause | false | 617c017bc09741e54970d7a1f25bddcf |
---\nversion: 2\nupdates:\n - package-ecosystem: "gomod"\n directory: "/"\n schedule:\n interval: "weekly"\n\n # Maintain dependencies for GitHub Actions\n - package-ecosystem: "github-actions"\n directory: "/"\n schedule:\n interval: weekly\n\n # Maintain dependencies for docker\n - package-ecosystem: "docker"\n directory: "/"\n schedule:\n interval: weekly\n | dataset_sample\yaml\derailed_k9s\.github\dependabot.yml | dependabot.yml | YAML | 381 | 0.8 | 0.105263 | 0.117647 | node-utils | 457 | 2025-04-03T15:22:27.210181 | GPL-3.0 | false | 8184608d5da9f1be69920f79f1dd1750 |
name: K9s Lint\n\non:\n pull_request:\n branches: [ master ]\n\njobs:\n golangci:\n runs-on: ubuntu-latest\n steps:\n - name: Checkout Code\n uses: actions/checkout@v4.2.2\n\n - name: Install Go\n uses: actions/setup-go@v5.4.0\n with:\n go-version-file: go.mod\n cache-dependency-path: go.sum\n\n - name: Lint\n uses: golangci/golangci-lint-action@v7.0.0\n with:\n github-token: ${{ secrets.GITHUB_TOKEN }}\n version: v2.1.1 | dataset_sample\yaml\derailed_k9s\.github\workflows\lint.yml | lint.yml | YAML | 498 | 0.7 | 0 | 0 | vue-tools | 888 | 2025-01-28T18:50:50.478157 | Apache-2.0 | false | 018f2252bbf677e42320533358b01ea6 |
name: Closeout Stale Issues\non:\n schedule:\n - cron: "0 2 * * *"\n\njobs:\n close-issues:\n runs-on: ubuntu-latest\n permissions:\n issues: write\n pull-requests: write\n steps:\n - uses: actions/stale@v9\n with:\n days-before-issue-stale: 30\n days-before-issue-close: 14\n stale-issue-label: "stale"\n stale-issue-message: "This issue is stale because it has been open for 30 days with no activity."\n close-issue-message: "This issue was closed because it has been inactive for 14 days since being marked as stale."\n repo-token: ${{ secrets.GITHUB_TOKEN }} | dataset_sample\yaml\derailed_k9s\.github\workflows\stales-issues.yml | stales-issues.yml | YAML | 633 | 0.7 | 0.105263 | 0 | vue-tools | 59 | 2023-07-28T14:26:11.776009 | Apache-2.0 | false | 3cbc6e500b6d1a37c26ccfe21752a4cd |
name: Closeout Stale PRs\non:\n schedule:\n - cron: "0 2 * * *"\n\njobs:\n close-issues:\n runs-on: ubuntu-latest\n permissions:\n issues: write\n pull-requests: write\n steps:\n - uses: actions/stale@v9\n with:\n days-before-pr-stale: 30\n days-before-pr-close: 14\n stale-pr-label: "stale"\n stale-pr-message: "This PR is stale because it has been open for 30 days with no activity."\n close-pr-message: "This PR was closed because it has been inactive for 14 days since being marked as stale."\n repo-token: ${{ secrets.GITHUB_TOKEN }} | dataset_sample\yaml\derailed_k9s\.github\workflows\stales-prs.yml | stales-prs.yml | YAML | 609 | 0.7 | 0.105263 | 0 | vue-tools | 243 | 2024-08-10T05:37:15.178087 | BSD-3-Clause | false | 17676ecb0738c8ed80e6e402ad827a07 |
version: v1.0\nname: First pipeline example\nagent:\n machine:\n type: e1-standard-2\n os_image: ubuntu1804\n\nblocks:\n - name: "Build"\n task:\n env_vars:\n - name: APP_ENV\n value: prod\n jobs:\n - name: Docker build\n commands:\n - checkout\n - ls -1\n - echo $APP_ENV\n - echo "Docker build..."\n - echo "done"\n\n - name: "Smoke tests"\n task:\n jobs:\n - name: Smoke\n commands:\n - checkout\n - echo "make smoke"\n\n - name: "Unit tests"\n task:\n jobs:\n - name: RSpec\n commands:\n - checkout\n - echo "make rspec"\n\n - name: Lint code\n commands:\n - checkout\n - echo "make lint"\n\n - name: Check security\n commands:\n - checkout\n - echo "make security"\n\n - name: "Integration tests"\n task:\n jobs:\n - name: Cucumber\n commands:\n - checkout\n - echo "make cucumber"\n\n - name: "Push Image"\n task:\n jobs:\n - name: Push\n commands:\n - checkout\n - echo "make docker.push"\n | dataset_sample\yaml\derailed_k9s\.semaphore\semaphore.yml | semaphore.yml | YAML | 1,144 | 0.7 | 0 | 0 | vue-tools | 288 | 2024-11-14T02:43:13.174583 | Apache-2.0 | false | 191084b04a3e044eb795e03fb4a0a89a |
files:\n - source: /src/**/*.resx\n ignore:\n - /**/*.%locale%.%file_extension%\n - /**/*.%two_letters_code%.%file_extension%\n - /**/DevToys.Tools/**/*.*\n - /**/DevToys.UnitTests/**/*.*\n translation: /src/**/%file_name%.%locale%.%file_extension%\n | dataset_sample\yaml\DevToys-app_DevToys\crowdin.yml | crowdin.yml | YAML | 269 | 0.8 | 0 | 0 | python-kit | 175 | 2025-06-01T18:46:30.806926 | BSD-3-Clause | false | 3fb563fca79896b6988ae26bc1c53e19 |
---\nengines:\n brakeman:\n enabled: true\n bundler-audit:\n enabled: true\n csslint:\n enabled: true\n duplication:\n enabled: true\n config:\n languages:\n - ruby\n - javascript\n eslint:\n enabled: true\n channel: "eslint-2"\n fixme:\n enabled: true\n rubocop:\n enabled: true\nratings:\n paths:\n - Gemfile.lock\n - "**.erb"\n - "**.haml"\n - "**.rb"\n - "**.css"\n - "**.js"\nexclude_paths:\n- config/\n- db/\n- features/\n- script/\n- spec/\n- vendor/\n | dataset_sample\yaml\diaspora_diaspora\.codeclimate.yml | .codeclimate.yml | YAML | 479 | 0.7 | 0 | 0 | python-kit | 362 | 2024-05-10T23:56:54.798801 | BSD-3-Clause | false | bf52365b7eb1318136edf0c1ba63c12c |
require: rubocop-rails\n\nAllCops:\n TargetRubyVersion: 3.0\n NewCops: enable\n Exclude:\n - "bin/**/*"\n - "db/schema.rb"\n - "config/locales/cldr/plurals.rb"\n\n# Disable for rails 4\nRails/HttpPositionalArguments:\n Enabled: false\n\n# Commonly used screens these days easily fit more than 80 characters.\nLayout/LineLength:\n Max: 120\n\n# Too short methods lead to extraction of single-use methods, which can make\n# the code easier to read (by naming things), but can also clutter the class\nMetrics/MethodLength:\n Max: 20\n\n# The guiding principle of classes is SRP, SRP can't be accurately measured by LoC\nMetrics/ClassLength:\n Max: 1500\nMetrics/ModuleLength:\n Max: 1500\n\n# Raise complexity metrics\nMetrics/AbcSize:\n Max: 20\n\nMetrics/CyclomaticComplexity:\n Max: 20\n\nMetrics/PerceivedComplexity:\n Max: 20\n\n# Some blocks are longer.\nMetrics/BlockLength:\n AllowedMethods:\n - "namespace"\n - "create_table"\n Exclude:\n - "config/**/*.rb"\n - "spec/**/*.rb"\n\n# No space makes the method definition shorter and differentiates\n# from a regular assignment.\nLayout/SpaceAroundEqualsInParameterDefault:\n EnforcedStyle: no_space\n\n# Single quotes being faster is hardly measurable and only affects parse time.\n# Enforcing double quotes reduces the times where you need to change them\n# when introducing an interpolation. Use single quotes only if their semantics\n# are needed.\nStyle/StringLiterals:\n EnforcedStyle: double_quotes\n Exclude:\n # These files are generated by rails, so it's best to keep them close to the original for smaller diffs\n - "config/application.rb"\n - "config/boot.rb"\n - "config/environment.rb"\n - "config/environments/*.rb"\n\n# We do not need to support Ruby 1.9, so this is good to use.\nStyle/SymbolArray:\n Enabled: true\n\n# Most readable form.\nLayout/HashAlignment:\n EnforcedHashRocketStyle: table\n EnforcedColonStyle: table\n\n# This rule makes haml files less readable, as there is no 'end' there.\nLayout/CaseIndentation:\n Exclude:\n - "app/views/**/*"\n\n# Mixing the styles looks just silly.\nStyle/HashSyntax:\n EnforcedStyle: ruby19_no_mixed_keys\n\n# has_key? and has_value? are far more readable than key? and value?\nStyle/PreferredHashMethods:\n Enabled: false\n\n# String#% is by far the least verbose and only object oriented variant.\nStyle/FormatString:\n EnforcedStyle: percent\n\nStyle/CollectionMethods:\n Enabled: true\n PreferredMethods:\n # inject seems more common in the community.\n reduce: "inject"\n\n# Either allow this style or don't. Marking it as safe with parenthesis\n# is silly. Let's try to live without them for now.\nStyle/ParenthesesAroundCondition:\n AllowSafeAssignment: false\nLint/AssignmentInCondition:\n AllowSafeAssignment: false\n\n# A specialized exception class will take one or more arguments and construct the message from it.\n# So both variants make sense.\nStyle/RaiseArgs:\n Enabled: false\n\n# Indenting the chained dots beneath each other is not supported by this cop,\n# see https://github.com/bbatsov/rubocop/issues/1633\nLayout/MultilineOperationIndentation:\n Enabled: false\n\n# Fail is an alias of raise. Avoid aliases, it's more cognitive load for no gain.\n# The argument that fail should be used to abort the program is wrong too,\n# there's Kernel#abort for that.\nStyle/SignalException:\n EnforcedStyle: only_raise\n\n# Suppressing exceptions can be perfectly fine, and be it to avoid to\n# explicitly type nil into the rescue since that's what you want to return,\n# or suppressing LoadError for optional dependencies\nLint/SuppressedException:\n Enabled: false\n\nLayout/SpaceInsideBlockBraces:\n # The space here provides no real gain in readability while consuming\n # horizontal space that could be used for a better parameter name.\n # Also {| differentiates better from a hash than { | does.\n SpaceBeforeBlockParameters: false\n\n# No trailing space differentiates better from the block:\n# foo} means hash, foo } means block.\nLayout/SpaceInsideHashLiteralBraces:\n EnforcedStyle: no_space\n\n# { ... } for multi-line blocks is okay, follow Weirichs rule instead:\n# https://web.archive.org/web/20140221124509/http://onestepback.org/index.cgi/Tech/Ruby/BraceVsDoEnd.rdoc\nStyle/BlockDelimiters:\n Enabled: false\n\n# Enforcing -> would be nice, but not at the cost of enforcing lambda { } for\n# multiline lambdas.\nStyle/Lambda:\n Enabled: false\n\n# do / end blocks should be used for side effects,\n# methods that run a block for side effects and have\n# a useful return value are rare, assign the return\n# value to a local variable for those cases.\nStyle/MethodCalledOnDoEndBlock:\n Enabled: true\n\n# Enforcing the names of variables? To single letter ones? Just no.\nStyle/SingleLineBlockParams:\n Enabled: false\n\n# Shadowing outer local variables with block parameters is often useful\n# to not reinvent a new name for the same thing, it highlights the relation\n# between the outer variable and the parameter. The cases where it's actually\n# confusing are rare, and usually bad for other reasons already, for example\n# because the method is too long.\nLint/ShadowingOuterLocalVariable:\n Enabled: false\n\n# Check with yard instead.\nStyle/Documentation:\n Enabled: false\n\n# This is just silly. Calling the argument `other` in all cases makes no sense.\nNaming/BinaryOperatorParameterName:\n Enabled: false\n\n# Defining constants in tests is fine, and it's good to have them close to the test where they are used.\nLint/ConstantDefinitionInBlock:\n Exclude:\n - "spec/**/*"\n\n# There are valid cases, for example debugging Cucumber steps,\n# also they'll fail CI anyway\nLint/Debugger:\n Enabled: false\n\n# We used comparison everywhere.\nStyle/NumericPredicate:\n EnforcedStyle: comparison\n\n# Reset some HoundCI changes back to Rubocop defaults\nLayout/DotPosition:\n EnforcedStyle: leading\n\n# Not enabled by default but good\nStyle/HashEachMethods:\n Enabled: true\n\n# It makes more sense to allow to structure and group them how it makes sense in the code\nStyle/AccessorGrouping:\n Enabled: false\n | dataset_sample\yaml\diaspora_diaspora\.rubocop.yml | .rubocop.yml | YAML | 5,967 | 0.95 | 0.105 | 0.376543 | react-lib | 789 | 2023-11-03T03:34:18.851779 | MIT | false | e36351918cfad4c8a43832e67e631a19 |
scss_files: 'app/assets/stylesheets/**/*.scss'\nplugin_directories: ['.scss-linters']\n\n# List of gem names to load custom linters from (make sure they are already\n# installed)\nplugin_gems: []\n\nlinters:\n BangFormat:\n enabled: true\n space_before_bang: true\n space_after_bang: false\n\n BemDepth:\n enabled: false\n max_elements: 1\n\n BorderZero:\n enabled: true\n convention: zero # or `none`\n\n ColorKeyword:\n enabled: true\n\n ColorVariable:\n enabled: true\n\n Comment:\n enabled: true\n\n DebugStatement:\n enabled: true\n\n DeclarationOrder:\n enabled: true\n\n DisableLinterReason:\n enabled: true\n\n DuplicateProperty:\n enabled: true\n\n ElsePlacement:\n enabled: true\n style: same_line # or 'new_line'\n\n EmptyLineBetweenBlocks:\n enabled: true\n ignore_single_line_blocks: true\n\n EmptyRule:\n enabled: true\n\n ExtendDirective:\n enabled: false\n\n FinalNewline:\n enabled: true\n present: true\n\n HexLength:\n enabled: true\n style: short # or 'long'\n\n HexNotation:\n enabled: true\n style: lowercase # or 'uppercase'\n\n HexValidation:\n enabled: true\n\n IdSelector:\n enabled: false\n\n ImportantRule:\n enabled: true\n\n ImportPath:\n enabled: true\n leading_underscore: false\n filename_extension: false\n\n Indentation:\n enabled: true\n allow_non_nested_indentation: false\n character: space # or 'tab'\n width: 2\n\n LeadingZero:\n enabled: true\n style: exclude_zero # or 'include_zero'\n\n MergeableSelector:\n enabled: true\n force_nesting: true\n\n NameFormat:\n enabled: true\n allow_leading_underscore: true\n convention: hyphenated_lowercase # or 'camel_case', or 'snake_case', or a regex pattern\n\n NestingDepth:\n enabled: true\n max_depth: 3\n ignore_parent_selectors: false\n\n PlaceholderInExtend:\n enabled: true\n\n PropertyCount:\n enabled: false\n include_nested: false\n max_properties: 10\n\n PropertySortOrder:\n enabled: true\n ignore_unspecified: false\n min_properties: 2\n separate_groups: false\n\n PropertySpelling:\n enabled: true\n extra_properties: []\n\n PropertyUnits:\n enabled: true\n global: [\n 'ch', 'em', 'ex', 'rem', # Font-relative lengths\n 'cm', 'in', 'mm', 'pc', 'pt', 'px', 'q', # Absolute lengths\n 'vh', 'vw', 'vmin', 'vmax', # Viewport-percentage lengths\n 'deg', 'grad', 'rad', 'turn', # Angle\n 'ms', 's', # Duration\n 'Hz', 'kHz', # Frequency\n 'dpi', 'dpcm', 'dppx', # Resolution\n '%'] # Other\n properties: {}\n\n QualifyingElement:\n enabled: true\n allow_element_with_attribute: false\n allow_element_with_class: false\n allow_element_with_id: false\n\n SelectorDepth:\n enabled: true\n max_depth: 3\n\n SelectorFormat:\n enabled: true\n convention: hyphenated_lowercase # or 'strict_BEM', or 'hyphenated_BEM', or 'snake_case', or 'camel_case', or a regex pattern\n\n Shorthand:\n enabled: true\n allowed_shorthands: [1, 2, 3]\n\n SingleLinePerProperty:\n enabled: true\n allow_single_line_rule_sets: true\n\n SingleLinePerSelector:\n enabled: true\n\n SpaceAfterComma:\n enabled: true\n\n SpaceAfterPropertyColon:\n enabled: true\n style: one_space # or 'no_space', or 'at_least_one_space', or 'aligned'\n\n SpaceAfterPropertyName:\n enabled: true\n\n SpaceAfterVariableName:\n enabled: true\n\n SpaceAroundOperator:\n enabled: true\n style: one_space # or 'no_space'\n\n SpaceBeforeBrace:\n enabled: true\n style: space # or 'new_line'\n allow_single_line_padding: false\n\n SpaceBetweenParens:\n enabled: true\n spaces: 0\n\n StringQuotes:\n enabled: true\n style: single_quotes # or double_quotes\n\n TrailingSemicolon:\n enabled: true\n\n TrailingWhitespace:\n enabled: true\n\n TrailingZero:\n enabled: false\n\n TransitionAll:\n enabled: false\n\n UnnecessaryMantissa:\n enabled: true\n\n UnnecessaryParentReference:\n enabled: true\n\n UrlFormat:\n enabled: true\n\n UrlQuotes:\n enabled: true\n\n VariableForProperty:\n enabled: false\n properties: []\n\n VendorPrefix:\n enabled: true\n identifier_list: base\n additional_identifiers: []\n excluded_identifiers: []\n\n ZeroUnit:\n enabled: true\n\n Compass::*:\n enabled: false\n | dataset_sample\yaml\diaspora_diaspora\.scss-lint.yml | .scss-lint.yml | YAML | 4,355 | 0.8 | 0 | 0.011561 | vue-tools | 710 | 2024-06-17T00:52:36.820147 | GPL-3.0 | false | a759a353ab1db61433b739854fdf5a93 |
name: CI\non:\n push:\n branches:\n - develop\n - next-minor\n - main\n - master\n pull_request:\n\njobs:\n test:\n runs-on: ubuntu-latest\n strategy:\n fail-fast: false\n matrix:\n ruby:\n - '3.2'\n - '3.3'\n db:\n - mysql\n - postgresql\n kind:\n - cucumber\n - other\n env:\n DB: ${{ matrix.db }}\n RAILS_ENV: test\n BUNDLE_WITH: ${{ matrix.db }}\n BUNDLE_WITHOUT: development\n BUNDLE_FROZEN: true\n BUNDLE_DISABLE_SHARED_GEMS: true\n services:\n postgres:\n image: postgres\n env:\n POSTGRES_PASSWORD: postgres\n options: >-\n --health-cmd pg_isready\n --health-interval 10s\n --health-timeout 5s\n --health-retries 5\n ports:\n - 5432:5432\n steps:\n - name: Install system dependencies\n run: sudo apt update && sudo apt install -y build-essential curl git gsfonts imagemagick libcurl4-openssl-dev libidn11-dev libmagickwand-dev libssl-dev libxml2-dev libxslt1-dev yarnpkg\n - name: Start MySQL\n run: sudo systemctl start mysql.service\n if: matrix.db == 'mysql'\n - uses: actions/checkout@v3\n - uses: ruby/setup-ruby@v1\n with:\n ruby-version: ${{ matrix.ruby }}\n bundler-cache: true\n - name: Prepare\n run: script/ci/prepare.sh\n - name: Run tests\n run: bin/rake --trace ci:${{ matrix.kind }}\n - name: Run Jasmine\n run: bin/rake jasmine:ci\n timeout-minutes: 2\n if: matrix.kind == 'other'\n | dataset_sample\yaml\diaspora_diaspora\.github\workflows\ci.yml | ci.yml | YAML | 1,603 | 0.7 | 0.031746 | 0 | react-lib | 25 | 2024-02-07T04:13:46.604159 | BSD-3-Clause | false | 95926aaaa9983047f12efbb60ed29743 |
## SECURITY WARNING:\n##\n## Do not change this job unless you know what you're doing.\n##\n## This GitHub Action runs on: pull_request_target, which means the jobs run in\n## a context where they have access to a Access Token with write access to the\n## target repo, even if the PR is opened from an external contributor from their\n## fork.\n##\n## This means that if we're not careful, we could be running third-party code\n## within an authenticated scope, which isn't good. To mitigate this, this\n## implementation does:\n##\n## 1. checkout the target branch (i.e. the project's original sources)\n## 2. install the Gems from there, and install them into a directory that's\n## outside the repository contents.\n## 3. checkout the PRs HEAD\n## 4. restore a bunch of files that would allow code execution from the\n## project's upstream sources, namely:\n## - bin/bundle - we'll run that in our Job\n## - Gemfile/Gemfile.lock - to avoid loading a gem with an identical\n## version number from a in-repo vendored directory\n\nname: Lint\non:\n pull_request_target:\n\npermissions:\n contents: read\n statuses: write\n pull-requests: write\n\njobs:\n pronto:\n runs-on: ubuntu-latest\n steps:\n - name: Install system dependencies\n run: sudo apt update && sudo apt install -y build-essential curl git gsfonts imagemagick libcurl4-openssl-dev libidn11-dev libmagickwand-dev libssl-dev libxml2-dev libxslt1-dev yarnpkg\n - name: Checkout Target branch\n uses: actions/checkout@v3\n with:\n ref: ${{ github.base_ref }}\n fetch-depth: 0\n - uses: ruby/setup-ruby@v1\n with:\n ruby-version: "3.1"\n bundler-cache: true\n - name: Checkout PR HEAD\n run: |\n git fetch -q origin +refs/pull/${{ github.event.pull_request.number }}/head:\n git checkout -qf FETCH_HEAD\n - name: Restore the bundle binstub and Gemfiles from the target branch\n run: |\n git restore -s ${{ github.base_ref }} -- bin/bundle\n git restore -s ${{ github.base_ref }} -- Gemfile\n git restore -s ${{ github.base_ref }} -- Gemfile.lock\n - name: Run Pronto\n run: bin/bundle exec pronto run -f github_status github_pr_review -c ${{ github.base_ref }}\n env:\n PRONTO_PULL_REQUEST_ID: ${{ github.event.pull_request.number }}\n PRONTO_GITHUB_ACCESS_TOKEN: ${{ github.token }}\n\n | dataset_sample\yaml\diaspora_diaspora\.github\workflows\lint.yml | lint.yml | YAML | 2,417 | 0.8 | 0.032258 | 0.37931 | react-lib | 876 | 2024-10-11T04:21:55.168028 | BSD-3-Clause | false | d8785364918fe82115cb6fb000f3de33 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.