commit
stringlengths
40
40
old_file
stringlengths
4
237
new_file
stringlengths
4
237
old_contents
stringlengths
1
4.24k
new_contents
stringlengths
1
4.87k
subject
stringlengths
15
778
message
stringlengths
15
8.75k
lang
stringclasses
266 values
license
stringclasses
13 values
repos
stringlengths
5
127k
ca882ab59a4474ca33e780595221b21e8f3bc53a
.kitchen.yml
.kitchen.yml
--- provisioner: name: puppet_apply files_path: files manifests_path: test/integration driver: name: docker platforms: - name: centos-6 driver_config: image: dynatrace/centos-testing:6 - name: debian-7.8 driver_config: image: dynatrace/debian-testing:7.8 - name: ubuntu-12.04 driver_config: image: dynatrace/ubuntu-testing:12.04 suites: - name: apache_wsagent_debian provisioner: manifest: apache_wsagent_debian/init.pp includes: - debian-7.8 - ubuntu-12.04 - name: apache_wsagent_redhat provisioner: manifest: apache_wsagent_redhat/init.pp includes: - centos-6 - name: agents_package provisioner: manifest: agents_package/init.pp - name: collector provisioner: manifest: collector/init.pp - name: java_agent provisioner: manifest: java_agent/init.pp - name: memory_analysis_server provisioner: manifest: memory_analysis_server/init.pp - name: server provisioner: manifest: server/init.pp - name: wsagent_package provisioner: manifest: wsagent_package/init.pp
--- provisioner: name: puppet_apply files_path: files manifests_path: test/integration driver: name: docker platforms: - name: centos-6 driver_config: image: dynatrace/centos-puppet-testing:6 - name: debian-7.8 driver_config: image: dynatrace/debian-puppet-testing:7.8 - name: ubuntu-12.04 driver_config: image: dynatrace/ubuntu-puppet-testing:12.04 suites: - name: apache_wsagent_debian provisioner: manifest: apache_wsagent_debian/init.pp includes: - debian-7.8 - ubuntu-12.04 - name: apache_wsagent_redhat provisioner: manifest: apache_wsagent_redhat/init.pp includes: - centos-6 - name: agents_package provisioner: manifest: agents_package/init.pp - name: collector provisioner: manifest: collector/init.pp - name: java_agent provisioner: manifest: java_agent/init.pp - name: memory_analysis_server provisioner: manifest: memory_analysis_server/init.pp - name: server provisioner: manifest: server/init.pp - name: wsagent_package provisioner: manifest: wsagent_package/init.pp
Make use of 'dynatrace/$os-puppet-testing' Docker images.
Make use of 'dynatrace/$os-puppet-testing' Docker images.
YAML
mit
nespresso/Dynatrace-Puppet,dynaTrace/Dynatrace-Puppet,dynaTrace/Dynatrace-Puppet,nespresso/Dynatrace-Puppet,nespresso/Dynatrace-Puppet,dynaTrace/Dynatrace-Puppet
cb408b717c17fd7c99b3cb93b003ae4ee52fb828
.kitchen.yml
.kitchen.yml
driver: name: vagrant provisioner: name: chef_zero deprecations_as_errors: true verifier: name: inspec platforms: - name: amazonlinux driver_config: box: mvbcoding/awslinux - name: centos-6 - name: centos-7 - name: debian-8 - name: debian-9 - name: fedora-26 - name: freebsd-10 - name: freebsd-11 - name: opensuse-leap-42 - name: ubuntu-14.04 - name: ubuntu-16.04 - name: solaris-11 driver: box: chef/solaris-11 - name: windows-2008r2-standard driver: box: chef/windows-server-2008r2-standard # private box in Chef's Atlas account - name: windows-2012r2 driver_config: box: chef/windows-server-2012r2-standard - name: macosx-10.11 driver: box: chef/macosx-10.11 # private box in Chef's Atlas account provider: vmware_fusion suites: - name: default run_list: test::default - name: compile_time run_list: test::default attributes: build-essential: compile_time: true
driver: name: vagrant provisioner: name: chef_zero deprecations_as_errors: true verifier: name: inspec platforms: - name: amazonlinux driver_config: box: mvbcoding/awslinux - name: centos-6 - name: centos-7 - name: debian-8 - name: debian-9 - name: fedora-27 - name: freebsd-10 - name: freebsd-11 - name: opensuse-leap-42 - name: ubuntu-14.04 - name: ubuntu-16.04 - name: solaris-11 driver: box: chef/solaris-11 - name: windows-2012r2-standard driver: box: chef/windows-server-2012r2-standard # private box in Chef's Atlas account - name: windows-2016 driver_config: box: chef/windows-server-2016-standard # private box in Chef's Atlas account - name: macos-10.12 driver: box: chef/macos-10.12 # private box in Chef's Atlas account provider: vmware_fusion suites: - name: default run_list: test::default - name: compile_time run_list: test::default attributes: build-essential: compile_time: true
Update Fedora, Windows, and MacOS test boxes
Update Fedora, Windows, and MacOS test boxes Signed-off-by: Tim Smith <764ef62106582a09ed09dfa0b6bff7c05fd7d1e4@chef.io>
YAML
apache-2.0
opscode-cookbooks/build-essential,chef-cookbooks/build-essential
76a7917dcfd862abf406432f77865400f8fb9b9d
.kitchen.yml
.kitchen.yml
driver: name: vagrant customize: cpus: 2 memory: 4096 provisioner: name: chef_zero verifier: name: pester platforms: - name: windows-2008r2 driver_config: box: chef/windows-server-2008r2-standard - name: windows-2012r2 driver_config: box: chef/windows-server-2012r2-standard suites: - name: autorun run_list: - recipe[test::autorun] - name: certificate run_list: - recipe[test::certificate] - name: feature run_list: - recipe[test::feature] - name: font run_list: - recipe[test::font] - name: http_acl run_list: - recipe[test::http_acl] - name: package run_list: - recipe[test::package] - name: path run_list: - recipe[test::path] - name: tasks run_list: - recipe[test::tasks]
driver: name: vagrant customize: cpus: 2 memory: 4096 transport: name: winrm elevated: true provisioner: name: chef_zero verifier: name: pester platforms: - name: windows-7 driver_config: box: chef/windows-7-professional - name: windows-8.1 driver_config: box: chef/windows-8.1-professional - name: windows-2008r2 driver_config: box: chef/windows-server-2008r2-standard - name: windows-2012r2 driver_config: box: chef/windows-server-2012r2-standard - name: windows-2012r2-chef-12.1 driver_config: box: chef/windows-server-2012r2-standard provisioner: require_chef_omnibus: 12.1.0 suites: - name: autorun run_list: - recipe[test::autorun] - name: certificate run_list: - recipe[test::certificate] - name: feature run_list: - recipe[test::feature] - name: font run_list: - recipe[test::font] - name: http_acl run_list: - recipe[test::http_acl] - name: package run_list: - recipe[test::package] - name: path run_list: - recipe[test::path] - name: tasks run_list: - recipe[test::tasks]
Test on more platforms / chef releases
Test on more platforms / chef releases Signed-off-by: Tim Smith <764ef62106582a09ed09dfa0b6bff7c05fd7d1e4@chef.io>
YAML
apache-2.0
protec-cookbooks/windows,juliandunn/windows,blairham/windows-cookbook,stuartpreston/windows,daptiv/windows,chef-cookbooks/windows,ruicovelo/windows,criteo-forks/windows,autoclone/windows,opscode-cookbooks/windows,urban-institute/windows,GannettDigital/chef-windows,shoekstra/chef-windows,rlaveycal/windows
01f4152a7aeb927dce940d3ea2e29df0a623baf1
.resinci.yml
.resinci.yml
docker: publish: false
npm: platforms: - name: linux os: ubuntu architecture: x86_64 node_versions: - "8" - "10" - name: linux os: alpine architecture: x86_64 node_versions: - "8" - "10" - name: linux os: ubuntu architecture: x86 node_versions: - "8" - "10" docker: publish: false
Exclude node 6 builds from CI
Exclude node 6 builds from CI Change-type: patch Signed-off-by: Giovanni Garufi <85dd344344db4c4ce024b7c00609e96849fa96e3@balena.io>
YAML
mit
resin-io-modules/resin-components,resin-io-modules/resin-components,resin-io-modules/resin-components
f6612ba12ab794d25020e476e4f56e465df4837f
.rubocop.yml
.rubocop.yml
Encoding: Enabled: false # The default 79 character limit is sometimes unproductive. LineLength: Max: 120 # The default 10 lines limit is unrealistic. MethodLength: Max: 50 # Lining up the attributes of some methods is advantageous for readability SingleSpaceBeforeFirstArg: Enabled: false AllCops: Exclude: - 'Guardfile' - 'Rakefile' - 'vendor/bundle/**/*' FileName: Enabled: false Documentation: Enabled: false AlignParameters: Exclude: - 'metadata.rb'
Encoding: Enabled: false # The default 79 character limit is sometimes unproductive. LineLength: Max: 120 # The default 10 lines limit is unrealistic. MethodLength: Max: 50 # Lining up the attributes of some methods is advantageous for readability SpaceBeforeFirstArg: Enabled: false AllCops: Exclude: - 'Guardfile' - 'Rakefile' - 'vendor/bundle/**/*' FileName: Enabled: false Documentation: Enabled: false AlignParameters: Exclude: - 'metadata.rb'
Update RuboCop config as per warning
Update RuboCop config as per warning
YAML
mit
evertrue/mesos-cookbook,evertrue/mesos-cookbook,evertrue/mesos-cookbook
09a0360ad980e4108d753f94fd3e43c6293ea18d
.rubocop.yml
.rubocop.yml
# Disable all metrics. Metrics/AbcSize: Enabled: false Metrics/BlockNesting: Enabled: false Metrics/ClassLength: Enabled: false Metrics/ModuleLength: Enabled: false Metrics/CyclomaticComplexity: Enabled: false Metrics/LineLength: Enabled: false Metrics/MethodLength: Enabled: false Metrics/ParameterLists: Enabled: false Metrics/PerceivedComplexity: Enabled: false # Prefer compact module/class defs Style/ClassAndModuleChildren: Enabled: false # Allow CamelCase constants as well as SCREAMING_SNAKE_CASE Style/ConstantName: Enabled: false # Allow 'Pokémon-style' exception handling Lint/RescueException: Enabled: false AllCops: TargetRubyVersion: 2.1 # Apparently setting the target version to 2.1 is not enough to prevent this cop... Style/FrozenStringLiteralComment: Enabled: false # http://stackoverflow.com/questions/4763121/should-i-use-alias-or-alias-method Style/Alias: Enabled: false
# Disable all metrics. Metrics/AbcSize: Enabled: false Metrics/BlockNesting: Enabled: false Metrics/ClassLength: Enabled: false Metrics/ModuleLength: Enabled: false Metrics/CyclomaticComplexity: Enabled: false Metrics/LineLength: Enabled: false Metrics/MethodLength: Enabled: false Metrics/ParameterLists: Enabled: false Metrics/PerceivedComplexity: Enabled: false # Prefer compact module/class defs Style/ClassAndModuleChildren: Enabled: false # Allow CamelCase constants as well as SCREAMING_SNAKE_CASE Style/ConstantName: Enabled: false # Allow 'Pokémon-style' exception handling Lint/RescueException: Enabled: false AllCops: TargetRubyVersion: 2.1 # Apparently setting the target version to 2.1 is not enough to prevent this cop... Style/FrozenStringLiteralComment: Enabled: false # http://stackoverflow.com/questions/4763121/should-i-use-alias-or-alias-method Style/Alias: Enabled: false # So RuboCop doesn't complain about application IDs Style/NumericLiterals: Exclude: - examples/*
Make RuboCop ignore numeric literals in examples
Make RuboCop ignore numeric literals in examples
YAML
mit
Roughsketch/discordrb,megumisonoda/discordrb,Roughsketch/discordrb,meew0/discordrb,megumisonoda/discordrb,VxJasonxV/discordrb,VxJasonxV/discordrb,meew0/discordrb
3dcbd78755e3ad85b7fa6b3f53a5e0d813fa65dd
slimmer_headers.yml
slimmer_headers.yml
:section: "Search" :format: "search" :proposition: "citizen" :template: "wrapper_with_fat_footer"
:section: "Search" :format: "search" :proposition: "citizen" :template: "wrapper"
Use standard slimmer template, not fat_footer version.
Use standard slimmer template, not fat_footer version.
YAML
mit
alphagov/rummager,alphagov/rummager,theodi/rummager,theodi/rummager
f9146697ee4b848cf988900fd115fa9fabd3aff2
hiera/data/repositories.yaml
hiera/data/repositories.yaml
# Repository Management openstack_extras::repo::redhat::redhat::repo_hash: 'CentOS-Base': 'descr': 'CentOS-$releasever - Base' 'baseurl': "%{hiera('yum_base_mirror')}/$releasever/os/$basearch/" 'CentOS-Updates': 'descr': 'CentOS-$releasever - Updates' 'baseurl': "%{hiera('yum_base_mirror')}/$releasever/updates/$basearch/" 'CentOS-Extras': 'descr': 'CentOS-$releasever - Extras' 'baseurl': "%{hiera('yum_base_mirror')}/$releasever/extras/$basearch/" 'epel': 'descr': 'Extra Packages for Enterprise Linux 6 - $basearch' 'baseurl': "%{hiera('yum_epel_mirror')}/$releasever/$basearch/" 'gpgkey': 'file:///etc/pki/rpm-gpg/RPM-GPG-KEY-EPEL-6' 'failovermethod': 'priority' openstack_extras::repo::redhat::redhat::gpgkey_hash: '/etc/pki/rpm-gpg/RPM-GPG-KEY-EPEL-6': 'source': 'puppet:///modules/openstack_extras/RPM-GPG-KEY-EPEL-6' openstack_extras::repo::redhat::redhat::purge_unmanaged: true openstack_extras::repo::redhat::redhat::package_require: true openstack_extras::repo::redhat::redhat::manage_epel: false
# Repository Management openstack_extras::repo::redhat::redhat::repo_hash: 'CentOS-Base': 'descr': 'CentOS-$releasever - Base' 'baseurl': "%{hiera('yum_base_mirror')}/$releasever/os/$basearch/" 'CentOS-Updates': 'descr': 'CentOS-$releasever - Updates' 'baseurl': "%{hiera('yum_base_mirror')}/$releasever/updates/$basearch/" 'CentOS-Extras': 'descr': 'CentOS-$releasever - Extras' 'baseurl': "%{hiera('yum_base_mirror')}/$releasever/extras/$basearch/" 'epel': 'descr': "Extra Packages for Enterprise Linux %{operatingsystemmajrelease} - $basearch" 'baseurl': "%{hiera('yum_epel_mirror')}/$releasever/$basearch/" 'gpgkey': 'file:///etc/pki/rpm-gpg/RPM-GPG-KEY-EPEL-7' 'failovermethod': 'priority' openstack_extras::repo::redhat::redhat::gpgkey_hash: '/etc/pki/rpm-gpg/RPM-GPG-KEY-EPEL-7': 'source': 'puppet:///modules/openstack_extras/RPM-GPG-KEY-EPEL-7' openstack_extras::repo::redhat::redhat::purge_unmanaged: true openstack_extras::repo::redhat::redhat::package_require: true openstack_extras::repo::redhat::redhat::manage_epel: false
Move to centos 7 from 6
Move to centos 7 from 6
YAML
apache-2.0
michaeltchapman/puppet_openstack_builder,michaeltchapman/puppet_openstack_builder,phchoic/puppet_openstack_builder,phchoic/puppet_openstack_builder,michaeltchapman/vagrant-consul,michaeltchapman/puppet_openstack_builder,michaeltchapman/vagrant-consul,phchoic/puppet_openstack_builder
0209bdc69a55a104e52b9d21356b20b95268dd84
.vsts-ci.yml
.vsts-ci.yml
variables: deployment: deploy nuget: packages steps: - task: NuGetCommand@2 inputs: command: 'restore' packDestination: $(nuget) projects: WhiteLibrary - script: | pip install -r requirements.txt displayName: Install needed compoments from PyPi - task: VSBuild@1 inputs: solution: UIAutomationTest\UIAutomationTest.csproj - task: Bash@3 inputs: targetType: inline script: echo $BUILD_BUILDNUMBER >> build.info - script: | set DEPLOYMENT=$(deployment) deploy.cmd continueOnError: false displayName: Copy build artifacts to deployment directory - script: | set DEPLOYMENT=$(deployment) run_tests.cmd --exclude no_ci --xunit xunit.xml continueOnError: true displayName: Execute Acceptance tests - task: PublishTestResults@2 inputs: testRunner: JUnit testResultsFiles: output/xunit.xml - task: PublishBuildArtifacts@1 displayName: Publish Robot Framework logs and results inputs: PathtoPublish: output ArtifactType: container TargetPath: "" ArtifactName: Robot Test results - task: PublishBuildArtifacts@1 displayName: Publish WhiteLibrary inputs: PathToPublish: $(deployment) ArtifactType: container TargetPath: "" ArtifactName: WhiteLibrary
variables: deployment: deploy nuget: packages steps: - task: NuGetCommand@2 inputs: command: 'restore' packDestination: $(nuget) projects: WhiteLibrary - script: | pip install -r requirements.txt displayName: Install needed compoments from PyPi - task: VSBuild@1 inputs: solution: UIAutomationTest\UIAutomationTest.csproj - task: Bash@3 inputs: targetType: inline script: echo $BUILD_BUILDNUMBER >> build.info displayName: Generate build.info - script: | set DEPLOYMENT=$(deployment) deploy.cmd continueOnError: false displayName: Copy build artifacts to deployment directory - script: | set DEPLOYMENT=$(deployment) run_tests.cmd --exclude no_ci --xunit xunit.xml continueOnError: true displayName: Execute Acceptance tests - task: PublishTestResults@2 inputs: testRunner: JUnit testResultsFiles: output/xunit.xml - task: PublishBuildArtifacts@1 displayName: Publish Robot Framework logs and results inputs: PathtoPublish: output ArtifactType: container TargetPath: "" ArtifactName: Robot Test results - task: PublishBuildArtifacts@1 displayName: Publish WhiteLibrary inputs: PathToPublish: $(deployment) ArtifactType: container TargetPath: "" ArtifactName: WhiteLibrary
Add display name to bash step
Add display name to bash step
YAML
apache-2.0
Omenia/robotframework-whitelibrary,Omenia/robotframework-whitelibrary
bde1ec3120116ff5a0f1299baeea751b08eb3530
recipes/umi_tools/meta.yaml
recipes/umi_tools/meta.yaml
package: name: "umi_tools" version: "0.2.6" source: fn: v0.2.6.tar.gz url: https://github.com/CGATOxford/UMI-tools/archive/v0.2.6.tar.gz build: number: 0 requirements: build: - python - setuptools - cython - numpy - pandas - pysam - future run: - python - cython - numpy - pandas - pysam - future test: commands: - umi_tools --help about: home: https://github.com/CGATOxford/UMI-tools license: MIT License summary: 'Tools for dealing with Unique Molecular Identifiers'
package: name: "umi_tools" version: "0.2.3" source: fn: v0.2.3.tar.gz url: https://github.com/CGATOxford/UMI-tools/archive/v0.2.3.tar.gz md5: ebf34023a9fedb0ce87a0b74e3e655a8 build: number: 2 skip: True # [py2k] requirements: build: - python - setuptools >=1.1 - cython >=0.19 - numpy >=1.7 - pandas >=0.12.0 - pysam >=0.8.4 - future run: - python - cython >=0.19 - numpy >=1.7 - pandas >=0.12.0 - pysam >=0.8.4 - future test: commands: - umi_tools --help about: home: https://github.com/CGATOxford/UMI-tools license: MIT License summary: 'Tools for dealing with Unique Molecular Identifiers'
Revert " bumped UMI-tools version"
Revert " bumped UMI-tools version" This reverts commit 399eef73392d14ec4290627b64e3e3aef2873d39.
YAML
mit
blankenberg/bioconda-recipes,dkoppstein/recipes,dmaticzka/bioconda-recipes,jasper1918/bioconda-recipes,joachimwolff/bioconda-recipes,chapmanb/bioconda-recipes,pinguinkiste/bioconda-recipes,jfallmann/bioconda-recipes,peterjc/bioconda-recipes,JenCabral/bioconda-recipes,rvalieris/bioconda-recipes,chapmanb/bioconda-recipes,lpantano/recipes,shenwei356/bioconda-recipes,CGATOxford/bioconda-recipes,bioconda/recipes,rob-p/bioconda-recipes,gregvonkuster/bioconda-recipes,joachimwolff/bioconda-recipes,JenCabral/bioconda-recipes,ThomasWollmann/bioconda-recipes,xguse/bioconda-recipes,matthdsm/bioconda-recipes,rvalieris/bioconda-recipes,omicsnut/bioconda-recipes,keuv-grvl/bioconda-recipes,gvlproject/bioconda-recipes,dmaticzka/bioconda-recipes,mcornwell1957/bioconda-recipes,Luobiny/bioconda-recipes,cokelaer/bioconda-recipes,bioconda/bioconda-recipes,instituteofpathologyheidelberg/bioconda-recipes,phac-nml/bioconda-recipes,acaprez/recipes,joachimwolff/bioconda-recipes,npavlovikj/bioconda-recipes,joachimwolff/bioconda-recipes,hardingnj/bioconda-recipes,rob-p/bioconda-recipes,chapmanb/bioconda-recipes,dkoppstein/recipes,instituteofpathologyheidelberg/bioconda-recipes,npavlovikj/bioconda-recipes,bow/bioconda-recipes,HassanAmr/bioconda-recipes,dmaticzka/bioconda-recipes,ostrokach/bioconda-recipes,blankenberg/bioconda-recipes,guowei-he/bioconda-recipes,ThomasWollmann/bioconda-recipes,bioconda/bioconda-recipes,ostrokach/bioconda-recipes,zachcp/bioconda-recipes,JenCabral/bioconda-recipes,bow/bioconda-recipes,peterjc/bioconda-recipes,gvlproject/bioconda-recipes,CGATOxford/bioconda-recipes,cokelaer/bioconda-recipes,pinguinkiste/bioconda-recipes,instituteofpathologyheidelberg/bioconda-recipes,oena/bioconda-recipes,bebatut/bioconda-recipes,bebatut/bioconda-recipes,jasper1918/bioconda-recipes,oena/bioconda-recipes,ostrokach/bioconda-recipes,abims-sbr/bioconda-recipes,lpantano/recipes,martin-mann/bioconda-recipes,mdehollander/bioconda-recipes,Luobiny/bioconda-recipes,joachimwolff/bioconda-recipes,matthdsm/bioconda-recipes,saketkc/bioconda-recipes,HassanAmr/bioconda-recipes,daler/bioconda-recipes,rvalieris/bioconda-recipes,phac-nml/bioconda-recipes,colinbrislawn/bioconda-recipes,bioconda/recipes,saketkc/bioconda-recipes,dkoppstein/recipes,JenCabral/bioconda-recipes,abims-sbr/bioconda-recipes,phac-nml/bioconda-recipes,ostrokach/bioconda-recipes,bebatut/bioconda-recipes,ivirshup/bioconda-recipes,zwanli/bioconda-recipes,gregvonkuster/bioconda-recipes,daler/bioconda-recipes,colinbrislawn/bioconda-recipes,JenCabral/bioconda-recipes,gvlproject/bioconda-recipes,gvlproject/bioconda-recipes,hardingnj/bioconda-recipes,bow/bioconda-recipes,ivirshup/bioconda-recipes,keuv-grvl/bioconda-recipes,mdehollander/bioconda-recipes,zwanli/bioconda-recipes,bow/bioconda-recipes,rvalieris/bioconda-recipes,npavlovikj/bioconda-recipes,ThomasWollmann/bioconda-recipes,saketkc/bioconda-recipes,bow/bioconda-recipes,zachcp/bioconda-recipes,ThomasWollmann/bioconda-recipes,hardingnj/bioconda-recipes,HassanAmr/bioconda-recipes,CGATOxford/bioconda-recipes,matthdsm/bioconda-recipes,dmaticzka/bioconda-recipes,BIMSBbioinfo/bioconda-recipes,omicsnut/bioconda-recipes,roryk/recipes,gvlproject/bioconda-recipes,mcornwell1957/bioconda-recipes,xguse/bioconda-recipes,ostrokach/bioconda-recipes,abims-sbr/bioconda-recipes,blankenberg/bioconda-recipes,colinbrislawn/bioconda-recipes,daler/bioconda-recipes,jfallmann/bioconda-recipes,guowei-he/bioconda-recipes,phac-nml/bioconda-recipes,BIMSBbioinfo/bioconda-recipes,bow/bioconda-recipes,shenwei356/bioconda-recipes,zwanli/bioconda-recipes,zwanli/bioconda-recipes,cokelaer/bioconda-recipes,martin-mann/bioconda-recipes,oena/bioconda-recipes,pinguinkiste/bioconda-recipes,jasper1918/bioconda-recipes,rvalieris/bioconda-recipes,guowei-he/bioconda-recipes,pinguinkiste/bioconda-recipes,jasper1918/bioconda-recipes,saketkc/bioconda-recipes,keuv-grvl/bioconda-recipes,blankenberg/bioconda-recipes,dmaticzka/bioconda-recipes,CGATOxford/bioconda-recipes,BIMSBbioinfo/bioconda-recipes,Luobiny/bioconda-recipes,roryk/recipes,daler/bioconda-recipes,omicsnut/bioconda-recipes,guowei-he/bioconda-recipes,instituteofpathologyheidelberg/bioconda-recipes,mdehollander/bioconda-recipes,ivirshup/bioconda-recipes,mcornwell1957/bioconda-recipes,colinbrislawn/bioconda-recipes,xguse/bioconda-recipes,acaprez/recipes,jfallmann/bioconda-recipes,gregvonkuster/bioconda-recipes,keuv-grvl/bioconda-recipes,mdehollander/bioconda-recipes,lpantano/recipes,HassanAmr/bioconda-recipes,colinbrislawn/bioconda-recipes,saketkc/bioconda-recipes,rob-p/bioconda-recipes,daler/bioconda-recipes,BIMSBbioinfo/bioconda-recipes,CGATOxford/bioconda-recipes,peterjc/bioconda-recipes,oena/bioconda-recipes,abims-sbr/bioconda-recipes,acaprez/recipes,martin-mann/bioconda-recipes,omicsnut/bioconda-recipes,lpantano/recipes,cokelaer/bioconda-recipes,joachimwolff/bioconda-recipes,shenwei356/bioconda-recipes,mcornwell1957/bioconda-recipes,ThomasWollmann/bioconda-recipes,pinguinkiste/bioconda-recipes,ostrokach/bioconda-recipes,gvlproject/bioconda-recipes,keuv-grvl/bioconda-recipes,peterjc/bioconda-recipes,mdehollander/bioconda-recipes,rob-p/bioconda-recipes,mcornwell1957/bioconda-recipes,JenCabral/bioconda-recipes,dmaticzka/bioconda-recipes,BIMSBbioinfo/bioconda-recipes,bioconda/bioconda-recipes,HassanAmr/bioconda-recipes,zwanli/bioconda-recipes,hardingnj/bioconda-recipes,matthdsm/bioconda-recipes,abims-sbr/bioconda-recipes,pinguinkiste/bioconda-recipes,instituteofpathologyheidelberg/bioconda-recipes,bioconda/recipes,xguse/bioconda-recipes,shenwei356/bioconda-recipes,npavlovikj/bioconda-recipes,phac-nml/bioconda-recipes,chapmanb/bioconda-recipes,mdehollander/bioconda-recipes,hardingnj/bioconda-recipes,matthdsm/bioconda-recipes,martin-mann/bioconda-recipes,colinbrislawn/bioconda-recipes,HassanAmr/bioconda-recipes,martin-mann/bioconda-recipes,roryk/recipes,rvalieris/bioconda-recipes,ThomasWollmann/bioconda-recipes,chapmanb/bioconda-recipes,jasper1918/bioconda-recipes,ivirshup/bioconda-recipes,daler/bioconda-recipes,matthdsm/bioconda-recipes,ivirshup/bioconda-recipes,zachcp/bioconda-recipes,oena/bioconda-recipes,Luobiny/bioconda-recipes,keuv-grvl/bioconda-recipes,gregvonkuster/bioconda-recipes,BIMSBbioinfo/bioconda-recipes,jfallmann/bioconda-recipes,peterjc/bioconda-recipes,zwanli/bioconda-recipes,xguse/bioconda-recipes,zachcp/bioconda-recipes,peterjc/bioconda-recipes,omicsnut/bioconda-recipes,bioconda/bioconda-recipes,bebatut/bioconda-recipes,acaprez/recipes,abims-sbr/bioconda-recipes,CGATOxford/bioconda-recipes,saketkc/bioconda-recipes,guowei-he/bioconda-recipes,ivirshup/bioconda-recipes
e55b89096f9f30c5a80f1fb8252c7239d236d24a
.github/FUNDING.yml
.github/FUNDING.yml
# These are supported funding model platforms github: # Replace with up to 4 GitHub Sponsors-enabled usernames e.g., [user1, user2] patreon: # Replace with a single Patreon username open_collective: # Replace with a single Open Collective username ko_fi: # Replace with a single Ko-fi username tidelift: # Replace with a single Tidelift platform-name/package-name e.g., npm/babel community_bridge: # Replace with a single Community Bridge project-name e.g., cloud-foundry liberapay: # Replace with a single Liberapay username issuehunt: # Replace with a single IssueHunt username otechie: # Replace with a single Otechie username custom: ['https://nlnetlabs.nl/funding/']
github: [NLnetLabs] custom: ['https://nlnetlabs.nl/funding/']
Add GitHub Sponsors for Organisations
Add GitHub Sponsors for Organisations
YAML
bsd-3-clause
NLnetLabs/unbound,NLnetLabs/unbound,NLnetLabs/unbound,NLnetLabs/unbound
fff54c506096fa54f8a442f1c139978039e7adce
.kitchen.travis.yml
.kitchen.travis.yml
--- driver: name: ec2 aws_ssh_key_id: travis ssh_key: travis.pem
--- driver: name: ec2 aws_ssh_key_id: travis ssh_key: travis.pem platforms: - name: ubuntu-12.04 driver_config: image_id: ami-4df9cf24
Use an instance store AMI.
Use an instance store AMI.
YAML
apache-2.0
balanced-cookbooks/balanced-omnibus
479f95741ea222c6f659bb819e4755cc6551a73e
roles/common/tasks/main.yml
roles/common/tasks/main.yml
--- - name: update apt index apt: update_cache=yes - name: python dependencies apt: pkg={{ item }} with_items: - python-pip - python-libxml2 - python-lxml - python-greenlet - python-openssl - python-dev - python-httplib2 - python-software-properties - python-virtualenv - python-mysqldb - apt: pkg={{ item }} with_items: - ntp - vlan - name: set UTC timezone copy: content='Etc/UTC' dest=/etc/timezone owner=root group=root mode=0644 notify: - update timezone - name: /opt/stack file: dest=/opt/stack state=directory - include: ssl.yml - include: ssh.yml - include: networking.yml - include: system_tools.yml # run this last so we only update if run was successful - name: drop an motd with ursula metadata action: template src=etc/motd.tail dest=/etc/motd.tail mode=0644
--- - name: update apt index apt: update_cache=yes - name: python dependencies apt: pkg={{ item }} with_items: - python-pip - python-libxml2 - python-lxml - python-greenlet - python-openssl - python-dev - python-httplib2 - python-software-properties - python-virtualenv - python-mysqldb - apt: pkg={{ item }} with_items: - cryptsetup - ntp - vlan - name: set UTC timezone copy: content='Etc/UTC' dest=/etc/timezone owner=root group=root mode=0644 notify: - update timezone - name: /opt/stack file: dest=/opt/stack state=directory - include: ssl.yml - include: ssh.yml - include: networking.yml - include: system_tools.yml # run this last so we only update if run was successful - name: drop an motd with ursula metadata action: template src=etc/motd.tail dest=/etc/motd.tail mode=0644
Add cryptsetup so that we can map encrypted vols
Add cryptsetup so that we can map encrypted vols Adding cryptsetup so that we can easily add encrypted volumes and images.
YAML
mit
davidcusatis/ursula,knandya/ursula,knandya/ursula,dlundquist/ursula,zrs233/ursula,channus/ursula,EricCrosson/ursula,greghaynes/ursula,fancyhe/ursula,MaheshIBM/ursula,MaheshIBM/ursula,nirajdp76/ursula,masteinhauser/ursula,jwaibel/ursula,lihkin213/ursula,blueboxjesse/ursula,davidcusatis/ursula,davidcusatis/ursula,narengan/ursula,aldevigi/ursula,persistent-ursula/ursula,channus/ursula,wupeiran/ursula,blueboxgroup/ursula,pbannister/ursula,pgraziano/ursula,greghaynes/ursula,paulczar/ursula,nirajdp76/ursula,edtubillara/ursula,msambol/ursula,j2sol/ursula,jwaibel/ursula,ryshah/ursula,ddaskal/ursula,rongzhus/ursula,edtubillara/ursula,j2sol/ursula,fancyhe/ursula,masteinhauser/ursula,persistent-ursula/ursula,ryshah/ursula,aldevigi/ursula,dlundquist/ursula,ryshah/ursula,j2sol/ursula,zrs233/ursula,channus/ursula,edtubillara/ursula,masteinhauser/ursula,ddaskal/ursula,narengan/ursula,panxia6679/ursula,sivakom/ursula,jwaibel/ursula,edtubillara/ursula,EricCrosson/ursula,mjbrewer/ursula,wupeiran/ursula,mjbrewer/ursula,twaldrop/ursula,msambol/ursula,pgraziano/ursula,msambol/ursula,ddaskal/ursula,wupeiran/ursula,andrewrothstein/ursula,rongzhus/ursula,EricCrosson/ursula,zrs233/ursula,kennjason/ursula,lihkin213/ursula,fancyhe/ursula,mjbrewer/ursula,allomov/ursula,panxia6679/ursula,andrewrothstein/ursula,kennjason/ursula,knandya/ursula,sivakom/ursula,andrewrothstein/ursula,rongzhus/ursula,blueboxjesse/ursula,ryshah/ursula,blueboxjesse/ursula,sivakom/ursula,pbannister/ursula,rongzhus/ursula,lihkin213/ursula,dlundquist/ursula,twaldrop/ursula,allomov/ursula,nirajdp76/ursula,nirajdp76/ursula,paulczar/ursula,paulczar/ursula,MaheshIBM/ursula,pgraziano/ursula,wupeiran/ursula,kennjason/ursula,panxia6679/ursula,persistent-ursula/ursula,narengan/ursula,lihkin213/ursula,zrs233/ursula,blueboxgroup/ursula,channus/ursula,blueboxgroup/ursula,j2sol/ursula,narengan/ursula,pgraziano/ursula,panxia6679/ursula,greghaynes/ursula,aldevigi/ursula,fancyhe/ursula,masteinhauser/ursula,twaldrop/ursula,knandya/ursula,twaldrop/ursula,blueboxgroup/ursula,pbannister/ursula,persistent-ursula/ursula,blueboxjesse/ursula,allomov/ursula,ddaskal/ursula
87d8e5fcef0a15599448042032d4ddba578223c4
roles/packer/tasks/main.yml
roles/packer/tasks/main.yml
--- - name: Download Packer sudo: yes get_url: url={{ packer_url }} dest={{ packer_install_dir }} - name: Extract Packer sudo: yes unarchive: copy=no src={{ packer_install_dir }}/{{ packer_archive }}.zip dest={{ packer_install_dir }} - name: Remove downloaded packer_archive sudo: yes shell: "rm -f {{ packer_install_dir }}/{{ packer_archive }}.zip"
--- - name: Download Packer sudo: yes shell: curl -o {{ packer_install_dir }}/{{ packer_archive }}.zip {{ packer_url }} - name: Extract Packer sudo: yes unarchive: copy=no src={{ packer_install_dir }}/{{ packer_archive }}.zip dest={{ packer_install_dir }} - name: Remove downloaded packer_archive sudo: yes shell: "rm -f {{ packer_install_dir }}/{{ packer_archive }}.zip"
Remove get_url in favour of curl
Remove get_url in favour of curl
YAML
apache-2.0
snowplow/ansible-playbooks
260274f82cea6856664d0adce151c221b78c2385
buildscripts/condarecipe.hsa/meta.yaml
buildscripts/condarecipe.hsa/meta.yaml
package: name: numba-hsa version: "0.20.0" source: path: ../.. build: number: {{ environ.get('GIT_DESCRIBE_NUMBER', 0) }} entry_points: - pycc = numba.pycc:main - numba = numba.numba_entry:main requirements: # build and run dependencies are duplicated to avoid setuptools issues # when we also set install_requires in setup.py build: - python - argparse [py26] - numpy # On channel https://binstar.org/numba/ - llvmlite 0.6* - funcsigs [py26 or py27] - singledispatch [py26 or py27 or py33] run: - python - argparse [py26] - numpy # On channel https://binstar.org/numba/ - llvmlite 0.6* - funcsigs [py26 or py27] - singledispatch [py26 or py27 or py33] - libhlc test: requires: - argparse [py26] - unittest2 [py26] - jinja2 files: - mandel.py commands: - pycc -h - numba -h
package: name: numba-hsa version: 0.21.0 source: path: ../.. build: number: 0 entry_points: - pycc = numba.pycc:main - numba = numba.numba_entry:main requirements: # build and run dependencies are duplicated to avoid setuptools issues # when we also set install_requires in setup.py build: - python - argparse [py26] - numpy x.x # On channel https://binstar.org/numba/ - llvmlite 0.8* - funcsigs [py26 or py27] - singledispatch [py26 or py27 or py33] run: - python - argparse [py26] - numpy x.x # On channel https://binstar.org/numba/ - llvmlite 0.8* - funcsigs [py26 or py27] - singledispatch [py26 or py27 or py33] - libhlc test: requires: - argparse [py26] - unittest2 [py26] - jinja2 files: - mandel.py commands: - pycc -h - numba -h
Update buildscript for custom HSA builds
Update buildscript for custom HSA builds
YAML
bsd-2-clause
stefanseefeld/numba,jriehl/numba,pombredanne/numba,sklam/numba,seibert/numba,numba/numba,gmarkall/numba,numba/numba,stefanseefeld/numba,stonebig/numba,stonebig/numba,IntelLabs/numba,numba/numba,stuartarchibald/numba,gmarkall/numba,jriehl/numba,gmarkall/numba,stonebig/numba,sklam/numba,stuartarchibald/numba,numba/numba,sklam/numba,cpcloud/numba,jriehl/numba,sklam/numba,gmarkall/numba,seibert/numba,pombredanne/numba,IntelLabs/numba,cpcloud/numba,jriehl/numba,IntelLabs/numba,pombredanne/numba,IntelLabs/numba,stuartarchibald/numba,numba/numba,gmarkall/numba,stonebig/numba,stefanseefeld/numba,IntelLabs/numba,seibert/numba,stefanseefeld/numba,stefanseefeld/numba,sklam/numba,pombredanne/numba,seibert/numba,jriehl/numba,seibert/numba,stonebig/numba,cpcloud/numba,stuartarchibald/numba,pombredanne/numba,cpcloud/numba,stuartarchibald/numba,cpcloud/numba
fce7a5af097120a4497317498d0099e0b878bc30
.codeclimate.yml
.codeclimate.yml
engines: csslint: enabled: true duplication: enabled: true config: languages: - python - javascript fixme: enabled: true pep8: enabled: true radon: enabled: true shellcheck: enabled: true ratings: paths: - app/** - bin/** exclude_paths: - config/
engines: csslint: enabled: true duplication: enabled: true config: languages: - python - javascript fixme: enabled: true pep8: enabled: true radon: enabled: true shellcheck: enabled: true ratings: paths: - app/** - bin/** exclude_paths: - config/ - app/static/css/bootstrap.css - app/static/css/normalize.css - app/static/tether.css - app/static/js/bootstrap.js - app/static/js/jquery.js - app/static/js/tether.js
Exclude third-party code from test coverage
Exclude third-party code from test coverage
YAML
mit
albertyw/base-flask,albertyw/base-flask,albertyw/base-flask,albertyw/base-flask
73859f3b540623eb2dd66a82314e9b73eb47195b
_data/projects.yml
_data/projects.yml
- name: cook description: CLI written in Node.js which allows you to easily install softwares on Ubuntu. url: https://github.com/mlcdf/cook img: 1f4e6 img_alt: package emoji color: 83dcff - name: ikarauraga description: 2D space shooter made with Unity for Ludum Dare 35 with Clément Stéphano. url: https://github.com/mlcdf/ikarauraga img: 1f3ae img_alt: video game controller emoji color: f51843 - name: dotfiles description: 'My config files: zsh, oh-my-zsh, z, git, node.js, python, atom, hyper, ...' url: https://github.com/mlcdf/dotfiles img: 1f527 img_alt: wrench emoji color: ffeb3b - name: generator-pym description: Scaffold out a Python module. Includes CI, license, Makefile, pylint, ... url: https://github.com/mlcdf/generator-pym img: 1f40d img_alt: snake emoji color: e9e9e9 - name: material-utilities description: Set of CSS classes & custom properties for Material shadows and opacities. url: https://github.com/mlcdf/material-utilities img: '2728' img_alt: sparkle emoji color: '000000'
- name: cook description: Easily install popular softwares that are missing from APT. url: https://github.com/cook-pkg/cook img: 1f4e6 img_alt: package emoji color: 83dcff - name: ikarauraga description: 2D space shooter made with Unity for Ludum Dare 35 with Clément Stéphano. url: https://github.com/mlcdf/ikarauraga img: 1f3ae img_alt: video game controller emoji color: f51843 - name: dotfiles description: 'My config files: zsh, oh-my-zsh, z, git, node.js, python, atom, hyper, ...' url: https://github.com/mlcdf/dotfiles img: 1f527 img_alt: wrench emoji color: ffeb3b - name: generator-pym description: Scaffold out a Python module. Includes CI, license, Makefile, pylint, ... url: https://github.com/mlcdf/generator-pym img: 1f40d img_alt: snake emoji color: e9e9e9 - name: material-utilities description: Set of CSS classes & custom properties for Material shadows and opacities. url: https://github.com/mlcdf/material-utilities img: '2728' img_alt: sparkle emoji color: '000000'
Update cook link & description
Update cook link & description
YAML
mit
mlcdf/website,mlcdf/website
49f58a7354800dfcdd231b54cc33642c3a2d95ef
.github/workflows/main.yml
.github/workflows/main.yml
name: dotfiles CI on: [push] jobs: build: runs-on: ${{ matrix.os }} strategy: matrix: os: [ubuntu-latest, macos-latest] node-version: [12.x] steps: - uses: actions/checkout@v1 - name: Use the Node.js ${{ matrix.node-version }} uses: actions/setup-node@v1 with: node-version: ${{ matrix.node-version }} - name: Update packages list and install zsh on Ubuntu if: runner.os == 'Linux' run: | sudo apt update -y sudo apt install -y zsh - name: Update packages list and install zsh on macOS if: runner.os == 'macOS' run: | brew update brew install zsh - name: Check syntax of initialize.zsh file run: zsh -n initialize.zsh env: CI: true - name: Check syntax of initialize script files run: "zsh -n ./init/*.zsh" env: CI: true - name: Check syntax of bin/pull-dotfiles file run: zsh -n bin/pull-dotfiles env: CI: true - name: Slack Notification uses: homoluctus/slatify@v1.5 if: always() with: job_name: '*dotfiles CI*' type: ${{ job.status }} icon_emoji: ":octocat:" url: ${{ secrets.SLACK_WEBHOOK }}
name: dotfiles CI on: [push] jobs: build: runs-on: ${{ matrix.os }} strategy: matrix: os: [ubuntu-latest, macos-latest] node-version: [12.x] steps: - uses: actions/checkout@v1 - name: Use the Node.js ${{ matrix.node-version }} uses: actions/setup-node@v1 with: node-version: ${{ matrix.node-version }} - name: Update packages list and install zsh on Ubuntu if: runner.os == 'Linux' run: | sudo apt update -y sudo apt install -y zsh - name: Update packages list and install zsh on macOS if: runner.os == 'macOS' run: | brew update brew install zsh - name: Check syntax of initialize.zsh file run: zsh -n initialize.zsh env: CI: true - name: Check syntax of initialize script files run: "zsh -n ./init/*.zsh" env: CI: true - name: Check syntax of bin/pull-dotfiles file run: zsh -n bin/pull-dotfiles env: CI: true - name: Slack Notification uses: homoluctus/slatify@v1.5 if: always() with: job_name: '*${{ github.workflow }}*' type: ${{ job.status }} icon_emoji: ":octocat:" url: ${{ secrets.SLACK_WEBHOOK }}
Use github.workflow value in slack notification
Use github.workflow value in slack notification
YAML
mit
Hiroto-K/dotfiles,Hiroto-K/dotfiles
930664471ca83855726957d5883a779c40db471d
.github/workflows/main.yml
.github/workflows/main.yml
# This is a basic workflow to help you get started with Actions name: CI on: push: branches: [ master ] pull_request: branches: [ master ] workflow_dispatch: jobs: test: runs-on: ubuntu-18.04 steps: - uses: actions/checkout@v3 - name: Setup Node.js environment uses: actions/setup-node@v3.5.0 with: node-version: 16.5.0 - name: Reuse npm cache folder uses: actions/cache@v3 env: cache-name: cache-node-modules with: path: | ~/.npm ./node_modules # invalidate cache when any package-lock.json changes key: ${{ runner.os }}-npm-test-x1-${{ hashFiles('**/package.json') }} restore-keys: | ${{ runner.os }}-npm-test-x1- - name: install npm dependencies # TODO remove legace peer deps once everything is up to date run: npm install --legacy-peer-deps - name: build run: npm run build - name: test:node run: npm run test:node - name: test:browser uses: GabrielBB/xvfb-action@v1 with: run: npm run test:browser - name: lint run: npm run lint
# This is a basic workflow to help you get started with Actions name: CI on: push: branches: [ master ] pull_request: branches: [ master ] workflow_dispatch: jobs: test: runs-on: ubuntu-18.04 steps: - uses: actions/checkout@v3 - name: Setup Node.js environment uses: actions/setup-node@v3.5.1 with: node-version: 16.5.0 - name: Reuse npm cache folder uses: actions/cache@v3 env: cache-name: cache-node-modules with: path: | ~/.npm ./node_modules # invalidate cache when any package-lock.json changes key: ${{ runner.os }}-npm-test-x1-${{ hashFiles('**/package.json') }} restore-keys: | ${{ runner.os }}-npm-test-x1- - name: install npm dependencies # TODO remove legace peer deps once everything is up to date run: npm install --legacy-peer-deps - name: build run: npm run build - name: test:node run: npm run test:node - name: test:browser uses: GabrielBB/xvfb-action@v1 with: run: npm run test:browser - name: lint run: npm run lint
Update actions/setup-node action to v3.5.1
Update actions/setup-node action to v3.5.1
YAML
apache-2.0
pubkey/async-test-util
9c67323b4bc94da3f7a7ce75fa56296f6d10a2c9
.github/workflows/main.yml
.github/workflows/main.yml
name: E-Mail Release Notification on: release: types: [released] jobs: notify: runs-on: ubuntu-latest steps: - name: Notify about a new release uses: daysofwonder/sendgrid-notification-action@v1.4 env: SENDGRID_API_TOKEN: ${{ secrets.SENDGRID_API_TOKEN }} RECIPIENTS_URL: ${{ secrets.RECIPIENTS_URL }} SENDER_EMAIL: ${{ secrets.SENDER_EMAIL }} DISTRIBUTION_LISTS: ${{ secrets.DISTRIBUTION_LISTS }}
name: E-Mail Release Notification on: release: types: [released] jobs: notify: runs-on: ubuntu-latest steps: - name: Notify about a new release uses: daysofwonder/sendgrid-notification-action@v1.5 env: SENDGRID_API_TOKEN: ${{ secrets.SENDGRID_API_TOKEN }} RECIPIENTS_URL: ${{ secrets.RECIPIENTS_URL }} SENDER_EMAIL: ${{ secrets.SENDER_EMAIL }} DISTRIBUTION_LISTS: ${{ secrets.DISTRIBUTION_LISTS }}
Update sendgrid notif uses version in action notify
Update sendgrid notif uses version in action notify
YAML
mit
daysofwonder/asmodeenet_sso_js,daysofwonder/asmodeenet_sso_js,daysofwonder/asmodeenet_sso_js
780b1a9326e55d4524c2052aa2bb4564fd251bf4
build/docker-compose.yml
build/docker-compose.yml
version: '2' services: # the web container web-server: image: venveo/baseimage:7.0-apache volumes: - "./php:/var/www/public" ports: - "80:80" - "443:443" links: - mysql-server - redis-server - memcached-server depends_on: - mysql-server - redis-server - memcached-server environment: - APP_ENV=local - APP_CPTRIGGER=manage # mysql container mysql-server: image: venveo/baseimage:5.7-mysql volumes: - "./mysql/data:/var/lib/mysql" ports: - "3306:3306" environment: - MYSQL_ROOT_PASSWORD=TestingRootPassword - MYSQL_DATABASE=testing_db - MYSQL_USER=testing_user - MYSQL_PASSWORD=testing_password - MYSQL_ALLOW_EMPTY_PASSWORD=no # redis container redis-server: image: venveo/baseimage:3.2-redis ports: - "6379:6379" # memcached container memcached-server: image: venveo/baseimage:1.4-memcached ports: - "11211:11211"
version: '2' services: # the web container web-server: image: venveo/baseimage:1.10-nginx # image: venveo/baseimage:7.0-apache volumes: - "./php:/var/www/public" # - "./php:/usr/share/nginx/html" ports: - "80:80" - "443:443" links: - mysql-server - redis-server - memcached-server - php-fpm depends_on: - mysql-server - redis-server - memcached-server - php-fpm environment: - APP_ENV=local - APP_CPTRIGGER=manage # mysql container mysql-server: image: venveo/baseimage:5.7-mysql volumes: - "./mysql/data:/var/lib/mysql" ports: - "3306:3306" environment: - MYSQL_ROOT_PASSWORD=TestingRootPassword - MYSQL_DATABASE=testing_db - MYSQL_USER=testing_user - MYSQL_PASSWORD=testing_password - MYSQL_ALLOW_EMPTY_PASSWORD=no # redis container redis-server: image: venveo/baseimage:3.2-redis ports: - "6379:6379" # memcached container memcached-server: image: venveo/baseimage:1.4-memcached ports: - "11211:11211" # php-fpm container php-fpm: image: venveo/baseimage:7.0-php-fpm ports: - "9000:9000"
Update compose to test nginx and php-fpm
Update compose to test nginx and php-fpm
YAML
mit
venveo/docker-baseimage,venveo/docker-baseimage
a5fcc710a799e63e7c2163f7d9b12c3e1c93294d
appveyor.yml
appveyor.yml
version: 1.0.{build} configuration: - Debug - Release clone_folder: C:\Fruit environment: BOOST_DIR: C:\Libraries\boost matrix: - image: Visual Studio 2017 CMAKE_PATH: '"C:\Program Files (x86)\CMake\bin"' CMAKE_GENERATOR: '"MinGW Makefiles"' - image: Visual Studio 2015 CMAKE_PATH: '"C:\Program Files (x86)\CMake\bin"' CMAKE_GENERATOR: '"Visual Studio 14 2015 Win64"' VCVARSALL_SCRIPT: '"C:\Program Files (x86)\Microsoft Visual Studio 14.0\VC\vcvarsall.bat"' - image: Visual Studio 2017 CMAKE_PATH: '"C:\Program Files (x86)\Microsoft Visual Studio\2017\Community\Common7\IDE\CommonExtensions\Microsoft\CMake\CMake\bin"' CMAKE_GENERATOR: '"Visual Studio 15 2017 Win64"' VCVARSALL_SCRIPT: '"C:\Program Files (x86)\Microsoft Visual Studio\2017\Community\VC\Auxiliary\Build\vcvarsall.bat"' install: - cmd: setx PATH %PATH:C:\Program Files\Git\usr\bin;=% build_script: - cmd: C:\Fruit\extras\scripts\postsubmit.bat
version: 1.0.{build} configuration: - Debug - Release clone_folder: C:\Fruit environment: BOOST_DIR: C:\Libraries\boost matrix: - image: Visual Studio 2017 CMAKE_PATH: '"C:\Program Files (x86)\CMake\bin"' CMAKE_GENERATOR: '"MinGW Makefiles"' - image: Visual Studio 2015 CMAKE_PATH: '"C:\Program Files (x86)\CMake\bin"' CMAKE_GENERATOR: '"Visual Studio 14 2015 Win64"' VCVARSALL_SCRIPT: '"C:\Program Files (x86)\Microsoft Visual Studio 14.0\VC\vcvarsall.bat"' - image: Visual Studio 2017 CMAKE_PATH: '"C:\Program Files (x86)\Microsoft Visual Studio\2017\Community\Common7\IDE\CommonExtensions\Microsoft\CMake\CMake\bin"' CMAKE_GENERATOR: '"Visual Studio 15 2017 Win64"' VCVARSALL_SCRIPT: '"C:\Program Files (x86)\Microsoft Visual Studio\2017\Community\VC\Auxiliary\Build\vcvarsall.bat"' install: - cmd: setx PATH "%PATH:C:\Program Files\Git\usr\bin;=%" build_script: - cmd: C:\Fruit\extras\scripts\postsubmit.bat
Fix quoting in a setx command used when testing Fruit on Windows in AppVeyor.
Fix quoting in a setx command used when testing Fruit on Windows in AppVeyor.
YAML
apache-2.0
google/fruit,google/fruit,google/fruit
25fbf5304019458c30835ae8ab933610a055c3f0
appveyor.yml
appveyor.yml
version: 0.1.{build} configuration: Release platform: Any CPU build_script: - cmd: .\build.bat Package test: off artifacts: - path: '**\CertiPay.*.nupkg'
version: 0.1.{build} configuration: Release platform: Any CPU build_script: - cmd: .\build.bat Package test: off artifacts: - path: '**\CertiPay.Services.PDF.*.nupkg'
Tweak publish artifcats to only catch the service
Tweak publish artifcats to only catch the service
YAML
mit
kylebjones/CertiPay.Services.PDF
58155cefe86ccc7b2082d40f669fec2a7dc1b9e5
appveyor.yml
appveyor.yml
version: 2.1.18.{build} pull_requests: do_not_increment_build_number: true branches: only: - master image: Visual Studio 2017 configuration: Release init: - -cmd: git config --global core.autocrlf true dotnet_csproj: patch: true file: '**\*.csproj' version: '{version}' package_version: '{version}' assembly_version: '{version}' file_version: '{version}' informational_version: '{version}' before_build: - cmd: choco install codecov - cmd: dotnet restore build: publish_nuget: true publish_nuget_symbols: true verbosity: minimal test_script: - cmd: dotnet test /p:CollectCoverage=true /p:CoverletOutputFormat=opencover ./test/ArgSentry.Test/ArgSentry.Test.csproj --logger AppVeyor - cmd: codecov -f "./test/ArgSentry.Test/coverage.opencover.xml" artifacts: - path: '**\*.nupkg' deploy: - provider: NuGet api_key: secure: 92tF47KjBR7PvZKHSvOyde7y5i17Ciwai1+X4gVtNKqGEK1ZrVl8EHQkatu/BjcH skip_symbols: false artifact: /.*\.nupkg/ on: branch: master
version: 2.1.{build} pull_requests: do_not_increment_build_number: true branches: only: - master image: Visual Studio 2017 configuration: Release init: - cmd: git config --global core.autocrlf true install: - cmd: choco install codecov dotnet_csproj: patch: true file: '**\*.csproj' version: '{version}' package_version: '{version}' assembly_version: '{version}' file_version: '{version}' informational_version: '{version}' before_build: - cmd: dotnet restore build_script: - cmd: dotnet build ./src/ArgSentry/ArgSentry.csproj --configuration Release --no-restore test_script: - cmd: dotnet test /p:CollectCoverage=true /p:CoverletOutputFormat=opencover ./test/ArgSentry.Test/ArgSentry.Test.csproj --logger AppVeyor - cmd: codecov -f "./test/ArgSentry.Test/coverage.opencover.xml" after_test: - cmd: dotnet pack ./src/ArgSentry/ArgSentry.csproj --configuration Release --no-build --no-restore --output ../../dist artifacts: - path: './dist/*.nupkg' deploy: - provider: NuGet api_key: secure: 92tF47KjBR7PvZKHSvOyde7y5i17Ciwai1+X4gVtNKqGEK1ZrVl8EHQkatu/BjcH skip_symbols: true artifact: /.*\.nupkg/ on: branch: master
Build & deploy process cleanup
Build & deploy process cleanup
YAML
mit
eric-davis/ArgSentry
4b471f61de3d5c4876a4e5e10b6c6fb3296ed416
tasks/main.yml
tasks/main.yml
--- - name: "++++++++++ install and config dehydrated and all challenge hooks ++++++++++" include_tasks: setup.yml become: true when: - le_certificates is defined - le_deploy == false - ansible_os_family == "Debian" tags: - setup - name: "++++++++++ sign/renew non-existant/changed/expiring certificates ++++++++++" include_tasks: certificate.yml become: true when: - le_certificates is defined - le_deploy == false - ansible_os_family == "Debian" with_items: - "{{ le_certificates }}" loop_control: loop_var: le_certificate_loop tags: - certificate - name: "++++++++++ deploy signed/renewed certificates to other webservers +++++++++" include_tasks: deploy.yml when: (le_deploy == true or le_deploy == "force") and le_deploy_certificates is defined with_items: - "{{ le_deploy_certificates }}" loop_control: loop_var: le_deploy_certificate_loop tags: - deploy
--- - name: "++++++++++ install and config dehydrated and all challenge hooks ++++++++++" include_tasks: setup.yml when: - le_certificates is defined - le_deploy == false - ansible_os_family == "Debian" tags: - setup - name: "++++++++++ sign/renew non-existant/changed/expiring certificates ++++++++++" include_tasks: certificate.yml when: - le_certificates is defined - le_deploy == false - ansible_os_family == "Debian" with_items: - "{{ le_certificates }}" loop_control: loop_var: le_certificate_loop tags: - certificate - name: "++++++++++ deploy signed/renewed certificates to other webservers +++++++++" include_tasks: deploy.yml when: (le_deploy == true or le_deploy == "force") and le_deploy_certificates is defined with_items: - "{{ le_deploy_certificates }}" loop_control: loop_var: le_deploy_certificate_loop tags: - deploy
Remove become statement in dynamic tasks include
Remove become statement in dynamic tasks include
YAML
mit
rembik/ansible-role-letsencrypt,rembik/ansible-role-letsencrypt
9cf2ef41a908a1a126b5418806a2ba4de31ebb86
.conda/meta.yaml
.conda/meta.yaml
package: name: csdms-dakota version: {{"0.2" + environ.get('BUILD_STR', '')}} source: path: .. requirements: build: - python - numpy - pyyaml - basic-modeling-interface run: - python - numpy - pyyaml - basic-modeling-interface - nose
package: name: csdms-dakota version: {{"0.2" + environ.get('BUILD_STR', '')}} source: path: .. requirements: build: - python - numpy - pyyaml - basic-modeling-interface - nose run: - python - numpy - pyyaml - basic-modeling-interface - nose
Include nose in build requirements
Include nose in build requirements
YAML
mit
csdms/dakota,csdms/dakota
4ad385f3bb4504cc156485ed941a67aab4133423
deploy/kubernetes/elasticsearch/snapshot-cronjob.yaml
deploy/kubernetes/elasticsearch/snapshot-cronjob.yaml
apiVersion: batch/v1beta1 kind: CronJob metadata: name: es-monthly-snapshotter labels: name: es-monthly-snapshotter spec: schedule: "@monthly" jobTemplate: spec: template: spec: containers: - name: es-monthly-snapshotter image: {{ DOCKER_IMAGE_PREFIX }}/curl{{ DOCKER_IMAGE_TAG }} imagePullPolicy: {{ IMAGE_PULL_POLICY }} env: - name: ES_PASSWORD valueFrom: secretKeyRef: name: kibana-secrets key: elasticsearch.password - name: ELASTICSEARCH_SERVICE_HOSTNAME value: "{{ ELASTICSEARCH_SERVICE_HOSTNAME }}" args: - /bin/sh - -c - 'curl -s -i -u "kibana:${ES_PASSWORD}" -XPUT "http://${ELASTICSEARCH_SERVICE_HOSTNAME}:9200/_snapshot/snapshot_storage/%3Csnapshot-%7Bnow%2Fd%7D%3E" | tee /dev/stderr | grep "200 OK"' restartPolicy: OnFailure
apiVersion: batch/v1beta1 kind: CronJob metadata: name: es-monthly-snapshotter labels: name: es-monthly-snapshotter spec: schedule: "@monthly" concurrencyPolicy: Forbid jobTemplate: spec: template: metadata: labels: app: es-monthly-snapshotter spec: containers: - name: es-monthly-snapshotter image: {{ DOCKER_IMAGE_PREFIX }}/curl{{ DOCKER_IMAGE_TAG }} imagePullPolicy: {{ IMAGE_PULL_POLICY }} env: - name: ES_PASSWORD valueFrom: secretKeyRef: name: kibana-secrets key: elasticsearch.password - name: ELASTICSEARCH_SERVICE_HOSTNAME value: "{{ ELASTICSEARCH_SERVICE_HOSTNAME }}" args: - /bin/sh - -c - 'curl -s -i -u "kibana:${ES_PASSWORD}" -XPUT "http://${ELASTICSEARCH_SERVICE_HOSTNAME}:9200/_snapshot/snapshot_storage/%3Csnapshot-%7Bnow%2Fd%7D%3E" | tee /dev/stderr | grep "200 OK"' restartPolicy: OnFailure
Add tagging for easier log alerting, forbid concurrent pods
Add tagging for easier log alerting, forbid concurrent pods
YAML
agpl-3.0
ssadedin/seqr,ssadedin/seqr,ssadedin/seqr,ssadedin/seqr,ssadedin/seqr
5ca82302472740c4bc577889e5b2fda96bd93d2d
.forestry/settings.yml
.forestry/settings.yml
--- new_page_extension: md auto_deploy: false admin_path: "/admin" webhook_url: sections: - type: jekyll-pages label: Pages create: all - type: jekyll-posts label: Posts create: all upload_dir: uploads public_path: "/uploads" front_matter_path: '' use_front_matter_path: file_template: ":filename:" build: preview_command: bundle exec jekyll build --drafts --unpublished --future -d _site publish_command: bundle exec jekyll build -d _site preview_env: - JEKYLL_ENV=staging publish_env: - JEKYLL_ENV=production preview_output_directory: _site output_directory: _site
--- new_page_extension: md auto_deploy: true admin_path: "/admin" webhook_url: sections: - type: jekyll-pages label: Pages create: all - type: jekyll-posts label: Posts create: all upload_dir: uploads public_path: "/uploads" front_matter_path: '' use_front_matter_path: file_template: ":filename:" build: preview_command: bundle exec jekyll build --drafts --unpublished --future -d _site publish_command: bundle exec jekyll build -d _site preview_env: - JEKYLL_ENV=staging publish_env: - JEKYLL_ENV=production preview_output_directory: _site output_directory: _site
Update from Forestry.io - Updated Forestry configuration
Update from Forestry.io - Updated Forestry configuration
YAML
mit
digitalfort/goulburnspeedway
bc91abff849d5085025bd202fd77911b6e5e358e
metadata/me.lucky.duress.yml
metadata/me.lucky.duress.yml
Categories: - Security License: GPL-3.0-only AuthorName: lucky AuthorEmail: 44uaanjm0@relay.firefox.com AuthorWebSite: https://github.com/x13a SourceCode: https://github.com/x13a/Duress IssueTracker: https://github.com/x13a/Duress/issues AutoName: Duress RepoType: git Repo: https://github.com/x13a/Duress.git Builds: - versionName: 1.0.1 versionCode: 2 commit: cd358b81b57ac772ccf1b723e7a9a7625c17545e subdir: app sudo: - apt-get update || apt-get update - apt-get install -y openjdk-11-jdk-headless - update-alternatives --auto java gradle: - yes AutoUpdateMode: Version v%v UpdateCheckMode: Tags CurrentVersion: 1.0.1 CurrentVersionCode: 2
Categories: - Security License: GPL-3.0-only AuthorName: lucky AuthorEmail: 44uaanjm0@relay.firefox.com AuthorWebSite: https://github.com/x13a SourceCode: https://github.com/x13a/Duress IssueTracker: https://github.com/x13a/Duress/issues AutoName: Duress RepoType: git Repo: https://github.com/x13a/Duress.git Builds: - versionName: 1.0.1 versionCode: 2 commit: cd358b81b57ac772ccf1b723e7a9a7625c17545e subdir: app sudo: - apt-get update || apt-get update - apt-get install -y openjdk-11-jdk-headless - update-alternatives --auto java gradle: - yes - versionName: 1.0.2 versionCode: 3 commit: 326e3c7a073a4ba48ef69dfaa327556ffd2343c8 subdir: app sudo: - apt-get update || apt-get update - apt-get install -y openjdk-11-jdk-headless - update-alternatives --auto java gradle: - yes AutoUpdateMode: Version v%v UpdateCheckMode: Tags CurrentVersion: 1.0.2 CurrentVersionCode: 3
Update Duress to 1.0.2 (3)
Update Duress to 1.0.2 (3)
YAML
agpl-3.0
f-droid/fdroiddata,f-droid/fdroiddata
d6e138cd384aad02f37851c8d748fc81262e7af4
devtools/conda-recipe/meta.yaml
devtools/conda-recipe/meta.yaml
package: name: openmoltools-dev version: !!str dev source: path: ../../ build: preserve_egg_dir: True number: 0 requirements: build: - python - setuptools - pandas - six - mdtraj - numpy - numpydoc - scipy - openmm - ambermini - pytables - parmed # - rdkit # rdkit is an optional dependency, may want to comment this out for the release version. run: - python - setuptools - pandas - six - mdtraj - numpy - numpydoc - scipy - openmm - ambermini - pytables - parmed # - rdkit # rdkit is an optional dependency, may want to comment this out for the release version. test: requires: - nose imports: - openmoltools #commands: # - nosetests openmoltools -v -a '!slow' about: home: https://github.com/choderalab/openmoltools license: GNU Lesser General Public License v2 or later (LGPLv2+)
package: name: openmoltools-dev version: !!str dev source: path: ../../ build: preserve_egg_dir: True number: 0 requirements: build: - python - setuptools - pandas - six - mdtraj - numpy - numpydoc - scipy - openmm-dev - ambermini - pytables - parmed # - rdkit # rdkit is an optional dependency, may want to comment this out for the release version. run: - python - setuptools - pandas - six - mdtraj - numpy - numpydoc - scipy - openmm-dev - ambermini - pytables - parmed # - rdkit # rdkit is an optional dependency, may want to comment this out for the release version. test: requires: - nose imports: - openmoltools #commands: # - nosetests openmoltools -v -a '!slow' about: home: https://github.com/choderalab/openmoltools license: GNU Lesser General Public License v2 or later (LGPLv2+)
Change openmm test dependency to openmm-dev
Change openmm test dependency to openmm-dev
YAML
mit
choderalab/openmoltools,choderalab/openmoltools
448fe2dd64b52e41e647320ee002c5a6ff84e23b
devtools/conda-recipe/meta.yaml
devtools/conda-recipe/meta.yaml
package: name: perses-dev version: !!str 0.0.0 source: path: ../../ build: preserve_egg_dir: True number: 0 requirements: build: - python - setuptools - openmm-dev - parmed-dev - numpy - scipy - openmoltools-dev - alchemy - numexpr - openmmtools - pymbar - autograd run: - python - openmm-dev - parmed-dev - numpy - scipy - openmoltools-dev - alchemy - numexpr - openmmtools - pymbar - autograd test: requires: - nose - openmmtools - pymbar imports: - perses about: home: https://github.com/choderalab/perses license: GNU Lesser General Public License v2 or later (LGPLv2+)
package: name: perses-dev version: !!str 0.0.0 source: path: ../../ build: preserve_egg_dir: True number: 0 requirements: build: - python - setuptools - numpy - scipy - numexpr - autograd - pymbar-dev - openmm-dev - parmed-dev - openmoltools-dev - alchemy-dev - openmmtools-dev run: - python - setuptools - numpy - scipy - numexpr - autograd - pymbar-dev - openmm-dev - parmed-dev - openmoltools-dev - alchemy-dev - openmmtools-dev test: requires: - nose imports: - perses about: home: https://github.com/choderalab/perses license: GNU Lesser General Public License v2 or later (LGPLv2+)
Fix conda recipe dependencies to ensure we get openmm-dev and parmed-dev
Fix conda recipe dependencies to ensure we get openmm-dev and parmed-dev
YAML
mit
choderalab/perses,choderalab/perses
9c6e2ec6a2fb4a457b3cd3e7ca08be468110fdc4
.github/dependabot.yml
.github/dependabot.yml
# https://help.github.com/github/administering-a-repository/configuration-options-for-dependency-updates version: 2 updates: - package-ecosystem: "maven" directory: "/" schedule: interval: "weekly" - package-ecosystem: "github-actions" directory: "/" schedule: interval: "weekly"
# https://help.github.com/github/administering-a-repository/configuration-options-for-dependency-updates version: 2 updates: - package-ecosystem: "maven" open-pull-requests-limit: 10 directory: "/" schedule: interval: "weekly" - package-ecosystem: "github-actions" directory: "/" schedule: interval: "weekly"
Increase open pull requests limit
Increase open pull requests limit
YAML
mit
jan-molak/jenkins-build-monitor-plugin,jan-molak/jenkins-build-monitor-plugin,jan-molak/jenkins-build-monitor-plugin,jan-molak/jenkins-build-monitor-plugin
b224c285e684ced74f7ba332f97bf6c9aa3f306b
stacks/resource-groups.yml
stacks/resource-groups.yml
# stacks/resource-groups.yml AWSTemplateFormatVersion: "2010-09-09" Description: >- Creates resource groups for resources in the stack, based on various sets of tag and resource type filters. Parameters: EnvironmentType: { Type: String } RootStackName: { Type: String } RootStackId: { Type: String } Resources: StackResourceGroup: Type: AWS::ResourceGroups::Group Properties: Description: >- Inlcudes all supported resources that belong to the root stack Name: !Sub ${RootStackName} ResourceQuery: Query: ResourceTypeFilters: - AWS::AllSupported TagFilters: - Key: prx:cloudformation:root-stack-name Values: - !Ref RootStackName Type: TAG_FILTERS_1_0 Tags: - { Key: prx:meta:tagging-version, Value: "2021-04-07" } - { Key: prx:cloudformation:stack-name, Value: !Ref AWS::StackName } - { Key: prx:cloudformation:stack-id, Value: !Ref AWS::StackId } # - { Key: prx:cloudformation:root-stack-name, Value: !Ref RootStackName } # - { Key: prx:cloudformation:root-stack-id, Value: !Ref RootStackId } - { Key: prx:ops:environment, Value: !Ref EnvironmentType } - { Key: prx:dev:application, Value: Infrastructure } Outputs: StackResourceGroupName: Value: !Ref StackResourceGroup
# stacks/resource-groups.yml AWSTemplateFormatVersion: "2010-09-09" Description: >- Creates resource groups for resources in the stack, based on various sets of tag and resource type filters. Parameters: EnvironmentType: { Type: String } RootStackName: { Type: String } RootStackId: { Type: String } Resources: StackResourceGroup: Type: AWS::ResourceGroups::Group Properties: Description: >- Inlcudes all supported resources that belong to the root stack Name: !Sub ${RootStackName} ResourceQuery: Query: ResourceTypeFilters: - AWS::AllSupported TagFilters: - Key: prx:cloudformation:root-stack-name Values: - !Ref RootStackName Type: TAG_FILTERS_1_0 # Tags: # - { Key: prx:meta:tagging-version, Value: "2021-04-07" } # - { Key: prx:cloudformation:stack-name, Value: !Ref AWS::StackName } # - { Key: prx:cloudformation:stack-id, Value: !Ref AWS::StackId } # # - { Key: prx:cloudformation:root-stack-name, Value: !Ref RootStackName } # # - { Key: prx:cloudformation:root-stack-id, Value: !Ref RootStackId } # - { Key: prx:ops:environment, Value: !Ref EnvironmentType } # - { Key: prx:dev:application, Value: Infrastructure } Outputs: StackResourceGroupName: Value: !Ref StackResourceGroup
Remove all resource group tags?
Remove all resource group tags?
YAML
mit
PRX/Infrastructure,PRX/Infrastructure,PRX/Infrastructure,PRX/Infrastructure,PRX/Infrastructure
d3f7b177e0e90eed6e153d582544660cf743e828
.github/dependabot.yml
.github/dependabot.yml
# To get started with Dependabot version updates, you'll need to specify which # package ecosystems to update and where the package manifests are located. # Please see the documentation for all configuration options: # https://help.github.com/github/administering-a-repository/configuration-options-for-dependency-updates version: 2 updates: - package-ecosystem: "docker" directory: "/" schedule: interval: "daily" open-pull-requests-limit: 10 - package-ecosystem: "npm" directory: "/" schedule: interval: "daily" open-pull-requests-limit: 10 - package-ecosystem: "pip" directory: "/" schedule: interval: "daily" open-pull-requests-limit: 10 - package-ecosystem: "github-actions" directory: "/" schedule: interval: "daily" open-pull-requests-limit: 10
# To get started with Dependabot version updates, you'll need to specify which # package ecosystems to update and where the package manifests are located. # Please see the documentation for all configuration options: # https://help.github.com/github/administering-a-repository/configuration-options-for-dependency-updates version: 2 updates: - package-ecosystem: "docker" directory: "/" schedule: interval: "daily" open-pull-requests-limit: 10 - package-ecosystem: "npm" directory: "/csunplugged/" schedule: interval: "daily" open-pull-requests-limit: 10 - package-ecosystem: "pip" directory: "/" schedule: interval: "daily" open-pull-requests-limit: 10 - package-ecosystem: "github-actions" directory: "/" schedule: interval: "daily" open-pull-requests-limit: 10
Update location of JavaScript dependency files
Update location of JavaScript dependency files
YAML
mit
uccser/cs-unplugged,uccser/cs-unplugged,uccser/cs-unplugged,uccser/cs-unplugged
a645b82cddadce0f0a702a01774b12d0d477748b
config/outgoing_mail.yml
config/outgoing_mail.yml
development: address: "smtp.example.com" port: "25" user_name: "user" password: "password" authentication: "plain" # plain, login, or cram_md5 domain: "example.com" outgoing_address: "canvas@example.com" default_name: "Instructure Canvas" # or to use sendmail -- # # development: # delivery_method: "sendmail" test: delivery_method: "test" production: address: "localhost" port: "25" # user_name: "user" # password: "password" # authentication: "plain" # plain, login, or cram_md5 domain: "canvas.localdomain" outgoing_address: "canvas@canvas.localdomain" default_name: "Instructure Canvas" # If receiving mail from multiple inboxes (see incoming_mail.yml.example), # you'll want to include those addresses in a reply_to_addresses array so # Canvas will select the Reply-To field of outgoing messages from all of the # incoming mailboxes. multiple_inboxes: address: "smtp.example.com" port: "25" user_name: "user" password: "password" authentication: "plain" # plain, login, or cram_md5 domain: "example.com" outgoing_address: "canvas@example.com" default_name: "Instructure Canvas" reply_to_addresses: - "canvas1@example.com" - "canvas2@example.com" - "canvas3@example.com" - "canvas4@example.com"
development: delivery_method: "sendmail" # or to use sendmail -- # # development: # delivery_method: "sendmail" test: delivery_method: "test" production: address: "localhost" port: "25" # user_name: "user" # password: "password" # authentication: "plain" # plain, login, or cram_md5 domain: "canvas.localdomain" outgoing_address: "canvas@canvas.localdomain" default_name: "Instructure Canvas" # If receiving mail from multiple inboxes (see incoming_mail.yml.example), # you'll want to include those addresses in a reply_to_addresses array so # Canvas will select the Reply-To field of outgoing messages from all of the # incoming mailboxes. multiple_inboxes: address: "smtp.example.com" port: "25" user_name: "user" password: "password" authentication: "plain" # plain, login, or cram_md5 domain: "example.com" outgoing_address: "canvas@example.com" default_name: "Instructure Canvas" reply_to_addresses: - "canvas1@example.com" - "canvas2@example.com" - "canvas3@example.com" - "canvas4@example.com"
Use sendmail in dev mode
Use sendmail in dev mode
YAML
agpl-3.0
HotChalk/canvas-lms,HotChalk/canvas-lms,HotChalk/canvas-lms,HotChalk/canvas-lms
301c82a49bbdd9c21de2c8b03bc75db41bdef971
metadata/t20kdc.offlinepuzzlesolver.yml
metadata/t20kdc.offlinepuzzlesolver.yml
Categories: - Games License: GPL-3.0-or-later AuthorName: 20kdc SourceCode: https://gitlab.com/20kdc/offline-puzzle-solver IssueTracker: https://gitlab.com/20kdc/offline-puzzle-solver/-/issues AutoName: Offline Puzzle Solver RepoType: git Repo: https://gitlab.com/20kdc/offline-puzzle-solver.git Builds: - versionName: '1.0' versionCode: 1 commit: v1 subdir: app gradle: - yes - versionName: '2.0' versionCode: 2 commit: 49d9c3f4f28071b8ee406f487e6f00ee2b8eb0cf subdir: app gradle: - yes AutoUpdateMode: Version v%c UpdateCheckMode: Tags CurrentVersion: '2.0' CurrentVersionCode: 2
Categories: - Games License: GPL-3.0-or-later AuthorName: 20kdc SourceCode: https://gitlab.com/20kdc/offline-puzzle-solver IssueTracker: https://gitlab.com/20kdc/offline-puzzle-solver/-/issues AutoName: Offline Puzzle Solver RepoType: git Repo: https://gitlab.com/20kdc/offline-puzzle-solver.git Builds: - versionName: '1.0' versionCode: 1 commit: v1 subdir: app gradle: - yes - versionName: '2.0' versionCode: 2 commit: 49d9c3f4f28071b8ee406f487e6f00ee2b8eb0cf subdir: app gradle: - yes - versionName: '4.0' versionCode: 4 commit: 06320725744f782eac6bfc2fcd504a01e5a801b7 subdir: app gradle: - yes AutoUpdateMode: Version v%c UpdateCheckMode: Tags CurrentVersion: '4.0' CurrentVersionCode: 4
Update Offline Puzzle Solver to 4.0 (4)
Update Offline Puzzle Solver to 4.0 (4)
YAML
agpl-3.0
f-droid/fdroiddata,f-droid/fdroiddata
e16d451f5b7dd56f9047268fe0a2be23f608dd5c
config/jobs/image-pushing/k8s-staging-dns.yaml
config/jobs/image-pushing/k8s-staging-dns.yaml
postsubmits: # this is the github repo we'll build from; this block needs to be repeated for each repo. kubernetes/dns: - name: dns-push-images cluster: k8s-infra-prow-build-trusted annotations: # this is the name of some testgrid dashboard to report to. testgrid-dashboards: sig-network-dns decorate: true branches: - ^master$ spec: serviceAccountName: gcb-builder containers: - image: gcr.io/k8s-testimages/image-builder:v20200901-ab141a0 command: - /run.sh args: # this is the project GCB will run in, which is the same as the GCR images are pushed to. - --project=k8s-staging-dns - --scratch-bucket=gs://k8s-staging-dns-gcb - --env-passthrough=PULL_BASE_REF - .
postsubmits: # this is the github repo we'll build from; this block needs to be repeated for each repo. kubernetes/dns: - name: dns-push-images cluster: k8s-infra-prow-build-trusted annotations: # this is the name of some testgrid dashboard to report to. testgrid-dashboards: sig-network-dns decorate: true # match on regex '^\d+\.\d+\.\d+$' to trigger job on new tags branches: - ^master$ - ^\d+\.\d+\.\d+$ spec: serviceAccountName: gcb-builder containers: - image: gcr.io/k8s-testimages/image-builder:v20200901-ab141a0 command: - /run.sh args: # this is the project GCB will run in, which is the same as the GCR images are pushed to. - --project=k8s-staging-dns - --scratch-bucket=gs://k8s-staging-dns-gcb - --env-passthrough=PULL_BASE_REF - .
Update DNS image push yaml to trigger on new tags
Update DNS image push yaml to trigger on new tags
YAML
apache-2.0
cblecker/test-infra,cjwagner/test-infra,fejta/test-infra,brahmaroutu/test-infra,michelle192837/test-infra,BenTheElder/test-infra,kubernetes/test-infra,cblecker/test-infra,pwittrock/test-infra,jessfraz/test-infra,pwittrock/test-infra,kubernetes/test-infra,BenTheElder/test-infra,dims/test-infra,cjwagner/test-infra,dims/test-infra,BenTheElder/test-infra,michelle192837/test-infra,BenTheElder/test-infra,cjwagner/test-infra,cblecker/test-infra,michelle192837/test-infra,cjwagner/test-infra,monopole/test-infra,brahmaroutu/test-infra,kubernetes/test-infra,BenTheElder/test-infra,BenTheElder/test-infra,cblecker/test-infra,dims/test-infra,dims/test-infra,monopole/test-infra,fejta/test-infra,michelle192837/test-infra,monopole/test-infra,fejta/test-infra,jessfraz/test-infra,jessfraz/test-infra,fejta/test-infra,cblecker/test-infra,pwittrock/test-infra,dims/test-infra,fejta/test-infra,dims/test-infra,monopole/test-infra,jessfraz/test-infra,jessfraz/test-infra,kubernetes/test-infra,michelle192837/test-infra,brahmaroutu/test-infra,monopole/test-infra,pwittrock/test-infra,kubernetes/test-infra,brahmaroutu/test-infra,michelle192837/test-infra,brahmaroutu/test-infra,pwittrock/test-infra,jessfraz/test-infra,fejta/test-infra,kubernetes/test-infra,cjwagner/test-infra,monopole/test-infra,brahmaroutu/test-infra,cblecker/test-infra,cjwagner/test-infra
2625b41203d796312cfd9639f3bf8e25235e32bb
.bazelci/presubmit.yml
.bazelci/presubmit.yml
--- # NOTE: To avoid listing the same things for build_flags/test_flags for each # of these tasks, they are listed in the .bazelrc instead. tasks: macos_latest: name: "Latest Bazel" platform: macos bazel: latest build_targets: - "tools/..." - "test/..." test_targets: - "tools/..." - "test/..." - "examples/..." macos_last_green: name: "Last Green Bazel" platform: macos bazel: last_green build_targets: - "tools/..." - "test/..." test_targets: - "tools/..." - "test/..." - "examples/..." macos_latest_head_deps: name: "Latest Bazel with Head Deps" platform: macos bazel: latest shell_commands: # Update the WORKSPACE to use head versions of some deps to ensure nothing # has landed on them breaking this project. - .bazelci/update_workspace_to_deps_heads.sh build_targets: - "tools/..." - "test/..." test_targets: - "tools/..." - "test/..." - "examples/..." buildifier: version: latest # Function doc string requirements are a little hard to follow, they did get # split into multiple warnings, so there is now just one left that causes # warnings. For the history of some of the issues, see: # https://github.com/bazelbuild/buildtools/issues/576 warnings: -function-docstring-args
--- x_defaults: # YAML has a feature for "repeated nodes", BazelCI is fine with extra nodes # it doesn't know about; so that is used to avoid repeating common subparts. common: &common platform: macos build_targets: - "tools/..." - "test/..." test_targets: - "tools/..." - "test/..." - "examples/..." # NOTE: To avoid listing the same things for build_flags/test_flags for each # of these tasks, they are listed in the .bazelrc instead. tasks: macos_latest: name: "Latest Bazel" bazel: latest <<: *common macos_last_green: name: "Last Green Bazel" bazel: last_green <<: *common macos_latest_head_deps: name: "Latest Bazel with Head Deps" bazel: latest shell_commands: # Update the WORKSPACE to use head versions of some deps to ensure nothing # has landed on them breaking this project. - .bazelci/update_workspace_to_deps_heads.sh <<: *common buildifier: version: latest # Function doc string requirements are a little hard to follow, they did get # split into multiple warnings, so there is now just one left that causes # warnings. For the history of some of the issues, see: # https://github.com/bazelbuild/buildtools/issues/576 warnings: -function-docstring-args
Use yaml anchors, aliases, and merge to avoid repetition.
Use yaml anchors, aliases, and merge to avoid repetition. RELNOTES: None. PiperOrigin-RevId: 252430973
YAML
apache-2.0
bazelbuild/rules_apple,bazelbuild/rules_apple,bazelbuild/rules_apple,bazelbuild/rules_apple
4fea6d16e0904fc4970e637cb0ce89d98afb73ab
packages/bi/bizzlelude.yaml
packages/bi/bizzlelude.yaml
homepage: http://github.com/TheBizzle changelog-type: '' hash: 81d83dde07d0a8164793272dfd76f84e821111c1f31c8f78577e55c598974587 test-bench-deps: {} maintainer: jason.bertsche@gmail.com synopsis: A lousy Prelude replacement by a lousy dude changelog: '' basic-deps: text: ! '>=1.2.2 && <1.3.0' base-noprelude: ! '>=4.8 && <5' containers: ! '>=0.5.7 && <0.6.0' directory: ! '>=1.3.0 && <1.4.0' all-versions: - '0.0.3' - '0.0.4' - '0.1.0' - '1.0.1' - '1.0.2' - '1.0.3' - '1.0.4' - '1.1.0' author: Jason Bertsche latest: '1.1.0' description-type: haddock description: A lousy Prelude replacement by a lousy dude. Imports common things while trying to avoid importing things could cause naming conflicts. Favors `Text` over `String`. Wants using `IO` to be painful. license-name: BSD3
homepage: http://github.com/TheBizzle changelog-type: '' hash: 20595839189ef17dd0d6992654dfb7fdb33af25e4bd467f792fd7e01160ee357 test-bench-deps: {} maintainer: jason.bertsche@gmail.com synopsis: A lousy Prelude replacement by a lousy dude changelog: '' basic-deps: text: ! '>=1.2.3 && <1.3.0' base-noprelude: ! '>=4.8 && <5' containers: ! '>=0.5.11 && <0.6.0' directory: ! '>=1.3.2 && <1.4.0' all-versions: - '0.0.3' - '0.0.4' - '0.1.0' - '1.0.1' - '1.0.2' - '1.0.3' - '1.0.4' - '1.1.0' - '1.2.0' author: Jason Bertsche latest: '1.2.0' description-type: haddock description: A lousy Prelude replacement by a lousy dude. Imports common things while trying to avoid importing things could cause naming conflicts. Favors `Text` over `String`. Wants using `IO` to be painful. license-name: BSD3
Update from Hackage at 2018-03-05T00:55:55Z
Update from Hackage at 2018-03-05T00:55:55Z
YAML
mit
commercialhaskell/all-cabal-metadata
8b780c59f065339cb1c07cb4e88413e2ae067109
.forestry/settings.yml
.forestry/settings.yml
--- new_page_extension: md auto_deploy: false admin_path: webhook_url: sections: - type: directory path: journal label: Journal create: documents match: "**/*.md" - type: directory path: projects label: Projects create: documents match: "**/*.md" upload_dir: uploads public_path: "/uploads" front_matter_path: '' use_front_matter_path: false file_template: ":filename:"
--- new_page_extension: md auto_deploy: false admin_path: webhook_url: sections: - type: directory path: journal label: Journal create: documents match: "**/*.md" - type: directory path: projects label: Projects create: documents match: "**/*.md" upload_dir: uploads public_path: "/uploads" front_matter_path: '' use_front_matter_path: false file_template: ":filename:" build: publish_command: gridsome build output_directory: dist
Update from Forestry.io - Updated Forestry configuration
Update from Forestry.io - Updated Forestry configuration
YAML
mit
wkadwaikar/wkadwaikar.github.io,gatsbimantico/blog,gatsbimantico/blog,wkadwaikar/wkadwaikar.github.io
0c58c7f24800d0bd20c5e93f3ce05cb10e9195d2
.github/dependabot.yml
.github/dependabot.yml
version: 2 updates: - package-ecosystem: maven directory: "/" schedule: interval: daily open-pull-requests-limit: 10
version: 2 updates: - package-ecosystem: maven directory: "/" schedule: interval: daily open-pull-requests-limit: 10 - package-ecosystem: "github-actions" directory: "/" schedule: interval: daily
Use Dependabot to update GitHub Actions
Use Dependabot to update GitHub Actions
YAML
apache-2.0
maxmind/minfraud-api-java,maxmind/minfraud-api-java
1f6ab8b04f5926e7fc715d0fcfc7f1234723e28f
packages/co/constraint.yaml
packages/co/constraint.yaml
homepage: '' changelog-type: '' hash: f199db7d27874e311233788db8d4462d3857f0060def6e1ba989a12dabae2445 test-bench-deps: {} maintainer: m.farkasdyck@gmail.com synopsis: Reified constraints changelog: '' basic-deps: base: ! '>=4.10 && <5' category: ! '>=0.1.2 && <0.3' all-versions: - 0.1.0.0 - 0.1.0.1 - 0.1.1.0 - 0.1.1.1 - 0.1.2.0 - 0.1.2.1 author: M Farkas-Dyck latest: 0.1.2.1 description-type: haddock description: '' license-name: BSD-3-Clause
homepage: '' changelog-type: '' hash: 0467251601d0c10639d774052f3953f96c4f453501bb96a48d36f72dc039db98 test-bench-deps: {} maintainer: m.farkasdyck@gmail.com synopsis: Reified constraints changelog: '' basic-deps: base: ! '>=4.10 && <5' category: ! '>=0.1.2 && <0.3' unconstrained: ! '>=0.1 && <0.2' all-versions: - 0.1.0.0 - 0.1.0.1 - 0.1.1.0 - 0.1.1.1 - 0.1.2.0 - 0.1.2.1 - 0.1.3.0 author: M Farkas-Dyck latest: 0.1.3.0 description-type: haddock description: '' license-name: BSD-3-Clause
Update from Hackage at 2018-12-30T18:20:23Z
Update from Hackage at 2018-12-30T18:20:23Z
YAML
mit
commercialhaskell/all-cabal-metadata
3e8cb619b8aeb7d7d858081ac6681cb65034e2bb
tasks/main.yml
tasks/main.yml
# SPDX-License-Identifier: BSD-3-Clause - name: Set version specific variables include_vars: "{{ item }}" with_first_found: - "{{ ansible_distribution }}-{{ ansible_distribution_major_version }}.yml" - "{{ ansible_distribution }}.yml" - "{{ ansible_os_family }}.yml" - "default.yml" - name: Install packages package: name: "{{ network_packages }}" state: present - name: Enable network service service: name: "{{ network_service_name }}" state: started enabled: yes - name: Configure networking connection profiles network_connections: provider: "{{ network_provider | mandatory }}" ignore_errors: "{{ network_ignore_errors | default(omit) }}" force_state_change: "{{ network_force_state_change | default(omit) }}" connections: "{{ network_connections | default([]) }}" - name: Re-test connectivity ping:
# SPDX-License-Identifier: BSD-3-Clause - name: Set version specific variables include_vars: "{{ item }}" with_first_found: - "{{ ansible_distribution }}-{{ ansible_distribution_major_version }}.yml" - "{{ ansible_distribution }}.yml" - "{{ ansible_os_family }}.yml" - "default.yml" # needed for ansible_facst.packages - name: Check which packages are installed package_facts: # Depending on the plugins, checking installed packages might be slow # for example subscription manager might slow this down # Therefore install packages only when rpm does not find them - name: Install packages package: name: "{{ network_packages }}" state: present when: - not network_packages is subset(ansible_facts.packages.keys()) - name: Enable network service service: name: "{{ network_service_name }}" state: started enabled: yes - name: Configure networking connection profiles network_connections: provider: "{{ network_provider | mandatory }}" ignore_errors: "{{ network_ignore_errors | default(omit) }}" force_state_change: "{{ network_force_state_change | default(omit) }}" connections: "{{ network_connections | default([]) }}" - name: Re-test connectivity ping:
Work around slow package manager
Work around slow package manager
YAML
bsd-3-clause
linux-system-roles/network,linux-system-roles/network,linux-system-roles/network
e2cfe082a35a1c34db40cc088a63fbfadc295f46
tasks/main.yml
tasks/main.yml
--- # tasks file for ndenv - name: Run specified tasks on current platform include: "{{ ansible_os_family }}.yml" - block: - name: Check whether ndenv is installed shell: "{{ ndenv_login_shell }} -lc 'ndenv versions'" register: result_ndenv_versions ignore_errors: yes changed_when: no - name: Install ndenv shell: "{{ ndenv_login_shell }} -lc 'anyenv install ndenv'" when: "result_ndenv_versions.rc != 0" - name: Install nodejs shell: "{{ ndenv_login_shell }} -lc 'ndenv install {{ item }}'" when: "item not in result_ndenv_versions.stdout" with_items: "{{ ndenv_node_versions }}" - name: Check default version shell: "{{ ndenv_login_shell }} -lc 'ndenv global'" register: result_ndenv_global changed_when: no when: "ndenv_global_version is defined" - name: Set default version shell: "{{ ndenv_login_shell }} -lc 'ndenv global {{ ndenv_global_version }}'" when: "ndenv_global_version is defined and ndenv_global_version not in result_ndenv_global.stdout" become: yes become_user: "{{ anyenv_owner }}" # Ref. https://github.com/ansible/ansible/pull/13424 remote_user: "{{ anyenv_owner }}"
--- # tasks file for ndenv - name: Run specified tasks on current platform include: "{{ ansible_os_family }}.yml" - block: - name: Check whether ndenv is installed shell: "su -s {{ ndenv_login_shell }} -c 'ndenv versions' - {{ anyenv_owner }}" register: result_ndenv_versions ignore_errors: yes changed_when: no - name: Install ndenv shell: "su -s {{ ndenv_login_shell }} -c 'anyenv install ndenv' - {{ anyenv_owner }}" when: "result_ndenv_versions.rc != 0" - name: Install nodejs shell: "su -s {{ ndenv_login_shell }} -c 'ndenv install {{ item }}' - {{ anyenv_owner }}" when: "item not in result_ndenv_versions.stdout" with_items: "{{ ndenv_node_versions }}" - name: Check default version shell: "su -s {{ ndenv_login_shell }} -c 'ndenv global' - {{ anyenv_owner }}" register: result_ndenv_global changed_when: no when: "ndenv_global_version is defined" - name: Set default version shell: "su -s {{ ndenv_login_shell }} -c 'ndenv global {{ ndenv_global_version }}' - {{ anyenv_owner }}" when: "ndenv_global_version is defined and ndenv_global_version not in result_ndenv_global.stdout" become: yes
Refactor tasks to switch user without remote_user.
Refactor tasks to switch user without remote_user. - In several cases, remote_user can't be overwritten in this role. - Following warning occurs, but become_method doesn't work well on container. - When become_method=sudo(default), almost containers don't have sudo command. - When become_method=su, docker connection plugin of Ansible doesn't allow 'su' as become_method. ``` [WARNING]: Consider using 'become', 'become_method', and 'become_user' rather than running su ```
YAML
mit
FGtatsuro/ansible-ndenv
73a1e8e3f9fd6ee04becde0b0650de43d36a23c4
tasks/main.yml
tasks/main.yml
--- # tasks file for consul-template - name: Write install script template: src=install.j2 dest={{consul_template_install_script}} mode=0755 tags: - consul-template - name: Check consul-template version shell: consul-template --version 2>&1 | grep {{consul_template_ver}} failed_when: false changed_when: false register: consul_template_ver_matches tags: - consul-template - name: Install consul-template shell: "{{consul_template_install_script}}" when: consul_template_ver_matches.rc != 0 tags: - consul-template - name: Verify binary stat: path: "{{consul_template_bin_path}}//{{consul_template_bin_name}}" register: consul_template_verify_binary - name: Fail if binary is not present fail: msg: consul-template binary is not present when: not consul_template_verify_binary.stat.exists - name: Make templates dir file: path={{consul_template_templates_dir}} state=directory tags: - consul-template
--- # tasks file for consul-template - name: Write install script template: src=install.j2 dest={{consul_template_install_script}} mode=0755 tags: - consul-template - name: Check consul-template version shell: consul-template --version 2>&1 | grep {{consul_template_ver}} failed_when: false changed_when: false register: consul_template_ver_matches tags: - consul-template - name: Install consul-template shell: "{{consul_template_install_script}}" when: consul_template_ver_matches.rc != 0 tags: - consul-template - name: Verify binary stat: path: "{{consul_template_bin_path}}//{{consul_template_bin_name}}" register: consul_template_verify_binary tags: - consul-template - name: Fail if binary is not present fail: msg: consul-template binary is not present when: not consul_template_verify_binary.stat.exists tags: - consul-template - name: Make templates dir file: path={{consul_template_templates_dir}} state=directory tags: - consul-template
Add tag to all tasks
Add tag to all tasks
YAML
mit
mtchavez/ansible-consul-template
458de30c4f40050bbe52248f711bd6b955a30a32
packages/gl/glob-posix.yaml
packages/gl/glob-posix.yaml
homepage: https://github.com/rdnetto/glob-posix#readme changelog-type: '' hash: 8c87efe365f523a59422bcf80f157a2b59887d085d94021ce0579a3d1679ec23 test-bench-deps: MissingH: -any unix: -any base: -any tasty-expected-failure: -any filepath: -any criterion: -any glob-posix: -any tasty-hunit: -any tasty: -any Glob: -any directory: -any maintainer: rdnetto@gmail.com synopsis: Haskell bindings for POSIX glob library. changelog: '' basic-deps: base: ! '>=4.7 && <5' all-versions: - 0.1.0.0 - 0.1.0.1 - 0.2.0.0 author: R D'Netto latest: 0.2.0.0 description-type: haddock description: |- Wrapper for the glob(3) function. The key functions are glob and globMany. GNU extensions are supported but contained in a different module to encourage portability. license-name: LicenseRef-Apache
homepage: https://github.com/rdnetto/glob-posix#readme changelog-type: '' hash: 438df35fc614ea9d355ec1390046d7abcc8afabf23f8e589abc2c9ef54b705be test-bench-deps: MissingH: -any unix: -any base: -any tasty-expected-failure: -any filepath: -any criterion: -any glob-posix: -any tasty-hunit: -any tasty: -any Glob: -any directory: -any maintainer: rdnetto@gmail.com synopsis: Haskell bindings for POSIX glob library. changelog: '' basic-deps: base: ! '>=4.7 && <5' all-versions: - 0.1.0.0 - 0.1.0.1 - 0.2.0.0 - 0.2.0.1 author: R D'Netto latest: 0.2.0.1 description-type: haddock description: |- Wrapper for the glob(3) function. The key functions are glob and globMany. GNU extensions are supported but contained in a different module to encourage portability. license-name: LicenseRef-Apache
Update from Hackage at 2019-03-03T05:43:42Z
Update from Hackage at 2019-03-03T05:43:42Z
YAML
mit
commercialhaskell/all-cabal-metadata
4ad78cf014261e4ded2bf85799ce39466e46c5d4
build/ci/macos.yml
build/ci/macos.yml
parameters: mode: 'debug' sanitize: '' jobs: - job: 'macOS_${{ parameters.mode }}' displayName: 'macOS (${{ parameters.mode }})' variables: - name: CCACHE_DIR value: $(Build.SourcesDirectory)/ccache pool: vmImage: macOS-11 steps: - script: | brew install ccache displayName: 'Dependencies' - template: setup.yml - template: cache.yml parameters: mode: ${{ parameters.mode }} - script: | echo "##vso[task.setvariable variable=config]mode=${{ parameters.mode }} sanitize=${{ parameters.sanitize }} XCODE=/Applications/Xcode_13.0.app/Contents/Developer" displayName: 'Configure' - ${{ if eq(parameters.mode, 'debug') }}: - script: | make -j $(sysctl -n hw.ncpu) $(config) tests displayName: 'Test' - ${{ if eq(parameters.mode, 'release') }}: - script: | make -j $(sysctl -n hw.ncpu) $(config) libfly displayName: 'Build' - template: package.yml parameters: contents: 'build/release/clang/x64/etc/libfly-*.tar.bz2'
parameters: mode: 'debug' sanitize: '' jobs: - job: 'macOS_${{ parameters.mode }}' displayName: 'macOS (${{ parameters.mode }})' variables: - name: CCACHE_DIR value: $(Build.SourcesDirectory)/ccache pool: vmImage: macOS-11 steps: - script: | brew install ccache displayName: 'Dependencies' - template: setup.yml - template: cache.yml parameters: mode: ${{ parameters.mode }} - script: | echo "##vso[task.setvariable variable=config]mode=${{ parameters.mode }} sanitize=${{ parameters.sanitize }}" displayName: 'Configure' - ${{ if eq(parameters.mode, 'debug') }}: - script: | make -j $(sysctl -n hw.ncpu) $(config) tests displayName: 'Test' - ${{ if eq(parameters.mode, 'release') }}: - script: | make -j $(sysctl -n hw.ncpu) $(config) libfly displayName: 'Build' - template: package.yml parameters: contents: 'build/release/clang/x64/etc/libfly-*.tar.bz2'
Use default Xcode version on macOS CI
Use default Xcode version on macOS CI At one point, Xcode 13 wasn't the default.
YAML
mit
trflynn89/libfly,trflynn89/libfly
1cbdd28e57c4ff15785520dbfdc60a1fb6664fb3
tasks/main.yml
tasks/main.yml
--- # Tasks file for jenkins - name: Load the OS specific varibles include_vars: "{{ role_path }}/vars/{{ ansible_os_family }}.yml" - include: install_debian.yml tags: - install - jenkins when: ansible_os_family == "Debian" - name : Restart jenkins if new version installed include : "{{ role_path }}/handlers/restart_jenkins.yml" when : jenkins_task_package_install.changed - name : Waiting the application is available include : "{{ role_path }}/handlers/waiting_jenkins.yml" - include: configure.yml tags: - configure - jenkins - include: manage_cli.yml tags: - jenkins - install - include: plugins.yml tags: - jenkins - install
--- # Tasks file for jenkins - name: Load the OS specific varibles include_vars: "{{ role_path }}/vars/{{ ansible_os_family }}.yml" - include: install_debian.yml tags: - install - jenkins when: ansible_os_family == "Debian" - name : Restart jenkins if new version installed include : "{{ role_path }}/handlers/restart_jenkins.yml" when : jenkins_task_package_install.changed - name : Waiting the application is available include : "{{ role_path }}/handlers/waiting_jenkins.yml" - name : Configure jenkins launching settings include: configure.yml tags: - configure - jenkins - name : Install and configure Jenkins CLI include: manage_cli.yml tags: - jenkins - install - name : Manage plugins installations and upgrades include: plugins.yml tags: - jenkins - install
Add name key for included files
Add name key for included files
YAML
mit
infOpen/ansible-role-jenkins
f46121a22ab430ff0aa65361daede339a6391850
config/application.example.yml
config/application.example.yml
development: &base session_secret: xxx database_url: 'postgres://localhost/app_prototype_development' assets_server_url: "http://localhost:1234" precompiled_assets: false precompiled_assets_host: app_mailer_from_email: bugsnag_api_key: test: <<: *base database_url: 'postgres://localhost/app_prototype_test' production: <<: *base
development: &base session_secret: xxx database_url: 'postgres://localhost/app_prototype_development' assets_server_url: "http://localhost:1234" precompiled_assets: false precompiled_assets_host: app_mailer_from_email: bugsnag_api_key: postmark_api_key: test: <<: *base database_url: 'postgres://localhost/app_prototype_test' postmark_api_key: 'POSTMARK_API_TEST' production: <<: *base
Use a dummy Postmark API key in the test environment
Use a dummy Postmark API key in the test environment
YAML
mit
icelab/rodakase-skeleton,icelab/dry-web-skeleton,icelab/rodakase-skeleton,icelab/dry-web-skeleton,icelab/rodakase-skeleton,icelab/dry-web-skeleton
0f58621ca23ee51d0e02253917993d2c745e6b51
packages/ve/vec-optics.yaml
packages/ve/vec-optics.yaml
homepage: https://github.com/phadej/vec changelog-type: markdown hash: 1962a9943f8168ccee95c954c378b10eee884248ed9b05db5fa4e7e0c491853a test-bench-deps: {} maintainer: Oleg.Grenrus <oleg.grenrus@iki.fi> synopsis: ! 'Vec: length-indexed (sized) list: optics support' changelog: | # Revision history for vec ## 0.3 - Split out of `vec` basic-deps: optics-core: ^>=0.2 base: ! '>=4.9 && <4.14' vec: ^>=0.3 fin: ^>=0.1.1 all-versions: - '0.3' author: Oleg Grenrus <oleg.grenrus@iki.fi> latest: '0.3' description-type: haddock description: |- This package provides [optics](https://hackage.haskell.org/package/optics) and instances for data types in [vec](https://hackage.haskell.org/package/vec) package. "Data.Vec.Optics.Instances" provides all the instances. the other modules provide some named optics too. license-name: BSD-3-Clause
homepage: https://github.com/phadej/vec changelog-type: markdown hash: 60d786949e94b2812b4a5cc89080166dc8c5362939748a349a71414cc6b04b81 test-bench-deps: {} maintainer: Oleg.Grenrus <oleg.grenrus@iki.fi> synopsis: 'Vec: length-indexed (sized) list: optics support' changelog: | # Revision history for vec ## 0.3 - Split out of `vec` basic-deps: optics-core: '>=0.2 && <0.4' base: '>=4.9 && <4.15' vec: ^>=0.3 fin: ^>=0.1.1 all-versions: - '0.3' author: Oleg Grenrus <oleg.grenrus@iki.fi> latest: '0.3' description-type: haddock description: |- This package provides [optics](https://hackage.haskell.org/package/optics) and instances for data types in [vec](https://hackage.haskell.org/package/vec) package. "Data.Vec.Optics.Instances" provides all the instances. the other modules provide some named optics too. license-name: BSD-3-Clause
Update from Hackage at 2020-04-16T07:50:15Z
Update from Hackage at 2020-04-16T07:50:15Z
YAML
mit
commercialhaskell/all-cabal-metadata
e226cade79cdd40afec89e8c540339476661af76
packages/wa/wai-thrift.yaml
packages/wa/wai-thrift.yaml
homepage: https://github.com/yogeshsajanikar/wai-thrift changelog-type: '' hash: 5310da25fa2299cb667a796abc6176292236057193c49dcf60c04f515f371d12 test-bench-deps: {} maintainer: yogesh_sajanikar@yahoo.com synopsis: Thrift transport layer for Wai changelog: '' basic-deps: bytestring: -any wai: -any stm: -any base: ! '>=4.8 && <4.9' blaze-builder: -any thrift: -any http-types: -any all-versions: - '0.0.1.0' - '0.0.1.1' author: Yogesh Sajanikar latest: '0.0.1.1' description-type: haddock description: ! 'Implements a read-only transport layer for Wai Request, and write-only transport layer for Wai streaming body.' license-name: MIT
homepage: https://github.com/yogeshsajanikar/wai-thrift changelog-type: '' hash: f776d8cac8165b0099030013a75000c52d1f86fc8788b29f19fdfa99f927f0b8 test-bench-deps: {} maintainer: yogesh_sajanikar@yahoo.com synopsis: Thrift transport layer for Wai changelog: '' basic-deps: bytestring: ! '>=0.10' wai: ! '>=3.0' base: ! '>=4.8 && <4.9' blaze-builder: ! '>=0.4' thrift: ==0.9.3 http-types: ! '>=0.8' all-versions: - '0.0.1.0' - '0.0.1.1' - '0.0.1.2' - '0.0.1.3' author: Yogesh Sajanikar latest: '0.0.1.3' description-type: haddock description: ! 'Implements a read-only transport layer for Wai Request, and write-only transport layer for Wai streaming body.' license-name: MIT
Update from Hackage at 2015-10-21T16:37:19+0000
Update from Hackage at 2015-10-21T16:37:19+0000
YAML
mit
commercialhaskell/all-cabal-metadata
8d6835987baa0fba2ea50de5d297ef3adebc7d79
packages/ye/yesod-auth-hmac-keccak.yaml
packages/ye/yesod-auth-hmac-keccak.yaml
homepage: '' changelog-type: markdown hash: cbb31fadf64e20903d026b879e30ee273392b3634b6838626969e258cab8e45d test-bench-deps: {} maintainer: nek0@chelnok.de synopsis: ! 'An account authentication plugin for yesod with encrypted token transfer.' changelog: ! '# Revision history for yesod-auth-hmac-keccak ## 0.0.0.0 -- YYYY-mm-dd * First version. Released on an unsuspecting world. ' basic-deps: shakespeare: -any bytestring: -any yesod-core: -any base: ! '>=4.9 && <4.11' persistent: -any text: -any yesod-persistent: -any yesod-form: -any cryptonite: -any mtl: -any random: -any aeson: -any yesod-static: -any yesod-auth: -any all-versions: - '0.0.0.1' - '0.0.0.2' - '0.0.0.3' author: nek0 latest: '0.0.0.3' description-type: haddock description: ! 'This authentication plugin for Yesod uses a challenge-response authentication mechanism, where the password is encrypted via HMAC- Keccak on the client side.' license-name: MIT
homepage: '' changelog-type: markdown hash: a6424cb27dbdfce09e6d7563ea7d72bc81c2dd3eba633d5a2c5756a494cc2c12 test-bench-deps: {} maintainer: nek0@chelnok.de synopsis: ! 'An account authentication plugin for yesod with encrypted token transfer.' changelog: ! '# Revision history for yesod-auth-hmac-keccak ## 0.0.0.0 -- YYYY-mm-dd * First version. Released on an unsuspecting world. ' basic-deps: shakespeare: -any bytestring: -any yesod-core: -any base: ! '>=4.9 && <4.11' persistent: -any text: -any yesod-persistent: -any yesod-form: -any cryptonite: -any mtl: -any random: -any aeson: -any yesod-static: -any yesod-auth: -any all-versions: - '0.0.0.1' - '0.0.0.2' - '0.0.0.3' - '0.0.0.4' author: nek0 latest: '0.0.0.4' description-type: haddock description: ! 'This authentication plugin for Yesod uses a challenge-response authentication mechanism, where the password is encrypted via HMAC- Keccak on the client side.' license-name: MIT
Update from Hackage at 2017-12-24T07:01:21Z
Update from Hackage at 2017-12-24T07:01:21Z
YAML
mit
commercialhaskell/all-cabal-metadata
98ad481fbf5eb925fe6569719a4c32ce1febf22f
packages/pa/parsec-free.yaml
packages/pa/parsec-free.yaml
homepage: https://github.com/jwiegley/parsec-free changelog-type: '' hash: b6ecc93a03f98d6e062ac00a50d3616cbd9bff40d3b899d9f2b7c45710b19552 test-bench-deps: {} maintainer: johnw@newartisans.com synopsis: Parsec API encoded as a deeply-embedded DSL, for debugging and analysis changelog: '' basic-deps: free: -any bytestring: -any base: ! '>=4 && <5' text: ! '>=0.2 && <1.3' parsec: ! '>=3.1.11' mtl: -any transformers: ! '>=0.4 && <0.6' all-versions: - '3.1.11.1' - '3.1.11.2' - '3.1.11.3' author: John Wiegley latest: '3.1.11.3' description-type: haddock description: Parsec API encoded as a deeply-embedded DSL, for debugging and analysis license-name: MIT
homepage: https://github.com/jwiegley/parsec-free changelog-type: '' hash: 900acd53ec3d2610c3d8ad8d5cbcb6f61a458bd597c48cc9d0dcbc02ed7d33d5 test-bench-deps: {} maintainer: johnw@newartisans.com synopsis: Parsec API encoded as a deeply-embedded DSL, for debugging and analysis changelog: '' basic-deps: free: -any bytestring: -any base: ! '>=4 && <5' text: ! '>=0.2 && <1.3' parsec: ! '>=3.1.11' containers: -any lens: -any mtl: -any transformers: ! '>=0.4 && <0.6' all-versions: - '3.1.11.1' - '3.1.11.2' - '3.1.11.3' - '3.1.11.4' author: John Wiegley latest: '3.1.11.4' description-type: haddock description: Parsec API encoded as a deeply-embedded DSL, for debugging and analysis license-name: MIT
Update from Hackage at 2016-10-11T00:02:03+00:00
Update from Hackage at 2016-10-11T00:02:03+00:00
YAML
mit
commercialhaskell/all-cabal-metadata
08e8d79d689fb393468fcabb63f7dcee25efa0e3
packages/pr/primes-type.yaml
packages/pr/primes-type.yaml
homepage: https://github.com/kindaro/primes-type#readme changelog-type: '' hash: 71f9a403061697291463bb459197e8e797a9fc7dc1a18b69eee1be4035b6d4f0 test-bench-deps: HTF: ! '>=0.13 && <0.14' base: ! '>=4.8 && <5' primes: ! '>=0.2 && <0.3' primes-type: -any maintainer: kindaro@gmail.com synopsis: Type-safe prime numbers. changelog: '' basic-deps: base: ! '>=4.8 && <4.11' primes: ! '>=0.2 && <0.3' all-versions: - '0.2.0.0' - '0.2.0.1' author: Ignat Insarov latest: '0.2.0.1' description-type: haddock description: This library provides type safe prime numbers. The idea is based upon the concept of a predicate type from type theory. license-name: ISC
homepage: https://github.com/kindaro/primes-type#readme changelog-type: '' hash: 1c389071fbe4bd78d0bb2ac205b69f88425ebd859bbdf79482ffa08dcd439d6c test-bench-deps: HTF: ! '>=0.13 && <0.14' base: ! '>=4.8 && <5' primes: ==0.2.* criterion: -any primes-type: -any maintainer: kindaro@gmail.com synopsis: Type-safe prime numbers. changelog: '' basic-deps: base: ! '>=4.8 && <4.11' primes: ! '>=0.2 && <0.3' deepseq: -any all-versions: - '0.2.0.0' - '0.2.0.1' - '0.2.0.2' author: Ignat Insarov latest: '0.2.0.2' description-type: haddock description: This library provides type safe prime numbers. The idea is based upon the concept of a predicate type from type theory. license-name: ISC
Update from Hackage at 2017-10-06T00:12:43Z
Update from Hackage at 2017-10-06T00:12:43Z
YAML
mit
commercialhaskell/all-cabal-metadata
b2b5fc1c897736c83ecc18bf59188b42ac887243
packages/sn/snap-stream.yaml
packages/sn/snap-stream.yaml
homepage: '' changelog-type: '' hash: acae39c1ad385b96de141a5264ef7c335e3b34d3d718d2cdaf348494b8f02448 test-bench-deps: {} maintainer: maintainer@obsidian.systems synopsis: Streaming Snap handlers changelog: '' basic-deps: bytestring: ==0.10.* base: ! '>=4.9 && <4.12' io-streams: ! '>=1.4 && <1.6' snap-core: ==1.0.* attoparsec: ==0.13.* all-versions: - '0.1' author: Obsidian Systems LLC latest: '0.1' description-type: haddock description: Snap handlers for streaming access with range requests license-name: BSD-3-Clause
homepage: '' changelog-type: markdown hash: 2096c688612cb1bef40a51d6dabfc1cd3304dcff546f1d423706d4dc7999234e test-bench-deps: {} maintainer: maintainer@obsidian.systems synopsis: Streaming Snap handlers changelog: | # Changelog for snap-stream ## 0.1.1.0 * Loosen `base` version bounds ## 0.1 * Initial release: Snap handlers for streaming access with range requests basic-deps: bytestring: '>=0.10 && <0.11' base: '>=4.9 && <4.15' io-streams: '>=1.4 && <1.6' snap-core: '>=1.0 && <1.1' attoparsec: '>=0.13 && <0.14' all-versions: - '0.1' - 0.1.1.0 author: Obsidian Systems LLC latest: 0.1.1.0 description-type: markdown description: | # snap-stream [Snap](https://hackage.haskell.org/package/snap) handlers for HTTP [range requests](https://developer.mozilla.org/en-US/docs/Web/HTTP/Range_requests). This package provides a way to respond to an HTTP request for a particular range of bytes. Range requests include a header that specifies which bytes the client is requesting (e.g., `Range: bytes=0-1023`). See the documentation for `serveStreamAs`. license-name: BSD-3-Clause
Update from Hackage at 2021-11-16T13:26:18Z
Update from Hackage at 2021-11-16T13:26:18Z
YAML
mit
commercialhaskell/all-cabal-metadata
eeced700efef1d5b43b3513d67f41cdab93f9b1b
packages/ve/vector-text.yaml
packages/ve/vector-text.yaml
homepage: https://github.com/luna/vector-text changelog-type: '' hash: 56eb6d61eb0b8b7ea8529c80f04faea5dae5922f7790b17ebf1b087cf2763f1d test-bench-deps: {} maintainer: Wojciech Danilo <wojciech.danilo@luna-lang.org> synopsis: Text implementation based on unboxed char vector. changelog: '' basic-deps: base: ! '>=4.10 && <4.12' text: -any prologue: -any vector-binary-instances: -any binary: -any vector: -any all-versions: - '1.1' - '1.1.1' - '1.1.2' - '1.1.3' author: Luna Team latest: '1.1.3' description-type: haddock description: '' license-name: Apache-2.0
homepage: https://github.com/luna/vector-text changelog-type: '' hash: 1db139021dae0cf901e2d920f59ad8321019bda7718f5e1245c968bb9169f7fa test-bench-deps: {} maintainer: Wojciech Danilo <wojciech.danilo@luna-lang.org> synopsis: Text implementation based on unboxed char vector. changelog: '' basic-deps: base: ! '>=4.10 && <4.12' text: -any prologue: -any vector-binary-instances: -any binary: -any vector: -any all-versions: - '1.1' - '1.1.1' - '1.1.2' - '1.1.3' - '1.1.5' author: Luna Team latest: '1.1.5' description-type: haddock description: '' license-name: Apache-2.0
Update from Hackage at 2018-08-24T10:06:01Z
Update from Hackage at 2018-08-24T10:06:01Z
YAML
mit
commercialhaskell/all-cabal-metadata
166e2fc5c03dd8e263dd2db32dd1739e25a27ddf
packages/in/interpolate.yaml
packages/in/interpolate.yaml
homepage: https://github.com/sol/interpolate#readme changelog-type: '' hash: bc243859cb5489b5ccd1e771714f79fc801bbd209c6c69b2059156c8f3dcb1ae test-bench-deps: bytestring: -any base: ==4.* hspec: ! '>=1.5' text: -any quickcheck-instances: -any base-compat: -any QuickCheck: -any haskell-src-meta: ! '>=0.8' template-haskell: -any maintainer: Simon Hengel <sol@typeful.net> synopsis: String interpolation done right changelog: '' basic-deps: base: ==4.* haskell-src-meta: ! '>=0.8' template-haskell: -any all-versions: - 0.0.0 - 0.0.1 - 0.0.2 - 0.0.3 - 0.0.4 - 0.1.0 - 0.1.1 - 0.2.0 author: Simon Hengel <sol@typeful.net> latest: 0.2.0 description-type: haddock description: String interpolation done right license-name: MIT
homepage: https://github.com/sol/interpolate#readme changelog-type: '' hash: e102d1870811afce16a339fc47f75cef6fc093bf931453569f326669186fb1db test-bench-deps: bytestring: -any base: ==4.* hspec: '>=1.5' text: -any quickcheck-instances: -any base-compat: -any QuickCheck: -any haskell-src-meta: '>=0.8' template-haskell: -any maintainer: Simon Hengel <sol@typeful.net> synopsis: String interpolation done right changelog: '' basic-deps: base: ==4.* haskell-src-meta: '>=0.8' template-haskell: -any all-versions: - 0.0.0 - 0.0.1 - 0.0.2 - 0.0.3 - 0.0.4 - 0.1.0 - 0.1.1 - 0.2.0 - 0.2.1 author: Simon Hengel <sol@typeful.net> latest: 0.2.1 description-type: haddock description: String interpolation done right license-name: MIT
Update from Hackage at 2020-06-05T11:52:20Z
Update from Hackage at 2020-06-05T11:52:20Z
YAML
mit
commercialhaskell/all-cabal-metadata
5fa02622ff1d83109715a856d5ecb7f6d9faa3be
fig.yml
fig.yml
db: image: orchardup/postgresql ports: - "5432" web: build: . command: gin volumes: - .:/go/src/github.com/hackedu/backend ports: - "3000:3000" links: - db
db: image: orchardup/postgresql ports: - "5432" redis: image: orchardup/redis web: build: . command: gin volumes: - .:/go/src/github.com/hackedu/backend ports: - "3000:3000" links: - db
Add Redis service to Fig.
Add Redis service to Fig.
YAML
mit
zachlatta/website,zachlatta/website,hackedu/website,zachlatta/website,hackedu/website,hackedu/website
743d6782e1ef8b8585b1daaabab82742c28577fb
playbooks/group_vars/all.yml
playbooks/group_vars/all.yml
# Copyright (c) 2015-2018 Agalmic Ventures LLC (www.agalmicventures.com) # # Permission is hereby granted, free of charge, to any person obtaining a copy # of this software and associated documentation files (the "Software"), to deal # in the Software without restriction, including without limitation the rights # to use, copy, modify, merge, publish, distribute, sublicense, and/or sell # copies of the Software, and to permit persons to whom the Software is # furnished to do so, subject to the following conditions: # # The above copyright notice and this permission notice shall be included in # all copies or substantial portions of the Software. # # THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR # IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, # FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE # AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER # LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, # OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE # SOFTWARE. is_digital_ocean: ansible_system_vendor == 'DigitalOcean' #NOTE: This check is to work around issues where ansible_virtualization_role is #incorrectly identified is_virtual_machine: ansible_virtualization_role == 'guest' or is_digital_ocean is_physical_machine: not is_virtual_machine
# Copyright (c) 2015-2018 Agalmic Ventures LLC (www.agalmicventures.com) # # Permission is hereby granted, free of charge, to any person obtaining a copy # of this software and associated documentation files (the "Software"), to deal # in the Software without restriction, including without limitation the rights # to use, copy, modify, merge, publish, distribute, sublicense, and/or sell # copies of the Software, and to permit persons to whom the Software is # furnished to do so, subject to the following conditions: # # The above copyright notice and this permission notice shall be included in # all copies or substantial portions of the Software. # # THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR # IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, # FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE # AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER # LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, # OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE # SOFTWARE. is_digital_ocean: "{{ ansible_system_vendor == 'DigitalOcean' }}" #NOTE: This check is to work around issues where ansible_virtualization_role is #incorrectly identified is_virtual_machine: "{{ ansible_virtualization_role == 'guest' or is_digital_ocean }}" is_physical_machine: "{{ not is_virtual_machine }}"
Fix missing interpolation in Ansible group_vars
Fix missing interpolation in Ansible group_vars
YAML
mit
AgalmicVentures/Environment,AgalmicVentures/Environment,AgalmicVentures/Environment
6ae8e9cc8e715eb3fd7dacb4478ef4066e76b4a7
readthedocs-environment.yml
readthedocs-environment.yml
channels: - conda-forge dependencies: - python=3.7 - ipython - numpy - numba - scipy - sparse - matplotlib - pandoc - ipykernel - nbsphinx - nbconvert=5.4.1
channels: - conda-forge dependencies: - python=3.7 - ipython - numpy - numba - scipy - sparse - matplotlib - pandoc - ipykernel - nbsphinx - nbconvert=5.4.1 - pip: - pyganja
Add pyganja dependency to ReadTheDocs yml config
Add pyganja dependency to ReadTheDocs yml config This was only added to the pip config, so didn't work when built in conda
YAML
bsd-3-clause
arsenovic/clifford,arsenovic/clifford
6b153ec01b1c27b258bcab9b89ba239ec2c46766
recipes/deeptools/meta.yaml
recipes/deeptools/meta.yaml
package: name: deeptools version: '2.0.0' source: fn: deepTools-2.0.0.tar.gz url: https://pypi.python.org/packages/source/d/deepTools/deepTools-2.0.0.tar.gz md5: 09821e73f56ddb744a35d01173c6eb29 build: number: 0 skip: True # [not py27] requirements: build: - python - setuptools - numpy x.x - scipy >=0.15.0 - matplotlib >=1.4.0 - pysam >=0.8.2 - bx-python >=0.7.2 run: - python - pybigwig >=0.2.3 - numpy x.x - scipy >=0.15.0 - matplotlib >=1.4.0 - pysam >=0.8.2 - bx-python >=0.7.2 test: imports: - deeptools commands: - nosetests --with-doctest deeptools requires: - nose about: home: https://github.com/fidelram/deepTools license: GPL3 summary: A set of user-friendly tools for normalization and visualzation of deep-sequencing data
package: name: deeptools version: '2.0.0' source: fn: deepTools-2.0.0.tar.gz url: https://pypi.python.org/packages/source/d/deepTools/deepTools-2.0.0.tar.gz md5: 09821e73f56ddb744a35d01173c6eb29 build: number: 0 skip: True # [not py27] requirements: build: - python - setuptools - numpy >=1.8.0 - scipy >=0.15.0 - matplotlib >=1.4.0 - pysam >=0.8.2 - bx-python >=0.7.2 run: - python - pybigwig >=0.2.3 - numpy >=1.8.0 - scipy >=0.15.0 - matplotlib >=1.4.0 - pysam >=0.8.2 - bx-python >=0.7.2 test: imports: - deeptools commands: - bamCompare --version about: home: https://github.com/fidelram/deepTools license: GPL3 summary: A set of user-friendly tools for normalization and visualzation of deep-sequencing data
Remove x.x from numpy and drop the nosetests in favor of just bamCompare --version
Remove x.x from numpy and drop the nosetests in favor of just bamCompare --version
YAML
mit
bebatut/bioconda-recipes,JenCabral/bioconda-recipes,bow/bioconda-recipes,joachimwolff/bioconda-recipes,JingchaoZhang/bioconda-recipes,acaprez/recipes,matthdsm/bioconda-recipes,cokelaer/bioconda-recipes,colinbrislawn/bioconda-recipes,roryk/recipes,saketkc/bioconda-recipes,keuv-grvl/bioconda-recipes,mdehollander/bioconda-recipes,blankenberg/bioconda-recipes,colinbrislawn/bioconda-recipes,HassanAmr/bioconda-recipes,zwanli/bioconda-recipes,ThomasWollmann/bioconda-recipes,CGATOxford/bioconda-recipes,guowei-he/bioconda-recipes,mcornwell1957/bioconda-recipes,chapmanb/bioconda-recipes,keuv-grvl/bioconda-recipes,hardingnj/bioconda-recipes,zwanli/bioconda-recipes,ivirshup/bioconda-recipes,daler/bioconda-recipes,gvlproject/bioconda-recipes,hardingnj/bioconda-recipes,mdehollander/bioconda-recipes,jasper1918/bioconda-recipes,abims-sbr/bioconda-recipes,ostrokach/bioconda-recipes,shenwei356/bioconda-recipes,CGATOxford/bioconda-recipes,gregvonkuster/bioconda-recipes,mcornwell1957/bioconda-recipes,BIMSBbioinfo/bioconda-recipes,ostrokach/bioconda-recipes,ivirshup/bioconda-recipes,instituteofpathologyheidelberg/bioconda-recipes,shenwei356/bioconda-recipes,xguse/bioconda-recipes,gvlproject/bioconda-recipes,BIMSBbioinfo/bioconda-recipes,abims-sbr/bioconda-recipes,lpantano/recipes,chapmanb/bioconda-recipes,chapmanb/bioconda-recipes,pinguinkiste/bioconda-recipes,jasper1918/bioconda-recipes,gvlproject/bioconda-recipes,acaprez/recipes,rob-p/bioconda-recipes,dmaticzka/bioconda-recipes,oena/bioconda-recipes,gvlproject/bioconda-recipes,rvalieris/bioconda-recipes,bow/bioconda-recipes,BIMSBbioinfo/bioconda-recipes,mdehollander/bioconda-recipes,mdehollander/bioconda-recipes,joachimwolff/bioconda-recipes,instituteofpathologyheidelberg/bioconda-recipes,joachimwolff/bioconda-recipes,Luobiny/bioconda-recipes,ivirshup/bioconda-recipes,bow/bioconda-recipes,gvlproject/bioconda-recipes,BIMSBbioinfo/bioconda-recipes,zachcp/bioconda-recipes,oena/bioconda-recipes,xguse/bioconda-recipes,saketkc/bioconda-recipes,zwanli/bioconda-recipes,keuv-grvl/bioconda-recipes,martin-mann/bioconda-recipes,dkoppstein/recipes,martin-mann/bioconda-recipes,chapmanb/bioconda-recipes,hardingnj/bioconda-recipes,rob-p/bioconda-recipes,HassanAmr/bioconda-recipes,pinguinkiste/bioconda-recipes,ostrokach/bioconda-recipes,ThomasWollmann/bioconda-recipes,oena/bioconda-recipes,guowei-he/bioconda-recipes,acaprez/recipes,CGATOxford/bioconda-recipes,saketkc/bioconda-recipes,matthdsm/bioconda-recipes,roryk/recipes,bow/bioconda-recipes,mcornwell1957/bioconda-recipes,keuv-grvl/bioconda-recipes,martin-mann/bioconda-recipes,pinguinkiste/bioconda-recipes,peterjc/bioconda-recipes,instituteofpathologyheidelberg/bioconda-recipes,phac-nml/bioconda-recipes,BIMSBbioinfo/bioconda-recipes,rvalieris/bioconda-recipes,peterjc/bioconda-recipes,BIMSBbioinfo/bioconda-recipes,HassanAmr/bioconda-recipes,gvlproject/bioconda-recipes,rvalieris/bioconda-recipes,dkoppstein/recipes,martin-mann/bioconda-recipes,xguse/bioconda-recipes,daler/bioconda-recipes,omicsnut/bioconda-recipes,blankenberg/bioconda-recipes,npavlovikj/bioconda-recipes,gregvonkuster/bioconda-recipes,mcornwell1957/bioconda-recipes,gregvonkuster/bioconda-recipes,saketkc/bioconda-recipes,joachimwolff/bioconda-recipes,bow/bioconda-recipes,JenCabral/bioconda-recipes,npavlovikj/bioconda-recipes,blankenberg/bioconda-recipes,JingchaoZhang/bioconda-recipes,chapmanb/bioconda-recipes,CGATOxford/bioconda-recipes,Luobiny/bioconda-recipes,cokelaer/bioconda-recipes,lpantano/recipes,phac-nml/bioconda-recipes,hardingnj/bioconda-recipes,JingchaoZhang/bioconda-recipes,cokelaer/bioconda-recipes,daler/bioconda-recipes,rvalieris/bioconda-recipes,bioconda/recipes,bioconda/bioconda-recipes,bioconda/bioconda-recipes,oena/bioconda-recipes,ivirshup/bioconda-recipes,zachcp/bioconda-recipes,abims-sbr/bioconda-recipes,omicsnut/bioconda-recipes,colinbrislawn/bioconda-recipes,xguse/bioconda-recipes,bioconda/bioconda-recipes,bioconda/recipes,ThomasWollmann/bioconda-recipes,colinbrislawn/bioconda-recipes,jasper1918/bioconda-recipes,acaprez/recipes,abims-sbr/bioconda-recipes,phac-nml/bioconda-recipes,HassanAmr/bioconda-recipes,pinguinkiste/bioconda-recipes,jfallmann/bioconda-recipes,pinguinkiste/bioconda-recipes,peterjc/bioconda-recipes,peterjc/bioconda-recipes,ostrokach/bioconda-recipes,npavlovikj/bioconda-recipes,dmaticzka/bioconda-recipes,zwanli/bioconda-recipes,zwanli/bioconda-recipes,yesimon/bioconda-recipes,saketkc/bioconda-recipes,colinbrislawn/bioconda-recipes,matthdsm/bioconda-recipes,oena/bioconda-recipes,zachcp/bioconda-recipes,bebatut/bioconda-recipes,rob-p/bioconda-recipes,joachimwolff/bioconda-recipes,rvalieris/bioconda-recipes,hardingnj/bioconda-recipes,xguse/bioconda-recipes,lpantano/recipes,pinguinkiste/bioconda-recipes,keuv-grvl/bioconda-recipes,cokelaer/bioconda-recipes,daler/bioconda-recipes,ThomasWollmann/bioconda-recipes,instituteofpathologyheidelberg/bioconda-recipes,yesimon/bioconda-recipes,keuv-grvl/bioconda-recipes,JingchaoZhang/bioconda-recipes,HassanAmr/bioconda-recipes,yesimon/bioconda-recipes,bioconda/recipes,dmaticzka/bioconda-recipes,jfallmann/bioconda-recipes,bebatut/bioconda-recipes,dmaticzka/bioconda-recipes,guowei-he/bioconda-recipes,shenwei356/bioconda-recipes,phac-nml/bioconda-recipes,zwanli/bioconda-recipes,guowei-he/bioconda-recipes,matthdsm/bioconda-recipes,blankenberg/bioconda-recipes,mdehollander/bioconda-recipes,bioconda/bioconda-recipes,phac-nml/bioconda-recipes,ivirshup/bioconda-recipes,jfallmann/bioconda-recipes,dmaticzka/bioconda-recipes,CGATOxford/bioconda-recipes,jasper1918/bioconda-recipes,guowei-he/bioconda-recipes,matthdsm/bioconda-recipes,ThomasWollmann/bioconda-recipes,saketkc/bioconda-recipes,npavlovikj/bioconda-recipes,dkoppstein/recipes,rob-p/bioconda-recipes,yesimon/bioconda-recipes,abims-sbr/bioconda-recipes,colinbrislawn/bioconda-recipes,dmaticzka/bioconda-recipes,ThomasWollmann/bioconda-recipes,JenCabral/bioconda-recipes,peterjc/bioconda-recipes,Luobiny/bioconda-recipes,peterjc/bioconda-recipes,daler/bioconda-recipes,zachcp/bioconda-recipes,lpantano/recipes,martin-mann/bioconda-recipes,bow/bioconda-recipes,bebatut/bioconda-recipes,ostrokach/bioconda-recipes,mcornwell1957/bioconda-recipes,instituteofpathologyheidelberg/bioconda-recipes,matthdsm/bioconda-recipes,mdehollander/bioconda-recipes,rvalieris/bioconda-recipes,jasper1918/bioconda-recipes,ivirshup/bioconda-recipes,Luobiny/bioconda-recipes,omicsnut/bioconda-recipes,joachimwolff/bioconda-recipes,roryk/recipes,HassanAmr/bioconda-recipes,ostrokach/bioconda-recipes,omicsnut/bioconda-recipes,daler/bioconda-recipes,shenwei356/bioconda-recipes,abims-sbr/bioconda-recipes,jfallmann/bioconda-recipes,JenCabral/bioconda-recipes,JenCabral/bioconda-recipes,CGATOxford/bioconda-recipes,gregvonkuster/bioconda-recipes,JenCabral/bioconda-recipes,omicsnut/bioconda-recipes
2f0e7257f62d84d1849a1d957e5c8cd8e5c8aa49
tests/acceptance.suite.yml
tests/acceptance.suite.yml
# Codeception Test Suite Configuration # # Suite for acceptance tests. # Perform tests in browser using the WebDriver or PhpBrowser. # If you need both WebDriver and PHPBrowser tests - create a separate suite. class_name: AcceptanceTester modules: enabled: - PhpBrowser: url: 'https://registry.dev/' # put your local url - Laravel5: part: ORM cleanup: false # can't wrap into transaction - \Helper\Acceptance
# Codeception Test Suite Configuration # # Suite for acceptance tests. # Perform tests in browser using the WebDriver or PhpBrowser. # If you need both WebDriver and PHPBrowser tests - create a separate suite. class_name: AcceptanceTester modules: enabled: - PhpBrowser: url: 'http://registry.dev/' # put your local url - Laravel5: part: ORM environment_file: .env.testing - \Helper\Acceptance
Update the codecept acceptance config
Update the codecept acceptance config
YAML
agpl-3.0
jonphipps/Metadata-Registry,jonphipps/Metadata-Registry,jonphipps/Metadata-Registry,jonphipps/Metadata-Registry,jonphipps/Metadata-Registry,jonphipps/Metadata-Registry,jonphipps/Metadata-Registry,jonphipps/Metadata-Registry,jonphipps/Metadata-Registry,jonphipps/Metadata-Registry
2866b7b10457946bb8575f744a79ed655102df0e
codeship-steps.yml
codeship-steps.yml
- name: download_cache service: cache command: true - name: trigger_build service: herokudeployment command: true - service: integrationtest command: deploy_to_heroku_app codeship-heroku-deployment - service: integrationtest command: heroku run --app codeship-heroku-deployment -- ls - service: herokudeployment tag: master type: push image_name: codeship/heroku-deployment registry: https://index.docker.io/v1/ encrypted_dockercfg_path: dockercfg.encrypted
- name: download_cache service: cache command: true - name: trigger_build service: herokudeployment command: true - service: integrationtest command: deploy_to_heroku_app codeship-heroku-deployment - service: integrationtest command: heroku run --app codeship-heroku-deployment -- true - service: herokudeployment tag: master type: push image_name: codeship/heroku-deployment registry: https://index.docker.io/v1/ encrypted_dockercfg_path: dockercfg.encrypted
Use true instead of ls as test command
Use true instead of ls as test command
YAML
mit
codeship-library/heroku-deployment,codeship-library/heroku-deployment
d109924b8700b582aba6ad8fe01ee6d9b9b7c6ca
release/tag/goreleaser.yaml
release/tag/goreleaser.yaml
# Copyright 2019 Google LLC # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. builds: - env: - CGO_ENABLED=0 - GO111MODULE=on goos: - darwin - linux - windows goarch: - amd64 ldflags: -s -w -X github.com/GoogleContainerTools/kpt/run.version={{.Version}} archives: - files: - LICENSE* - README* - CHANGELOG* - lib.zip* name_template: "{{ .ProjectName }}_{{ .Os }}_{{ .Arch }}_{{ .Version }}" checksum: name_template: 'checksums.txt' snapshot: name_template: "master" changelog: sort: asc filters: exclude: - '(?i)^docs:' - '(?i)^test:' - '(?i)^regenerate docs' - Merge pull request - Merge branch release: github: owner: GoogleContainerTools name: kpt
# Copyright 2019 Google LLC # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. builds: - env: - CGO_ENABLED=0 - GO111MODULE=on goos: - darwin - linux - windows goarch: - amd64 ldflags: -s -w -X github.com/GoogleContainerTools/kpt/run.version={{.Version}} archives: - files: - LICENSES* - lib.zip* name_template: "{{ .ProjectName }}_{{ .Os }}_{{ .Arch }}-{{ .Version }}" checksum: name_template: 'checksums.txt' snapshot: name_template: "master" changelog: sort: asc filters: exclude: - '(?i)^docs:' - '(?i)^test:' - '(?i)^regenerate docs' - Merge pull request - Merge branch release: github: owner: GoogleContainerTools name: kpt
Remove files from release artifacts and improve naming
Remove files from release artifacts and improve naming
YAML
apache-2.0
GoogleContainerTools/kpt,GoogleContainerTools/kpt,GoogleContainerTools/kpt,GoogleContainerTools/kpt,GoogleContainerTools/kpt
d7dcb41b271e6b5d5e74988a979a286185a0093e
tests/model/SiteTreeBrokenLinksTest.yml
tests/model/SiteTreeBrokenLinksTest.yml
Page: content: Title: ContentPage Content: 'This is some partially happy content. It has one missing a skiplink, but does have another <a name="yes-anchor-here">skiplink here</a>.' about: Title: About URLSegment: about Content: 'about us here <a name="yes-anchor-here">about skiplinks here</a>.' brokenInternalRedirector: RedirectionType: Internal Title: RedirectorPageToBrokenInteralPage LinkToID: 0 workingInternalRedirector: RedirectionType: Internal Title: RedirectorPageToBrokenInteralPage LinkTo: =>Page.content File: privacypolicy: Name: privacypolicy.pdf Title: privacypolicy.pdf Filename: assets/privacypolicy.pdf ErrorPage: 404: Title: Page not Found ErrorCode: 404
Page: content: Title: ContentPage Content: 'This is some partially happy content. It has one missing a skiplink, but does have another <a name="yes-anchor-here">skiplink here</a>.' about: Title: About URLSegment: about Content: 'about us here <a name="yes-anchor-here">about skiplinks here</a>.' brokenInternalRedirector: RedirectionType: Internal Title: RedirectorPageToBrokenInteralPage LinkToID: 0 workingInternalRedirector: RedirectionType: Internal Title: RedirectorPageToBrokenInteralPage LinkTo: =>Page.content File: privacypolicy: Name: privacypolicy.pdf Title: privacypolicy.pdf Filename: assets/privacypolicy.pdf ErrorPage: 404: Title: Page not Found ErrorCode: 404
Remove tab - invalid YAML
FIX: Remove tab - invalid YAML
YAML
bsd-3-clause
jonom/silverstripe-cms,jonom/silverstripe-cms,silverstripe/silverstripe-cms,silverstripe/silverstripe-cms
30899b1e9d8c213d4d10d352aee242a3d632fd14
.github/workflows/generate-documentation.yml
.github/workflows/generate-documentation.yml
name: Manually generate documentation on: workflow_dispatch permissions: contents: read pages: write id-token: write packages: read concurrency: group: "update-standard-library-documentation" cancel-in-progress: true jobs: generate-documentation: name: Generate documentation for release environment: name: github-pages url: ${{ steps.deployment.outputs.page_url }} runs-on: ubuntu-latest container: image: ghcr.io/ponylang/ponyc-ci-stdlib-builder:latest credentials: username: ${{ github.repository_owner }} password: ${{ secrets.GITHUB_TOKEN }} steps: - name: Checkout uses: actions/checkout@v3 - name: Build run: "bash .ci-scripts/build-stdlib-documentation.bash" - name: Setup Pages uses: actions/configure-pages@v2 - name: Upload artifact uses: actions/upload-pages-artifact@v1 with: path: 'stdlib-docs/site/' - name: Deploy to GitHub Pages id: deployment uses: actions/deploy-pages@v1
name: Manually generate documentation on: workflow_dispatch permissions: contents: read pages: write id-token: write packages: read concurrency: group: "update-standard-library-documentation" cancel-in-progress: true jobs: generate-documentation: name: Generate documentation for release environment: name: github-pages url: ${{ steps.deployment.outputs.page_url }} runs-on: ubuntu-latest container: image: ghcr.io/ponylang/ponyc-ci-stdlib-builder:release credentials: username: ${{ github.repository_owner }} password: ${{ secrets.GITHUB_TOKEN }} steps: - name: Checkout uses: actions/checkout@v3 - name: Build run: "bash .ci-scripts/build-stdlib-documentation.bash" - name: Setup Pages uses: actions/configure-pages@v2 - name: Upload artifact uses: actions/upload-pages-artifact@v1 with: path: 'stdlib-docs/site/' - name: Deploy to GitHub Pages id: deployment uses: actions/deploy-pages@v1
Put generated documentation step back to using release
Put generated documentation step back to using release
YAML
bsd-2-clause
ponylang/ponyc,ponylang/ponyc,ponylang/ponyc
2e886f6c2c8483c673b329acd8052bc0dcabf449
image-yaml/overcloud-images.yaml
image-yaml/overcloud-images.yaml
disk_images: - imagename: overcloud arch: amd64 type: qcow2 elements: - hosts baremetal dhcp-all-interfaces os-collect-config overcloud-full overcloud-controller overcloud-compute overcloud-ceph-storage heat-config-puppet heat-config-script puppet-modules hiera os-net-config stable-interface-names grub2 element-manifest network-gateway dynamic-login enable-packages-install pip-and-virtualenv-override packages: - python-psutil - python-debtcollector - plotnetcfg - sos - python-networking-cisco - python-UcsSdk - device-mapper-multipath - python-networking-bigswitch - openstack-neutron-bigswitch-lldp - openstack-neutron-bigswitch-agent options: - "--min-tmpfs 5" - imagename: ramdisk_agent arch: amd64 type: qcow2 elements: - ironic-agent dynamic-login element-manifest network-gateway epel enable-packages-install pip-and-virtualenv-override packages: - python-hardware-detect options: - "--min-tmpfs=5"
disk_images: - imagename: overcloud arch: amd64 type: qcow2 elements: - hosts baremetal dhcp-all-interfaces os-collect-config overcloud-full overcloud-controller overcloud-compute overcloud-ceph-storage heat-config-puppet heat-config-script puppet-modules hiera os-net-config stable-interface-names grub2 element-manifest network-gateway dynamic-login enable-packages-install pip-and-virtualenv-override packages: - python-psutil - python-debtcollector - plotnetcfg - sos - python-networking-cisco - python-UcsSdk - device-mapper-multipath - python-networking-bigswitch - openstack-neutron-bigswitch-lldp - openstack-neutron-bigswitch-agent options: - "--min-tmpfs 5" - imagename: ramdisk_agent arch: amd64 type: qcow2 elements: - ironic-agent dynamic-login element-manifest network-gateway enable-packages-install pip-and-virtualenv-override packages: - python-hardware-detect options: - "--min-tmpfs=5"
Remove epel element from base yaml ramdisk agent image
Remove epel element from base yaml ramdisk agent image The epel element should not be in the base element. Change-Id: I843e9ed3bd027dd0b975d6b423ba3475631b7bd9
YAML
apache-2.0
openstack/tripleo-common,openstack/tripleo-common
dd03e75be462746978c910690947e84fc340e7b3
.goreleaser.yml
.goreleaser.yml
builds: - ldflags: - -s -w -X github.com/khlieng/dispatch/version.Tag=v{{.Version}} -X github.com/khlieng/dispatch/version.Commit={{.ShortCommit}} -X github.com/khlieng/dispatch/version.Date={{.Date}} env: - CGO_ENABLED=0 goos: - linux - darwin - windows goarch: - amd64 - arm - arm64 goarm: - 6 - 7 archive: files: - none* format_overrides: - goos: windows format: zip checksum: name_template: "checksums.txt" changelog: filters: exclude: - ^v.* - "(?i)^update.*dep" - Merge pull request - Merge branch release: name_template: "{{.Version}}"
builds: - ldflags: - -s -w -X github.com/khlieng/dispatch/version.Tag=v{{.Version}} -X github.com/khlieng/dispatch/version.Commit={{.ShortCommit}} -X github.com/khlieng/dispatch/version.Date={{.Date}} env: - CGO_ENABLED=0 goos: - linux - darwin - windows goarch: - amd64 - arm - arm64 goarm: - 6 - 7 archive: files: - none* format_overrides: - goos: windows format: zip replacements: amd64: x64 darwin: mac checksum: name_template: "checksums.txt" changelog: filters: exclude: - ^v.* - "(?i)^update.*dep" - Merge pull request - Merge branch release: name_template: "{{.Version}}"
Add amd64 and darwin archive name replacements
Add amd64 and darwin archive name replacements
YAML
mit
khlieng/dispatch,khlieng/dispatch,khlieng/name_pending,khlieng/name_pending,khlieng/dispatch,khlieng/name_pending,khlieng/name_pending
cb3d490b395df11d40cab7d00d40842e3be093df
util/ansible/roles/azuracast-fixtures/tasks/main.yml
util/ansible/roles/azuracast-fixtures/tasks/main.yml
--- - name: Find fixture ZIP files find: paths: "{{ util_base }}/fixtures" patterns: "*.zip" recurse: yes register: fixture_zip_files - name: Extract fixture ZIP files unarchive: src: "{{ item.path }}" dest: "{{ app_base }}/stations" owner: azuracast with_items: "{{ fixture_zip_files.files }}" when: fixture_zip_files.matched > 0 - name: Find fixture SQL files find: paths: "{{ util_base }}/fixtures" patterns: "*.sql" recurse: yes register: fixture_sql_files - name: Import fixture SQL files become: true become_user: azuracast shell: "php {{ util_base }}/cli.php dbal:import {{ item.path }}" with_items: "{{ fixture_sql_files.files }}" when: fixture_sql_files.matched > 0 - name: Clear AzuraCast Cache and Restart Stations become: true become_user: azuracast shell: php {{ util_base }}/cli.php cache:clear && php {{ util_base }}/cli.php azuracast:radio:restart when: fixture_zip_files.matched > 0 or fixture_sql_files.matched > 0
--- - name: Find fixture ZIP files find: paths: "{{ util_base }}/fixtures" patterns: "*.zip" recurse: yes register: fixture_zip_files - name: Extract fixture ZIP files unarchive: src: "{{ item.path }}" dest: "{{ app_base }}/stations" owner: azuracast with_items: "{{ fixture_zip_files.files }}" when: fixture_zip_files.matched > 0 - name: Find fixture SQL files find: paths: "{{ util_base }}/fixtures" patterns: "*.sql" recurse: yes register: fixture_sql_files - name: Import fixture SQL files become: true become_user: azuracast shell: "php {{ util_base }}/cli.php dbal:import {{ item.path }}" with_items: "{{ fixture_sql_files.files }}" when: fixture_sql_files.matched > 0 register: fixture_sql_command_result failed_when: "'exception' in command_result.stdout" - name: Clear AzuraCast Cache and Restart Stations become: true become_user: azuracast shell: php {{ util_base }}/cli.php cache:clear && php {{ util_base }}/cli.php azuracast:radio:restart when: fixture_zip_files.matched > 0 or fixture_sql_files.matched > 0
Add failure criteria to fixture install.
Add failure criteria to fixture install.
YAML
agpl-3.0
AzuraCast/AzuraCast,AzuraCast/AzuraCast,SlvrEagle23/AzuraCast,SlvrEagle23/AzuraCast,SlvrEagle23/AzuraCast,SlvrEagle23/AzuraCast,SlvrEagle23/AzuraCast,AzuraCast/AzuraCast,AzuraCast/AzuraCast
859fee51a0d121842a42fd12da7e617258cc8011
recipes/dotty-dict/meta.yaml
recipes/dotty-dict/meta.yaml
{% set name = "dotty-dict" %} {% set version = "1.3.0" %} package: name: {{ name|lower }} version: {{ version }} pypi_name: "dotty_dict" source: url: https://pypi.io/packages/source/{{ pypi_name[0] }}/{{ pypi_name }}/{{ pypi_name }}-{{ version }}.tar.gz sha256: eb0035a3629ecd84397a68f1f42f1e94abd1c34577a19cd3eacad331ee7cbaf0 build: number: 0 noarch: python script: {{ PYTHON }} -m pip install . -vv requirements: host: - pip - python >=3.5 - setuptools_scm run: - python >=3.5 test: imports: - dotty_dict commands: - pip check requires: - pip about: home: https://pypi.org/project/dotty-dict/ summary: Dictionary wrapper for quick access to deeply nested keys. license: MIT license_file: LICENSE dev_url: https://github.com/pawelzny/dotty_dict doc_url: https://dotty-dict.readthedocs.io extra: recipe-maintainers: - KingDarBoja
{% set name = "dotty-dict" %} {% set version = "1.3.0" %} {% set pypi_name = "dotty_dict" %} package: name: {{ name|lower }} version: {{ version }} pypi_name: {{ pypi_name }} source: url: https://pypi.io/packages/source/{{ pypi_name[0] }}/{{ pypi_name }}/{{ pypi_name }}-{{ version }}.tar.gz sha256: eb0035a3629ecd84397a68f1f42f1e94abd1c34577a19cd3eacad331ee7cbaf0 build: number: 0 noarch: python script: {{ PYTHON }} -m pip install . -vv requirements: host: - pip - python >=3.5 - setuptools_scm run: - python >=3.5 test: imports: - dotty_dict commands: - pip check requires: - pip about: home: https://pypi.org/project/dotty-dict/ summary: Dictionary wrapper for quick access to deeply nested keys. license: MIT license_file: LICENSE dev_url: https://github.com/pawelzny/dotty_dict doc_url: https://dotty-dict.readthedocs.io extra: recipe-maintainers: - KingDarBoja
Use jinja template variable at the top
Use jinja template variable at the top
YAML
bsd-3-clause
igortg/staged-recipes,scopatz/staged-recipes,mariusvniekerk/staged-recipes,goanpeca/staged-recipes,hadim/staged-recipes,goanpeca/staged-recipes,ReimarBauer/staged-recipes,jochym/staged-recipes,conda-forge/staged-recipes,jochym/staged-recipes,jakirkham/staged-recipes,igortg/staged-recipes,kwilcox/staged-recipes,jakirkham/staged-recipes,patricksnape/staged-recipes,stuertz/staged-recipes,ocefpaf/staged-recipes,johanneskoester/staged-recipes,ocefpaf/staged-recipes,stuertz/staged-recipes,kwilcox/staged-recipes,mariusvniekerk/staged-recipes,patricksnape/staged-recipes,hadim/staged-recipes,scopatz/staged-recipes,johanneskoester/staged-recipes,ReimarBauer/staged-recipes,conda-forge/staged-recipes
20eaa43d18b9dbaaa66f9bfb26453d4d872e910c
roles/dev/defaults/main.yml
roles/dev/defaults/main.yml
--- projects: - { name: idea, version: '2017.3.5', url_prefix: 'https://download.jetbrains.com/idea/ideaIC-' } - { name: pycharm, version: '2017.3.4', url_prefix: 'https://download.jetbrains.com/python/pycharm-professional-' } bashrc_path: "{{ ansible_env.HOME }}/.bashrc"
--- projects: - { name: idea, version: '2018.1', url_prefix: 'https://download.jetbrains.com/idea/ideaIC-' } - { name: pycharm, version: '2018.1', url_prefix: 'https://download.jetbrains.com/python/pycharm-professional-' } bashrc_path: "{{ ansible_env.HOME }}/.bashrc"
Upgrade pycharm and intellij idea
Upgrade pycharm and intellij idea
YAML
mit
andornaut/ansible-workstation
334627535ea1d9987bfa6aca762230af66cd8618
roles/fidimag/vars/main.yml
roles/fidimag/vars/main.yml
--- FIDIMAG_REPO_URL: https://github.com/computationalmodelling/fidimag.git FIDIMAG_INSTALL_PATH: /opt/fidimag FIDIMAG_VERSION: v0.0.1 XPRA_DISPLAYS: - 25 - 30 - 36 - 49 - 50 - 64 - 75 - 81 - 90 - 99
--- FIDIMAG_REPO_URL: https://github.com/computationalmodelling/fidimag.git FIDIMAG_INSTALL_PATH: /opt/fidimag FIDIMAG_VERSION: v0.0.2 XPRA_DISPLAYS: - 25 - 30 - 36 - 49 - 50 - 64 - 75 - 81 - 90 - 99
Build fidimag with the somewhat stable 0.0.2 version for consistency, which fixes a container installation issue.
Build fidimag with the somewhat stable 0.0.2 version for consistency, which fixes a container installation issue.
YAML
bsd-3-clause
fangohr/virtualmicromagnetics,computationalmodelling/virtualmicromagnetics,computationalmodelling/virtualmicromagnetics,fangohr/virtualmicromagnetics
461718186328d1b705636e57215a47e3fb59d27e
config/ansible/tasks/gerrit_replication_update.yaml
config/ansible/tasks/gerrit_replication_update.yaml
--- - include: update_configrepo.yaml - stat: path=/root/config/gerrit/replication.config register: rstat - name: Check if replication.config changed command: diff /root/config/gerrit/replication.config /home/gerrit/site_path/etc/replication.config register: diffstat no_log: True failed_when: diffstat.rc > 1 - name: Update Gerrit replication.config command: cp /root/config/gerrit/replication.config /home/gerrit/site_path/etc/ when: rstat.stat.exists == True and diffstat.rc == 1 - name: Reload the replication plugin command: ssh -p 29418 admin@{{ fqdn }} gerrit plugin reload replication when: rstat.stat.exists == True and diffstat.rc == 1 - name: Restart the replication command: ssh -p 29418 admin@{{ fqdn }} replication start --all when: rstat.stat.exists == True and diffstat.rc == 1
--- - include: update_configrepo.yaml - stat: path=/root/config/gerrit/replication.config register: rstat - name: Check if replication.config changed command: diff /root/config/gerrit/replication.config /home/gerrit/site_path/etc/replication.config register: diffstat no_log: True when: rstat.stat.exists == True failed_when: diffstat.rc > 1 - name: Update Gerrit replication.config command: cp /root/config/gerrit/replication.config /home/gerrit/site_path/etc/ when: rstat.stat.exists == True and diffstat.rc == 1 - name: Reload the replication plugin command: ssh -p 29418 admin@{{ fqdn }} gerrit plugin reload replication when: rstat.stat.exists == True and diffstat.rc == 1 - name: Restart the replication command: ssh -p 29418 admin@{{ fqdn }} replication start --all when: rstat.stat.exists == True and diffstat.rc == 1
Fix failing diff if /root/config/gerrit/replication.config missing
Fix failing diff if /root/config/gerrit/replication.config missing If /root/config/gerrit/replication.config is missing the diff will return with an exit code of 2, and therefore the task will fail. Due to the earlier stat it's already known if the file exists, so just checking this using a "when" condition. Change-Id: I22627f4af1189fdddbef5f8ee8bea5489dd7012f
YAML
apache-2.0
enovance/software-factory,enovance/software-factory,enovance/software-factory,enovance/software-factory,enovance/software-factory
58b151577bfbceb41fed5ba485cadd5ef28905be
roles/scripts/tasks/main.yml
roles/scripts/tasks/main.yml
- name: Install custom script helpers copy: src={{ item }} dest=/usr/local/bin/{{ item | basename }} mode=0755 with_fileglob: - ../files/* - name: Get wp-phptidy for html coding standards command: curl -L -o /usr/local/bin/wp-phptidy https://raw.githubusercontent.com/scribu/wp-phptidy/master/wp-phptidy.php - name: Give wp-phptidy execute permissions file: path: /usr/local/bin/wp-phptidy mode: 0755
- name: Install custom script helpers copy: src={{ item }} dest=/usr/local/bin/{{ item | basename }} mode=0755 with_fileglob: - ../files/*
Remove undocumented and unused wp-phptidy in favor of phpcs
Remove undocumented and unused wp-phptidy in favor of phpcs
YAML
mit
Seravo/wp-vagrant,Seravo/wp-palvelu-vagrant
f3a6b2c0fb62823da59eea3986dac6ad9aa7322b
.octorunner.yml
.octorunner.yml
image: golang:latest script: - go get -u github.com/kardianos/govendor - cd /go - mkdir -p src/github.com/boyvanduuren/ - ln -s /var/run/octorunner /go/src/github.com/boyvanduuren/octorunner - cd /go/src/github.com/boyvanduuren/octorunner - govendor sync - go test ./...
image: golang:latest script: - go get -u github.com/kardianos/govendor - cd /go - mkdir -p src/github.com/boyvanduuren/ - ln -s /var/run/octorunner /go/src/github.com/boyvanduuren/octorunner - cd /go/src/github.com/boyvanduuren/octorunner - govendor sync - go get - go test ./...
Make sure ql dep is downloaded before running tests
Make sure ql dep is downloaded before running tests
YAML
mit
boyvanduuren/octorunner
158a871e7d94638e368d190ca8e05a8656b5fc63
.overcommit.yml
.overcommit.yml
# # Select version of overcommit and the other tools from Gemfile # gemfile: Gemfile # # Hooks that are run against every commit message after a user has written it. # CommitMsg: HardTabs: enabled: true SingleLineSubject: enabled: true # # Hooks that are run after `git commit` is executed, before the commit message # editor is displayed. # PreCommit: BundleCheck: enabled: true LocalPathsInGemfile: enabled: true ExecutePermissions: enabled: true exclude: - bin/* RuboCop: enabled: true TrailingWhitespace: enabled: true YamlSyntax: enabled: true # # Hooks that are run before pushing changes # PrePush: RSpec: enabled: true
# # Select version of overcommit and the other tools from Gemfile # gemfile: Gemfile # # Hooks that are run against every commit message after a user has written it. # CommitMsg: HardTabs: enabled: true SingleLineSubject: enabled: true # # Hooks that are run after `git commit` is executed, before the commit message # editor is displayed. # PreCommit: BundleCheck: enabled: true LocalPathsInGemfile: enabled: true ExecutePermissions: enabled: true exclude: - bin/* RuboCop: enabled: true exclude: - 'db/schema.rb' - 'db/migrate/*.rb' TrailingWhitespace: enabled: true YamlSyntax: enabled: true # # Hooks that are run before pushing changes # PrePush: RSpec: enabled: true
Exclude migration related files from rubocop hook
Exclude migration related files from rubocop hook
YAML
apache-2.0
CredentialEngine/CredentialRegistry,coffeejunk/CredentialRegistry,learningtapestry/learningregistry,learningtapestry/learningregistry,CredentialEngine/CredentialRegistry,CredentialEngine/CredentialRegistry,coffeejunk/CredentialRegistry,coffeejunk/CredentialRegistry,learningtapestry/learningregistry
431723522dce7fb6fdfa9a3599a2d374724d3479
.pep8speaks.yml
.pep8speaks.yml
scanner: linter: flake8 flake8: max-line-length: 79 # Using the default of 79 in PEP8 ignore: - E741 # l and b are valid variable names for the galactic frame - E226 # Don't force "missing whitespace around arithmetic operator" - E402 # .conf has to be set in the __init__.py modules imports - W503 # line break before binary operator - W504 # we've been perpetually annoyed by W504 "line break after binary operator", since there's often no real alternative exclude: - _astropy_init.py - version.py
scanner: linter: flake8 flake8: max-line-length: 100 ignore: - E741 # l and b are valid variable names for the galactic frame - E226 # Don't force "missing whitespace around arithmetic operator" - E402 # .conf has to be set in the __init__.py modules imports - W503 # line break before binary operator - W504 # we've been perpetually annoyed by W504 "line break after binary operator", since there's often no real alternative exclude: - _astropy_init.py - version.py
Make PEP8speaks max line length agree with tox flake8 codestyle test
Make PEP8speaks max line length agree with tox flake8 codestyle test
YAML
bsd-3-clause
larrybradley/photutils,astropy/photutils
7f4727cfa3b307f7500ef08da2ce5907d4c57215
config/mongoid.yml
config/mongoid.yml
development: host: localhost database: cartoque_development test: host: localhost database: cartoque_test # set these environment variables on your prod server production: host: <%= ENV['MONGOID_HOST'] || 'localhost' %> port: <%= ENV['MONGOID_PORT'] || '27017' %> username: <%= ENV['MONGOID_USERNAME'] %> password: <%= ENV['MONGOID_PASSWORD'] %> database: <%= ENV['MONGOID_DATABASE'] || 'cartoque' %> # slaves: # - host: slave1.local # port: 27018 # - host: slave2.local # port: 27019
development: host: localhost database: cartoque test: host: localhost database: cartoque_test # set these environment variables on your prod server production: host: <%= ENV['MONGOID_HOST'] || 'localhost' %> port: <%= ENV['MONGOID_PORT'] || '27017' %> username: <%= ENV['MONGOID_USERNAME'] %> password: <%= ENV['MONGOID_PASSWORD'] %> database: <%= ENV['MONGOID_DATABASE'] || 'cartoque' %> # slaves: # - host: slave1.local # port: 27018 # - host: slave2.local # port: 27019
Use the same mongodb database name in development and production modes for simplicity
Use the same mongodb database name in development and production modes for simplicity
YAML
mit
cartoque/cartoque,jbbarth/cartoque,skylost/cartoque,jbbarth/cartoque,skylost/cartoque,cartoque/cartoque,jbbarth/cartoque,cartoque/cartoque
e30604311f71e4ac84500b15f54b661165a5287f
test/config/database.yml
test/config/database.yml
<% raise "ENV['DB'] not specified!" unless ENV['DB'] %> # a bit hackish - tests would require a refactoring instead mysql2: &mysql2 host: localhost adapter: mysql2 database: activewarehouse_etl_test username: root encoding: utf8 local_infile: true # the tests would require a rework: disabling casting for now cast: false postgresql: &postgresql adapter: postgresql database: activewarehouse_etl_test username: postgres # TODO - refactor test to avoid using 2 databases maybe? operational_database: <<: *<%= ENV['DB'] %> data_warehouse: <<: *<%= ENV['DB'] %> etl_execution: adapter: mysql2 database: etl_execution
<% raise "ENV['DB'] not specified!" unless ENV['DB'] %> # a bit hackish - tests would require a refactoring instead mysql2: &mysql2 host: 127.0.0.1 adapter: mysql2 database: activewarehouse_etl_test username: root encoding: utf8 local_infile: true # the tests would require a rework: disabling casting for now cast: false postgresql: &postgresql adapter: postgresql database: activewarehouse_etl_test username: postgres # TODO - refactor test to avoid using 2 databases maybe? operational_database: <<: *<%= ENV['DB'] %> data_warehouse: <<: *<%= ENV['DB'] %> etl_execution: <<: *mysql2 database: etl_execution
Make the test suite start on my machine
Make the test suite start on my machine
YAML
mit
activewarehouse/activewarehouse-etl
62c0abcb38c5c836b24c06b78c01bfd655c0bfdb
metadata/bluepie.ad_silence.yml
metadata/bluepie.ad_silence.yml
AntiFeatures: - NonFreeAdd Categories: - System License: GPL-3.0-only AuthorName: BluePie AuthorEmail: iambluepie@gmail.com AuthorWebSite: https://github.com/aghontpi WebSite: https://github.com/aghontpi/ad-silence SourceCode: https://github.com/aghontpi/ad-silence IssueTracker: https://github.com/aghontpi/ad-silence/issues AutoName: Ad-silence RepoType: git Repo: https://github.com/aghontpi/ad-silence.git Builds: - versionName: 0.3.0 versionCode: 7 commit: a28e8eb5751431da5dcf22d185674073fc0e1f6f subdir: app sudo: - apt-get update || apt-get update - apt-get install -y openjdk-11-jdk - update-alternatives --auto java gradle: - yes AutoUpdateMode: Version %v UpdateCheckMode: Tags CurrentVersion: 0.3.0 CurrentVersionCode: 7
AntiFeatures: - NonFreeAdd Categories: - System License: GPL-3.0-only AuthorName: BluePie AuthorEmail: iambluepie@gmail.com AuthorWebSite: https://github.com/aghontpi WebSite: https://github.com/aghontpi/ad-silence SourceCode: https://github.com/aghontpi/ad-silence IssueTracker: https://github.com/aghontpi/ad-silence/issues AutoName: Ad-silence RepoType: git Repo: https://github.com/aghontpi/ad-silence.git Builds: - versionName: 0.3.0 versionCode: 7 commit: a28e8eb5751431da5dcf22d185674073fc0e1f6f subdir: app sudo: - apt-get update || apt-get update - apt-get install -y openjdk-11-jdk - update-alternatives --auto java gradle: - yes - versionName: 0.4.0 versionCode: 11 commit: 11f63efe95a71fa93770b23d98ae059e984b1e6b subdir: app sudo: - apt-get update || apt-get update - apt-get install -y openjdk-11-jdk - update-alternatives --auto java gradle: - yes AutoUpdateMode: Version %v UpdateCheckMode: Tags CurrentVersion: 0.4.0 CurrentVersionCode: 11
Update Ad-silence to 0.4.0 (11)
Update Ad-silence to 0.4.0 (11)
YAML
agpl-3.0
f-droid/fdroiddata,f-droid/fdroiddata
f5532e8520212b55dbb46bbc7736aa1d34fe656b
environment/external.yaml
environment/external.yaml
resources: httpproxy: number: 1 flavor: large image: trusty networks: default assign_floating_ip: true
resources: httpproxy: number: 1 flavor: large image: trusty networks: default boot_volume: default_volume assign_floating_ip: true
Make httpproxy boot from volume
Make httpproxy boot from volume so that we can create a new external layout
YAML
apache-2.0
devendermishrajio/puppet-rjil,amar266/puppet-rjil,JioCloud/puppet-rjil-keystone,swamireddy/puppet-rjil,bafna/puppet-rjil,swamireddy/puppet-rjil,rakeshmi/puppet-rjil,ajayaa/puppet-rjil,sajuptpm/puppet-rjil-temp,amar266/puppet-rjil,d4devops/puppet-openstack_cloud,sajuptpm/puppet-rjil-temp,anshprat/puppet-rjil,anshprat/puppet-rjil,roopali8/puppet-rjil,vpramo/puppet-rjil,upadhyay-prashant/puppet-rjil,bafna/puppet-rjil,JioCloud/puppet-rjil,upadhyay-prashant/puppet-rjil,vpramo/puppet-rjil,sajuptpm/puppet-rjil,JioCloud/puppet-rjil,d4devops/puppet-openstack_cloud,alokjani/puppet-rjil,rakeshmi/puppet-rjil,sajuptpm/puppet-rjil-temp,rakeshmi/puppet-rjil,JioCloud/puppet-rjil,JioCloud/puppet-rjil-keystone,JioCloud/puppet-rjil-keystone,jiocloudDSS/puppet-dss,akash1808/puppet-rjil,devendermishrajio/puppet-rjil,punituee/puppet-rjil,rahul1aggarwal/puppet-rjil,rahul1aggarwal/puppet-rjil,punituee/puppet-rjil,rakeshmi/puppet-rjil,devendermishrajio/puppet-rjil,jiocloudDSS/puppet-dss,saneax/puppet-rjil,sajuptpm/puppet-rjil,ajayaa/puppet-rjil,rahul1aggarwal/puppet-rjil,bafna/try-dss,bafna/try-dss,rohit-k/puppet-rjil,JioCloud/puppet-rjil,roopali8/puppet-rjil,d4devops/puppet-openstack_cloud,rohit-k/puppet-rjil,anshprat/puppet-rjil,bafna/puppet-rjil,alokjani/puppet-rjil,akash1808/puppet-rjil,devendermishrajio/puppet-rjil,vpramo/puppet-rjil,JioCloud/puppet-rjil-keystone,alokjani/puppet-rjil,upadhyay-prashant/puppet-rjil,ajayaa/puppet-rjil,akash1808/puppet-rjil,sajuptpm/puppet-rjil,upadhyay-prashant/puppet-rjil,swamireddy/puppet-rjil,punituee/puppet-rjil,bafna/puppet-rjil,roopali8/puppet-rjil,jiocloudDSS/puppet-dss,sajuptpm/puppet-rjil-temp,saneax/puppet-rjil,ajayaa/puppet-rjil,bafna/try-dss,saneax/puppet-rjil,rahul1aggarwal/puppet-rjil,alokjani/puppet-rjil,amar266/puppet-rjil,roopali8/puppet-rjil,rohit-k/puppet-rjil,saneax/puppet-rjil,swamireddy/puppet-rjil,jiocloudDSS/puppet-dss,punituee/puppet-rjil,akash1808/puppet-rjil,sajuptpm/puppet-rjil,bafna/try-dss,anshprat/puppet-rjil,vpramo/puppet-rjil,amar266/puppet-rjil,d4devops/puppet-openstack_cloud,rohit-k/puppet-rjil
239fa847ca7d091b5a8d8d9341e433a211c43de4
actions/workflows/st2_integration_tests.yaml
actions/workflows/st2_integration_tests.yaml
--- chain: - name: "core_local_date" ref: "st2cd.action_run" params: name: "core_local_date" action: "core.local" params: "date" hosts: "{{hostname}}" on-success: "core_http_google" on-failure: "core_http_google" - name: "core_http_google" ref: "st2cd.action_run" params: name: "core_http_google" action: "core.http" params: "url=http://www.google.com" hosts: "{{hostname}}" on-success: "core_remote_single_host" on-failure: "core_remote_single_host" - name: "core_remote_single_host" ref: "st2cd.action_run" params: name: "core_remote_single_host" action: "core.remote" params: "hosts={{hostname}} hostname" hosts: "{{hostname}}" on-success: "linux_diag_loadavg" on-failure: "linux_diag_loadavg" - name: "linux_diag_loadavg" ref: "st2cd.action_run" params: name: "linux_diag_loadavg" action: "linux.diag_loadavg" params: "hostname={{hostname}}" hosts: "{{hostname}}" on-success: "aggregator" on-failure: "aggregator" - name: "aggregator" ref: "st2cd.result_aggregator" params: results: "{{__results}}" default: "core_local_date"
--- chain: - name: "core_local_date" ref: "st2cd.action_run" params: name: "core_local_date" action: "core.local" params: "date" hosts: "{{hostname}}" on-success: "core_http_google" - name: "core_http_google" ref: "st2cd.action_run" params: name: "core_http_google" action: "core.http" params: "url=http://www.google.com" hosts: "{{hostname}}" on-success: "core_remote_single_host" - name: "core_remote_single_host" ref: "st2cd.action_run" params: name: "core_remote_single_host" action: "core.remote" params: "hosts={{hostname}} hostname" hosts: "{{hostname}}" on-success: "action_chain_load" - name: "action_chain_load" ref: "st2cd.action_run" params: name: "linux_diag_loadavg" action: "linux.diag_loadavg" params: "hostname={{hostname}}" hosts: "{{hostname}}" default: "core_local_date"
Make this actually fail if there is a problem
Make this actually fail if there is a problem
YAML
apache-2.0
StackStorm/st2cd,StackStorm/st2cd
0f1ca99984de1b8d421475e7ac5714f954aa4945
.github/release-drafter.yml
.github/release-drafter.yml
name-template: "v$RESOLVED_VERSION 🌈" tag-template: "v$RESOLVED_VERSION" categories: - title: "🚀 Features" labels: - "feature" - "enhancement" - title: "🐛 Bug Fixes" labels: - "fix" - "bugfix" - "bug" - title: "🧰 Maintenance" label: "chore" change-template: "- $TITLE (#$NUMBER)" version-resolver: major: labels: - "major" minor: labels: - "minor" patch: labels: - "patch" default: patch template: | ## Installation The Third Coast `vendordeps` file is at: http://maven.strykeforce.org/thirdcoast.json To install, use **Install new libraries (online)** in VS Code or download manually to your project `vendordeps` directory. ## Changes $CHANGES
name-template: "v$RESOLVED_VERSION 🌈" tag-template: "v$RESOLVED_VERSION" categories: - title: "🚀 Features" labels: - "feature" - "enhancement" - title: "🐛 Bug Fixes" labels: - "fix" - "bugfix" - "bug" - title: "🧰 Maintenance" label: "chore" change-template: "- $TITLE (#$NUMBER)" version-resolver: major: labels: - "major" minor: labels: - "minor" patch: labels: - "patch" default: patch template: | ## Installation The Third Coast `vendordeps` file is at: http://maven.strykeforce.org/thirdcoast.json To install, use **Install new libraries (online)** in VS Code or download manually to your project `vendordeps` directory. ``` $ ./gradlew vendordep --url=http://maven.strykeforce.org/thirdcoast.json ``` ## Changes $CHANGES
Add GradleRIO command to download vendordep
Add GradleRIO command to download vendordep
YAML
mit
strykeforce/thirdcoast,strykeforce/thirdcoast,strykeforce/thirdcoast
e7143eface23933a10215f59e50325f5d264c458
.github/workflows/build.yml
.github/workflows/build.yml
on: push name: build jobs: test: runs-on: ubuntu-latest services: db: image: postgres:11 ports: ['5432:5432'] options: >- --health-cmd pg_isready --health-interval 10s --health-timeout 5s --health-retries 5 steps: - uses: actions/checkout@v2 - uses: actions/setup-elixir@v1.0.0 with: otp-version: 22.x elixir-version: 1.9.x - run: script/cibuild env: ACTIONS: true COVERALLS_REPO_TOKEN: ${{ secrets.COVERALLS_REPO_TOKEN }}
on: push name: build jobs: test: runs-on: ubuntu-latest services: db: image: postgres:12.1 ports: ['5432:5432'] options: >- --health-cmd pg_isready --health-interval 10s --health-timeout 5s --health-retries 5 steps: - uses: actions/checkout@v2 - uses: actions/setup-elixir@v1.0.0 with: otp-version: 22.x elixir-version: 1.9.x - run: script/cibuild env: ACTIONS: true COVERALLS_REPO_TOKEN: ${{ secrets.COVERALLS_REPO_TOKEN }}
Use Postgres v12.1 for now until they get their images fixed
Use Postgres v12.1 for now until they get their images fixed
YAML
mit
lee-dohm/atom-style-tweaks,lee-dohm/atom-style-tweaks,lee-dohm/atom-style-tweaks,lee-dohm/atom-style-tweaks
76d3d580b5b8f83b5e57183478744a42ea215d01
.github/workflows/build.yml
.github/workflows/build.yml
name: Build on: push: branches: - master pull_request: branches: - master jobs: build: runs-on: ubuntu-latest steps: - uses: actions/checkout@v2 - uses: actions/cache@v1 with: path: vendor/bundle key: ${{ runner.os }}-gems-${{ hashFiles('**/Gemfile.lock') }} restore-keys: | ${{ runner.os }}-gems- - name: Set up Ruby 2.6 uses: actions/setup-ruby@v1 with: ruby-version: 2.6.x - name: Bundle install run: | bundle config path vendor/bundle bundle install --jobs 4 --retry 3 - name: Build with Jekyll run: | bundle exec jekyll build
name: Build on: push: branches: - master pull_request: branches: - master jobs: build: runs-on: ubuntu-latest steps: - uses: actions/checkout@v2 - uses: actions/cache@v1 with: path: vendor/bundle key: ${{ runner.os }}-gems-${{ hashFiles('Gemfile') }} restore-keys: | ${{ runner.os }}-gems- - name: Set up Ruby 2.6 uses: actions/setup-ruby@v1 with: ruby-version: 2.6.x - name: Bundle install run: | bundle config path vendor/bundle bundle install --jobs 4 --retry 3 - name: Build with Jekyll run: | bundle exec jekyll build
Use cache key of Gemfile
Use cache key of Gemfile Failure if Gemfile.lock is used: ##[error]The template is not valid. hashFiles('/home/runner/work/pixyll/pixyll/**/Gemfile.lock') failed. Search pattern '/home/runner/work/pixyll/pixyll/**/Gemfile.lock' doesn't match any file under '/home/runner/work/pixyll/pixyll'
YAML
mit
muodov/blog,galeone/galeone.github.io,muodov/blog,galeone/galeone.github.io,kongove/kongove.github.com,kongove/kongove.github.com,pepijn/pepijn.github.io,galeone/galeone.github.io,tanmaybinaykiya/tanmaybinaykiya.github.io,deepredsky/deepredsky.github.io,faustovaz/faustovaz.github.io,kongove/kongove.github.com,deepredsky/deepredsky.github.io,faustovaz/faustovaz.github.io,muodov/blog,grimsa/grimsa.github.io,tanmaybinaykiya/tanmaybinaykiya.github.io,pepijn/pepijn.github.io,galeone/galeone.github.io,pepijn/pepijn.github.io,kongove/kongove.github.com,grimsa/grimsa.github.io,galeone/galeone.github.io,tanmaybinaykiya/tanmaybinaykiya.github.io,kongove/kongove.github.com
9d6df3137a018426b9b125e908abb8a50e7e341b
.github/workflows/build.yml
.github/workflows/build.yml
name: Build on: pull_request: push: branches: - main - develop jobs: build: name: Build runs-on: ${{ matrix.os }} strategy: matrix: os: [windows-latest, ubuntu-latest, macos-latest] steps: - name: Get the sources uses: actions/checkout@v2 - name: Fetch all history for all tags and branches run: git fetch --prune --unshallow - name: Install .NET Core SDK 3.1.301 uses: actions/setup-dotnet@v1 with: dotnet-version: '3.1.301' - name: Run Cake script uses: ecampidoglio/cake-action@v1.1.1 env: DOTNET_ROLL_FORWARD_ON_NO_CANDIDATE_FX: 2 with: target: Run-Integration-Tests cake-bootstrap: true
name: Build on: pull_request: push: branches: - main - develop - hotfix/* jobs: build: name: Build runs-on: ${{ matrix.os }} strategy: matrix: os: [windows-latest, ubuntu-latest, macos-latest] steps: - name: Get the sources uses: actions/checkout@v2 - name: Fetch all history for all tags and branches run: git fetch --prune --unshallow - name: Install .NET Core SDK 3.1.301 uses: actions/setup-dotnet@v1 with: dotnet-version: '3.1.301' - name: Run Cake script uses: ecampidoglio/cake-action@v1.1.1 env: DOTNET_ROLL_FORWARD_ON_NO_CANDIDATE_FX: 2 with: target: Run-Integration-Tests cake-bootstrap: true
Build hotfix branches on GitHub actions
Build hotfix branches on GitHub actions
YAML
mit
gep13/cake,patriksvensson/cake,devlead/cake,cake-build/cake,patriksvensson/cake,devlead/cake,gep13/cake,cake-build/cake
3b2a3ad5117480294de822b18fffbc07c22ca75e
.github/workflows/build.yml
.github/workflows/build.yml
name: Build project on: [push] jobs: build_project: runs-on: ubuntu-20.04 steps: - name: Set up Git repository uses: actions/checkout@v1 - name: Build run: | mvn test --batch-mode
name: Build project on: [push] jobs: build_project: runs-on: ubuntu-20.04 steps: - name: Set up Git repository uses: actions/checkout@v1 - name: Build run: | mvn test --batch-mode - name: Install run: | mvn install --batch-mode
Test install step in CI
Test install step in CI
YAML
apache-2.0
tudelft-atlarge/graphalytics-platforms-reference,ldbc/ldbc_graphalytics_platforms_reference,ldbc/ldbc_graphalytics_platforms_reference,tudelft-atlarge/graphalytics-platforms-reference
a961f0ef25663f787096149ad978ca357152eace
.github/workflows/build.yml
.github/workflows/build.yml
name: build on: push: branches: [ main ] pull_request: branches: [ main ] workflow_dispatch: jobs: build: name: ${{ matrix.os }} runs-on: ${{ matrix.os }} strategy: fail-fast: false matrix: os: [ macos-latest, ubuntu-latest, windows-latest ] steps: - name: Checkout code uses: actions/checkout@v3 - name: Setup .NET Core SDK uses: actions/setup-dotnet@v2 - name: Build, Test and Publish shell: pwsh run: ./build.ps1 env: DOTNET_CLI_TELEMETRY_OPTOUT: true DOTNET_SKIP_FIRST_TIME_EXPERIENCE: true NUGET_XMLDOC_MODE: skip
name: build on: push: branches: [ main ] pull_request: branches: [ main ] workflow_dispatch: jobs: build: name: ${{ matrix.os }} runs-on: ${{ matrix.os }} strategy: fail-fast: false matrix: os: [ macos-latest, ubuntu-latest, windows-latest ] steps: - name: Checkout code uses: actions/checkout@v3 - name: Setup .NET Core SDK uses: actions/setup-dotnet@v2 - name: Build, Test and Publish shell: pwsh run: ./build.ps1 env: DOTNET_CLI_TELEMETRY_OPTOUT: true DOTNET_SKIP_FIRST_TIME_EXPERIENCE: true DOTNET_SYSTEM_CONSOLE_ALLOW_ANSI_COLOR_REDIRECTION: 1 NUGET_XMLDOC_MODE: skip TERM: xterm
Enable colour output from .NET CLI
Enable colour output from .NET CLI Enable coloured output to the terminal in GitHub Actions for Linux and macOS.
YAML
apache-2.0
martincostello/project-euler
dbbe2b9d6bdccb7b869a24532313e47aad49d01d
.github/workflows/maven.yml
.github/workflows/maven.yml
# Licensed to the Apache Software Foundation (ASF) under one or more # contributor license agreements. See the NOTICE file distributed with # this work for additional information regarding copyright ownership. # The ASF licenses this file to You under the Apache License, Version 2.0 # (the "License"); you may not use this file except in compliance with # the License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. name: Java CI on: [push, pull_request] jobs: build: runs-on: ${{ matrix.os }} continue-on-error: ${{ matrix.experimental }} strategy: matrix: os: [ubuntu-latest, windows-latest] java: [ 8, 11, 17 ] experimental: [false] # include: # - java: 18-ea # os: ubuntu-latest # experimental: true steps: - uses: actions/checkout@v3 - name: Set up JDK ${{ matrix.java }} uses: actions/setup-java@v3 with: distribution: adopt java-version: ${{ matrix.java }} - name: Build with Maven run: mvn -V --no-transfer-progress
# Licensed to the Apache Software Foundation (ASF) under one or more # contributor license agreements. See the NOTICE file distributed with # this work for additional information regarding copyright ownership. # The ASF licenses this file to You under the Apache License, Version 2.0 # (the "License"); you may not use this file except in compliance with # the License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. name: Java CI on: [push, pull_request] permissions: contents: read jobs: build: runs-on: ${{ matrix.os }} continue-on-error: ${{ matrix.experimental }} strategy: matrix: os: [ubuntu-latest, windows-latest] java: [ 8, 11, 17 ] experimental: [false] # include: # - java: 18-ea # os: ubuntu-latest # experimental: true steps: - uses: actions/checkout@v3 - name: Set up JDK ${{ matrix.java }} uses: actions/setup-java@v3 with: distribution: adopt java-version: ${{ matrix.java }} - name: Build with Maven run: mvn -V --no-transfer-progress
Set permissions for GH Actions
Set permissions for GH Actions
YAML
apache-2.0
apache/commons-imaging,apache/commons-imaging
e50a38e9d257067149e62695c6437754fa625562
.github/workflows/stale.yml
.github/workflows/stale.yml
name: Mark stale issues and pull requests on: schedule: - cron: "30 1 * * *" jobs: stale: runs-on: ubuntu-latest steps: - uses: actions/stale@v3 with: repo-token: ${{ secrets.GITHUB_TOKEN }} stale-issue-message: 'As this issue has received no new activity, it has been marked as stale' stale-pr-message: 'As this pull request has received no new activity, it has been marked as stale' stale-issue-label: 'no-issue-activity' stale-pr-label: 'no-pr-activity' exempt-issue-labels: 'Up for Grabs'
name: Mark stale issues and pull requests on: schedule: - cron: "30 1 * * *" jobs: stale: runs-on: ubuntu-latest steps: - uses: actions/stale@v3 with: repo-token: ${{ secrets.GITHUB_TOKEN }} stale-issue-message: 'As this issue has received no new activity, it has been marked as stale' stale-pr-message: 'As this pull request has received no new activity, it has been marked as stale' stale-issue-label: 'no-issue-activity' stale-pr-label: 'no-pr-activity' exempt-issue-labels: 'Up for Grabs,Blocked' exempt-pr-labels: 'Work in progress'
Expand list of exempt labels
Expand list of exempt labels
YAML
bsd-3-clause
zarlo/Cosmos,CosmosOS/Cosmos,CosmosOS/Cosmos,zarlo/Cosmos,CosmosOS/Cosmos,zarlo/Cosmos,CosmosOS/Cosmos,zarlo/Cosmos
1cf1b0248d23cd0789332df868c7690672291710
.github/workflows/tests.yml
.github/workflows/tests.yml
name: Node CI on: [push, pull_request] jobs: build: runs-on: ubuntu-latest strategy: matrix: node-version: [12.x, 14.x, 16.x] steps: - uses: actions/checkout@v1 - name: Use Node.js ${{ matrix.node-version }} uses: actions/setup-node@v1 with: node-version: ${{ matrix.node-version }} - name: npm install, build, and test run: | npm install npm run ci-test env: CI: true
name: Node CI on: push: branches: - master pull_request: branches: - master jobs: build: runs-on: ubuntu-latest strategy: matrix: node-version: [12.x, 14.x, 16.x] steps: - uses: actions/checkout@v1 - name: Use Node.js ${{ matrix.node-version }} uses: actions/setup-node@v1 with: node-version: ${{ matrix.node-version }} - name: npm install, build, and test run: | npm install npm run ci-test env: CI: true
Change PR to only run on master
Change PR to only run on master
YAML
mit
taylorhakes/promise-polyfill
d75f3a5391a46ea267602dbea79c780436d63320
ansible/main.yml
ansible/main.yml
--- - hosts: all vars: cf_cli_version: "6.32.0" maven_version: "3.5.2" intellij_version: "2017.2.6" sts_version: "3.9.1" eclipse_version: "4.7.1a" docker_compose_version: "1.17.1" atom_version: "1.20.1" gradle_version: "4.3.1" go_version: "1.9.2" compliance_masonry_version: "1.1.2" bosh_cli_version: "2.0.45" yaml_linux_version: "1.13.1" terraform_version: "0.10.8" bosh_bootloader_version: "5.4.1" postman_version: "5.3.2" gather_facts: yes become: yes roles: - disable_auto_update - developer_packages - git - atom - docker_service - docker_compose - cf_cli - maven - gradle - intellij - sts - unity_desktop - go - postman - gitbook - compliance_masonry - gcloud - newrelic - bosh
--- - hosts: all vars: cf_cli_version: "6.33.0" maven_version: "3.5.2" intellij_version: "2017.2.6" sts_version: "3.9.1" eclipse_version: "4.7.1a" docker_compose_version: "1.17.1" atom_version: "1.20.1" gradle_version: "4.3.1" go_version: "1.9.2" compliance_masonry_version: "1.1.2" bosh_cli_version: "2.0.45" yaml_linux_version: "1.13.1" terraform_version: "0.11.0" bosh_bootloader_version: "5.6.0" postman_version: "5.3.2" gather_facts: yes become: yes roles: - disable_auto_update - developer_packages - git - atom - docker_service - docker_compose - cf_cli - maven - gradle - intellij - sts - unity_desktop - go - postman - gitbook - compliance_masonry - gcloud - newrelic - bosh
Update cf-cli, bosh-bootloader & terraform
Update cf-cli, bosh-bootloader & terraform
YAML
apache-2.0
cagiti/developer-environment
e0ad4860f5461a9a1b7bbca45009d12cc6999a42
.github/workflows/continuous-integration-workflow.yml
.github/workflows/continuous-integration-workflow.yml
name: Continuous Integration on: [push] jobs: build: runs-on: ubuntu-latest strategy: fail-fast: false matrix: python-version: ['3.7', '3.8'] steps: - uses: actions/checkout@v2 - uses: conda-incubator/setup-miniconda@v2 with: auto-update-conda: true python-version: ${{ matrix.python }} - name: Install Tox and any other packages shell: bash -l {0} run: conda install -c conda-forge tox-conda - name: Run Tox shell: bash -l {0} run: tox
name: Continuous Integration on: [push] jobs: build: runs-on: ubuntu-latest strategy: fail-fast: false matrix: python-version: ['3.7', '3.8'] steps: - uses: actions/checkout@v2 - uses: conda-incubator/setup-miniconda@v2 with: auto-update-conda: true python-version: ${{ matrix.python }} - name: Install Tox and any other packages # shell: bash -l {0} run: conda install -c conda-forge tox-conda - name: Run Tox # shell: bash -l {0} run: tox -e pytest
Remove shell bash command and only run tox -e pytest.
Remove shell bash command and only run tox -e pytest.
YAML
mit
grmToolbox/grmpy
7a1ba5e4e1668750b89051d1220383ad7f6e8bbc
spec/acceptance/nodesets/fedora-22.yml
spec/acceptance/nodesets/fedora-22.yml
HOSTS: fedora-22: roles: - master platform: fedora-22-x86_64 box: boxcutter/fedora22 box_url: https://atlas.hashicorp.com/boxcutter/boxes/fedora22 hypervisor: vagrant CONFIG: log_level: verbose type: foss
HOSTS: fedora-22: roles: - master platform: fedora-22-x86_64 box: fedora/22-cloud-base box_url: https://dl.fedoraproject.org/pub/fedora/linux/releases/22/Cloud/x86_64/Images/Fedora-Cloud-Base-Vagrant-22-20150521.x86_64.vagrant-virtualbox.box hypervisor: vagrant CONFIG: log_level: verbose type: foss
Use official Fedora 22 base box for Beaker tests
Use official Fedora 22 base box for Beaker tests Signed-off-by: Daniel Farrell <65eaad88897e615d9128b5e890b98414aac720cc@redhat.com>
YAML
bsd-2-clause
srisankethu/puppet-opendaylight,srisankethu/puppet-opendaylight,trozet/puppet-opendaylight,trozet/puppet-opendaylight,umeshksingla/puppet-opendaylight,umeshksingla/puppet-opendaylight
a9f8fe9eaf2e4cec0189fa6db410b38f1a1ecb28
.pyup.yml
.pyup.yml
# see https://pyup.io/docs/configuration/ for all available options schedule: "every week on monday" # Security updates only - not just any newer version update: insecure search: False requirements: - requirements.in - requirements_for_test.txt
# see https://pyup.io/docs/configuration/ for all available options schedule: "every week on monday" # Security updates only - not just any newer version update: insecure search: False requirements: - requirements.in: update: insecure - requirements_for_test.txt: update: insecure
Fix PyUp insecure-only version bumps.
Fix PyUp insecure-only version bumps. PyUp docs: ``` requirements: - requirements/staging.txt: # update all dependencies and pin them update: all pin: True - requirements/dev.txt: # don't update dependencies, use global 'pin' default update: False - requirements/prod.txt: # update insecure only, pin all update: insecure pin: True ``` Despite their own documentation implying settings configured at a global level apply everywhere, it's not true. https://github.com/pyupio/pyup/issues/238
YAML
mit
alphagov/notifications-admin,alphagov/notifications-admin,alphagov/notifications-admin,alphagov/notifications-admin
f57be67dddf857c85e202abdea064a1fa9288ede
tests/all-roles.yml
tests/all-roles.yml
--- - name: Dummy playbook that includes all roles for syntax checking hosts: all sudo: True roles: - development - multimedia - office - vagrant
--- - name: Dummy playbook that includes all roles for syntax checking hosts: all sudo: True roles: - development - multimedia - office
Remove Vagrant roles of tests
Remove Vagrant roles of tests
YAML
mit
JeanFred/APPA
4ffa7e5c7a0ea1068b4bdead028224b837b573fd
spec/support/broken/site.yml
spec/support/broken/site.yml
:site: :name: My broken site :url: http://broke.it :copyright: 2010 :how-to-break-it: Just start writing like you don't know what you're doing.
:site: :name: My broken site :url: http://broke.it :copyright: 2010 :how-to-br eak-it: Just start writing like you don't know what you're doing.
Use really broken yml to test
Use really broken yml to test
YAML
mit
benschwarz/bonsai,benschwarz/bonsai,benschwarz/bonsai
1a73d55bae86fc746b8e64dafd4ca60007401522
roles/ceph-mon/tasks/docker/create_configs.yml
roles/ceph-mon/tasks/docker/create_configs.yml
--- - name: create ceph conf directory file: path: /etc/ceph state: directory owner: root group: root mode: 0644 - name: generate ceph configuration file config_template: src: "{{ playbook_dir }}/roles/ceph-common/templates/ceph.conf.j2" dest: /etc/ceph/ceph.conf owner: "root" group: "root" mode: "0644" config_overrides: "{{ ceph_conf_overrides }}" config_type: ini
--- - name: create ceph conf directory file: path: /etc/ceph state: directory owner: root group: root mode: 0644 - name: generate ceph configuration file action: config_template args: src: "{{ playbook_dir }}/roles/ceph-common/templates/ceph.conf.j2" dest: /etc/ceph/ceph.conf owner: "root" group: "root" mode: "0644" config_overrides: "{{ ceph_conf_overrides }}" config_type: ini
Fix mis-use of config_template in docker tasks
Fix mis-use of config_template in docker tasks As written, generating the config file for ceph-mon in Docker yielded: ERROR: config_template is not a legal parameter in an Ansible task or handler This fixes that error condition.
YAML
apache-2.0
guits/ceph-ansible,mcsage/ceph-ansible,ceph/ceph-ansible,rootfs/ceph-ansible,dang/ceph-ansible,jfcoz/ceph-ansible,font/ceph-ansible,travmi/ceph-ansible,bengland2/ceph-ansible,jsaintrocc/ceph-ansible,albertomurillo/ceph-ansible,fgal/ceph-ansible,jfcoz/ceph-ansible,jtaleric/ceph-ansible,fgal/ceph-ansible,verdurin/ceph-ansible-1,guits/ceph-ansible,ceph/ceph-ansible,WingkaiHo/ceph-ansible,jtaleric/ceph-ansible,albertomurillo/ceph-ansible,bengland2/ceph-ansible,jsaintrocc/ceph-ansible,rootfs/ceph-ansible,mcsage/ceph-ansible,travmi/ceph-ansible,WingkaiHo/ceph-ansible,verdurin/ceph-ansible-1,font/ceph-ansible,WingkaiHo/ceph-ansible,albertomurillo/ceph-ansible,dang/ceph-ansible
600d88a261ec4a04b822d4b3de4eefe07ea0b09d
charts/lagoon-logging/templates/logging.yaml
charts/lagoon-logging/templates/logging.yaml
apiVersion: logging.banzaicloud.io/v1beta1 kind: Logging metadata: namespace: {{ .Release.Namespace | quote }} name: {{ include "lagoon-logging.fullname" . }} labels: {{- include "lagoon-logging.labels" . | nindent 4 }} spec: fluentd: {} fluentbit: {} controlNamespace: {{ .Release.Namespace | quote }}
apiVersion: logging.banzaicloud.io/v1beta1 kind: Logging metadata: namespace: {{ .Release.Namespace | quote }} name: {{ include "lagoon-logging.fullname" . }} labels: {{- include "lagoon-logging.labels" . | nindent 4 }} spec: fluentd: security: podSecurityContext: runAsUser: 100 fluentbit: {} controlNamespace: {{ .Release.Namespace | quote }}
Define PodSecurityContext for OpenShift compatibility
Define PodSecurityContext for OpenShift compatibility
YAML
apache-2.0
amazeeio/lagoon,amazeeio/lagoon,amazeeio/lagoon,amazeeio/lagoon,amazeeio/lagoon,amazeeio/lagoon,amazeeio/lagoon
3591429347aaa2b8df502f37a6ef54da0ce6cc7b
tasks/promote-candidate-to-rc/task.yml
tasks/promote-candidate-to-rc/task.yml
--- platform: linux image_resource: type: docker-image source: repository: cfbuildpacks/ci inputs: - name: buildpacks-ci - name: candidate - name: version outputs: - name: release-candidate run: buildpacks-ci/tasks/promote-candidate-to-rc/run.sh
--- platform: linux image_resource: type: docker-image source: repository: cfbuildpacks/ci inputs: - name: buildpacks-ci - name: candidate - name: version outputs: - name: release-candidate run: path: buildpacks-ci/tasks/promote-candidate-to-rc/run.sh
Add correct path to promote-candidate-to-rc
Add correct path to promote-candidate-to-rc Co-authored-by: David Freilich <0106e5ee349bb58d7c51800b6acf5b57cf80aef6@pivotal.io>
YAML
apache-2.0
cloudfoundry/buildpacks-ci,cloudfoundry/buildpacks-ci,cloudfoundry/buildpacks-ci,cloudfoundry/buildpacks-ci,cloudfoundry/buildpacks-ci