commit stringlengths 40 40 | old_file stringlengths 4 237 | new_file stringlengths 4 237 | old_contents stringlengths 1 4.24k | new_contents stringlengths 1 4.87k | subject stringlengths 15 778 | message stringlengths 15 8.75k | lang stringclasses 266 values | license stringclasses 13 values | repos stringlengths 5 127k |
|---|---|---|---|---|---|---|---|---|---|
ca882ab59a4474ca33e780595221b21e8f3bc53a | .kitchen.yml | .kitchen.yml | ---
provisioner:
name: puppet_apply
files_path: files
manifests_path: test/integration
driver:
name: docker
platforms:
- name: centos-6
driver_config:
image: dynatrace/centos-testing:6
- name: debian-7.8
driver_config:
image: dynatrace/debian-testing:7.8
- name: ubuntu-12.04
driver_config:
image: dynatrace/ubuntu-testing:12.04
suites:
- name: apache_wsagent_debian
provisioner:
manifest: apache_wsagent_debian/init.pp
includes:
- debian-7.8
- ubuntu-12.04
- name: apache_wsagent_redhat
provisioner:
manifest: apache_wsagent_redhat/init.pp
includes:
- centos-6
- name: agents_package
provisioner:
manifest: agents_package/init.pp
- name: collector
provisioner:
manifest: collector/init.pp
- name: java_agent
provisioner:
manifest: java_agent/init.pp
- name: memory_analysis_server
provisioner:
manifest: memory_analysis_server/init.pp
- name: server
provisioner:
manifest: server/init.pp
- name: wsagent_package
provisioner:
manifest: wsagent_package/init.pp
| ---
provisioner:
name: puppet_apply
files_path: files
manifests_path: test/integration
driver:
name: docker
platforms:
- name: centos-6
driver_config:
image: dynatrace/centos-puppet-testing:6
- name: debian-7.8
driver_config:
image: dynatrace/debian-puppet-testing:7.8
- name: ubuntu-12.04
driver_config:
image: dynatrace/ubuntu-puppet-testing:12.04
suites:
- name: apache_wsagent_debian
provisioner:
manifest: apache_wsagent_debian/init.pp
includes:
- debian-7.8
- ubuntu-12.04
- name: apache_wsagent_redhat
provisioner:
manifest: apache_wsagent_redhat/init.pp
includes:
- centos-6
- name: agents_package
provisioner:
manifest: agents_package/init.pp
- name: collector
provisioner:
manifest: collector/init.pp
- name: java_agent
provisioner:
manifest: java_agent/init.pp
- name: memory_analysis_server
provisioner:
manifest: memory_analysis_server/init.pp
- name: server
provisioner:
manifest: server/init.pp
- name: wsagent_package
provisioner:
manifest: wsagent_package/init.pp
| Make use of 'dynatrace/$os-puppet-testing' Docker images. | Make use of 'dynatrace/$os-puppet-testing' Docker images.
| YAML | mit | nespresso/Dynatrace-Puppet,dynaTrace/Dynatrace-Puppet,dynaTrace/Dynatrace-Puppet,nespresso/Dynatrace-Puppet,nespresso/Dynatrace-Puppet,dynaTrace/Dynatrace-Puppet |
cb408b717c17fd7c99b3cb93b003ae4ee52fb828 | .kitchen.yml | .kitchen.yml | driver:
name: vagrant
provisioner:
name: chef_zero
deprecations_as_errors: true
verifier:
name: inspec
platforms:
- name: amazonlinux
driver_config:
box: mvbcoding/awslinux
- name: centos-6
- name: centos-7
- name: debian-8
- name: debian-9
- name: fedora-26
- name: freebsd-10
- name: freebsd-11
- name: opensuse-leap-42
- name: ubuntu-14.04
- name: ubuntu-16.04
- name: solaris-11
driver:
box: chef/solaris-11
- name: windows-2008r2-standard
driver:
box: chef/windows-server-2008r2-standard # private box in Chef's Atlas account
- name: windows-2012r2
driver_config:
box: chef/windows-server-2012r2-standard
- name: macosx-10.11
driver:
box: chef/macosx-10.11 # private box in Chef's Atlas account
provider: vmware_fusion
suites:
- name: default
run_list: test::default
- name: compile_time
run_list: test::default
attributes:
build-essential:
compile_time: true
| driver:
name: vagrant
provisioner:
name: chef_zero
deprecations_as_errors: true
verifier:
name: inspec
platforms:
- name: amazonlinux
driver_config:
box: mvbcoding/awslinux
- name: centos-6
- name: centos-7
- name: debian-8
- name: debian-9
- name: fedora-27
- name: freebsd-10
- name: freebsd-11
- name: opensuse-leap-42
- name: ubuntu-14.04
- name: ubuntu-16.04
- name: solaris-11
driver:
box: chef/solaris-11
- name: windows-2012r2-standard
driver:
box: chef/windows-server-2012r2-standard # private box in Chef's Atlas account
- name: windows-2016
driver_config:
box: chef/windows-server-2016-standard # private box in Chef's Atlas account
- name: macos-10.12
driver:
box: chef/macos-10.12 # private box in Chef's Atlas account
provider: vmware_fusion
suites:
- name: default
run_list: test::default
- name: compile_time
run_list: test::default
attributes:
build-essential:
compile_time: true
| Update Fedora, Windows, and MacOS test boxes | Update Fedora, Windows, and MacOS test boxes
Signed-off-by: Tim Smith <764ef62106582a09ed09dfa0b6bff7c05fd7d1e4@chef.io>
| YAML | apache-2.0 | opscode-cookbooks/build-essential,chef-cookbooks/build-essential |
76a7917dcfd862abf406432f77865400f8fb9b9d | .kitchen.yml | .kitchen.yml | driver:
name: vagrant
customize:
cpus: 2
memory: 4096
provisioner:
name: chef_zero
verifier:
name: pester
platforms:
- name: windows-2008r2
driver_config:
box: chef/windows-server-2008r2-standard
- name: windows-2012r2
driver_config:
box: chef/windows-server-2012r2-standard
suites:
- name: autorun
run_list:
- recipe[test::autorun]
- name: certificate
run_list:
- recipe[test::certificate]
- name: feature
run_list:
- recipe[test::feature]
- name: font
run_list:
- recipe[test::font]
- name: http_acl
run_list:
- recipe[test::http_acl]
- name: package
run_list:
- recipe[test::package]
- name: path
run_list:
- recipe[test::path]
- name: tasks
run_list:
- recipe[test::tasks]
| driver:
name: vagrant
customize:
cpus: 2
memory: 4096
transport:
name: winrm
elevated: true
provisioner:
name: chef_zero
verifier:
name: pester
platforms:
- name: windows-7
driver_config:
box: chef/windows-7-professional
- name: windows-8.1
driver_config:
box: chef/windows-8.1-professional
- name: windows-2008r2
driver_config:
box: chef/windows-server-2008r2-standard
- name: windows-2012r2
driver_config:
box: chef/windows-server-2012r2-standard
- name: windows-2012r2-chef-12.1
driver_config:
box: chef/windows-server-2012r2-standard
provisioner:
require_chef_omnibus: 12.1.0
suites:
- name: autorun
run_list:
- recipe[test::autorun]
- name: certificate
run_list:
- recipe[test::certificate]
- name: feature
run_list:
- recipe[test::feature]
- name: font
run_list:
- recipe[test::font]
- name: http_acl
run_list:
- recipe[test::http_acl]
- name: package
run_list:
- recipe[test::package]
- name: path
run_list:
- recipe[test::path]
- name: tasks
run_list:
- recipe[test::tasks]
| Test on more platforms / chef releases | Test on more platforms / chef releases
Signed-off-by: Tim Smith <764ef62106582a09ed09dfa0b6bff7c05fd7d1e4@chef.io>
| YAML | apache-2.0 | protec-cookbooks/windows,juliandunn/windows,blairham/windows-cookbook,stuartpreston/windows,daptiv/windows,chef-cookbooks/windows,ruicovelo/windows,criteo-forks/windows,autoclone/windows,opscode-cookbooks/windows,urban-institute/windows,GannettDigital/chef-windows,shoekstra/chef-windows,rlaveycal/windows |
01f4152a7aeb927dce940d3ea2e29df0a623baf1 | .resinci.yml | .resinci.yml | docker:
publish: false
| npm:
platforms:
- name: linux
os: ubuntu
architecture: x86_64
node_versions:
- "8"
- "10"
- name: linux
os: alpine
architecture: x86_64
node_versions:
- "8"
- "10"
- name: linux
os: ubuntu
architecture: x86
node_versions:
- "8"
- "10"
docker:
publish: false
| Exclude node 6 builds from CI | Exclude node 6 builds from CI
Change-type: patch
Signed-off-by: Giovanni Garufi <85dd344344db4c4ce024b7c00609e96849fa96e3@balena.io>
| YAML | mit | resin-io-modules/resin-components,resin-io-modules/resin-components,resin-io-modules/resin-components |
f6612ba12ab794d25020e476e4f56e465df4837f | .rubocop.yml | .rubocop.yml | Encoding:
Enabled: false
# The default 79 character limit is sometimes unproductive.
LineLength:
Max: 120
# The default 10 lines limit is unrealistic.
MethodLength:
Max: 50
# Lining up the attributes of some methods is advantageous for readability
SingleSpaceBeforeFirstArg:
Enabled: false
AllCops:
Exclude:
- 'Guardfile'
- 'Rakefile'
- 'vendor/bundle/**/*'
FileName:
Enabled: false
Documentation:
Enabled: false
AlignParameters:
Exclude:
- 'metadata.rb'
| Encoding:
Enabled: false
# The default 79 character limit is sometimes unproductive.
LineLength:
Max: 120
# The default 10 lines limit is unrealistic.
MethodLength:
Max: 50
# Lining up the attributes of some methods is advantageous for readability
SpaceBeforeFirstArg:
Enabled: false
AllCops:
Exclude:
- 'Guardfile'
- 'Rakefile'
- 'vendor/bundle/**/*'
FileName:
Enabled: false
Documentation:
Enabled: false
AlignParameters:
Exclude:
- 'metadata.rb'
| Update RuboCop config as per warning | Update RuboCop config as per warning
| YAML | mit | evertrue/mesos-cookbook,evertrue/mesos-cookbook,evertrue/mesos-cookbook |
09a0360ad980e4108d753f94fd3e43c6293ea18d | .rubocop.yml | .rubocop.yml | # Disable all metrics.
Metrics/AbcSize:
Enabled: false
Metrics/BlockNesting:
Enabled: false
Metrics/ClassLength:
Enabled: false
Metrics/ModuleLength:
Enabled: false
Metrics/CyclomaticComplexity:
Enabled: false
Metrics/LineLength:
Enabled: false
Metrics/MethodLength:
Enabled: false
Metrics/ParameterLists:
Enabled: false
Metrics/PerceivedComplexity:
Enabled: false
# Prefer compact module/class defs
Style/ClassAndModuleChildren:
Enabled: false
# Allow CamelCase constants as well as SCREAMING_SNAKE_CASE
Style/ConstantName:
Enabled: false
# Allow 'Pokémon-style' exception handling
Lint/RescueException:
Enabled: false
AllCops:
TargetRubyVersion: 2.1
# Apparently setting the target version to 2.1 is not enough to prevent this cop...
Style/FrozenStringLiteralComment:
Enabled: false
# http://stackoverflow.com/questions/4763121/should-i-use-alias-or-alias-method
Style/Alias:
Enabled: false
| # Disable all metrics.
Metrics/AbcSize:
Enabled: false
Metrics/BlockNesting:
Enabled: false
Metrics/ClassLength:
Enabled: false
Metrics/ModuleLength:
Enabled: false
Metrics/CyclomaticComplexity:
Enabled: false
Metrics/LineLength:
Enabled: false
Metrics/MethodLength:
Enabled: false
Metrics/ParameterLists:
Enabled: false
Metrics/PerceivedComplexity:
Enabled: false
# Prefer compact module/class defs
Style/ClassAndModuleChildren:
Enabled: false
# Allow CamelCase constants as well as SCREAMING_SNAKE_CASE
Style/ConstantName:
Enabled: false
# Allow 'Pokémon-style' exception handling
Lint/RescueException:
Enabled: false
AllCops:
TargetRubyVersion: 2.1
# Apparently setting the target version to 2.1 is not enough to prevent this cop...
Style/FrozenStringLiteralComment:
Enabled: false
# http://stackoverflow.com/questions/4763121/should-i-use-alias-or-alias-method
Style/Alias:
Enabled: false
# So RuboCop doesn't complain about application IDs
Style/NumericLiterals:
Exclude:
- examples/* | Make RuboCop ignore numeric literals in examples | Make RuboCop ignore numeric literals in examples
| YAML | mit | Roughsketch/discordrb,megumisonoda/discordrb,Roughsketch/discordrb,meew0/discordrb,megumisonoda/discordrb,VxJasonxV/discordrb,VxJasonxV/discordrb,meew0/discordrb |
3dcbd78755e3ad85b7fa6b3f53a5e0d813fa65dd | slimmer_headers.yml | slimmer_headers.yml | :section: "Search"
:format: "search"
:proposition: "citizen"
:template: "wrapper_with_fat_footer"
| :section: "Search"
:format: "search"
:proposition: "citizen"
:template: "wrapper"
| Use standard slimmer template, not fat_footer version. | Use standard slimmer template, not fat_footer version.
| YAML | mit | alphagov/rummager,alphagov/rummager,theodi/rummager,theodi/rummager |
f9146697ee4b848cf988900fd115fa9fabd3aff2 | hiera/data/repositories.yaml | hiera/data/repositories.yaml | # Repository Management
openstack_extras::repo::redhat::redhat::repo_hash:
'CentOS-Base':
'descr': 'CentOS-$releasever - Base'
'baseurl': "%{hiera('yum_base_mirror')}/$releasever/os/$basearch/"
'CentOS-Updates':
'descr': 'CentOS-$releasever - Updates'
'baseurl': "%{hiera('yum_base_mirror')}/$releasever/updates/$basearch/"
'CentOS-Extras':
'descr': 'CentOS-$releasever - Extras'
'baseurl': "%{hiera('yum_base_mirror')}/$releasever/extras/$basearch/"
'epel':
'descr': 'Extra Packages for Enterprise Linux 6 - $basearch'
'baseurl': "%{hiera('yum_epel_mirror')}/$releasever/$basearch/"
'gpgkey': 'file:///etc/pki/rpm-gpg/RPM-GPG-KEY-EPEL-6'
'failovermethod': 'priority'
openstack_extras::repo::redhat::redhat::gpgkey_hash:
'/etc/pki/rpm-gpg/RPM-GPG-KEY-EPEL-6':
'source': 'puppet:///modules/openstack_extras/RPM-GPG-KEY-EPEL-6'
openstack_extras::repo::redhat::redhat::purge_unmanaged: true
openstack_extras::repo::redhat::redhat::package_require: true
openstack_extras::repo::redhat::redhat::manage_epel: false
| # Repository Management
openstack_extras::repo::redhat::redhat::repo_hash:
'CentOS-Base':
'descr': 'CentOS-$releasever - Base'
'baseurl': "%{hiera('yum_base_mirror')}/$releasever/os/$basearch/"
'CentOS-Updates':
'descr': 'CentOS-$releasever - Updates'
'baseurl': "%{hiera('yum_base_mirror')}/$releasever/updates/$basearch/"
'CentOS-Extras':
'descr': 'CentOS-$releasever - Extras'
'baseurl': "%{hiera('yum_base_mirror')}/$releasever/extras/$basearch/"
'epel':
'descr': "Extra Packages for Enterprise Linux %{operatingsystemmajrelease} - $basearch"
'baseurl': "%{hiera('yum_epel_mirror')}/$releasever/$basearch/"
'gpgkey': 'file:///etc/pki/rpm-gpg/RPM-GPG-KEY-EPEL-7'
'failovermethod': 'priority'
openstack_extras::repo::redhat::redhat::gpgkey_hash:
'/etc/pki/rpm-gpg/RPM-GPG-KEY-EPEL-7':
'source': 'puppet:///modules/openstack_extras/RPM-GPG-KEY-EPEL-7'
openstack_extras::repo::redhat::redhat::purge_unmanaged: true
openstack_extras::repo::redhat::redhat::package_require: true
openstack_extras::repo::redhat::redhat::manage_epel: false
| Move to centos 7 from 6 | Move to centos 7 from 6
| YAML | apache-2.0 | michaeltchapman/puppet_openstack_builder,michaeltchapman/puppet_openstack_builder,phchoic/puppet_openstack_builder,phchoic/puppet_openstack_builder,michaeltchapman/vagrant-consul,michaeltchapman/puppet_openstack_builder,michaeltchapman/vagrant-consul,phchoic/puppet_openstack_builder |
0209bdc69a55a104e52b9d21356b20b95268dd84 | .vsts-ci.yml | .vsts-ci.yml | variables:
deployment: deploy
nuget: packages
steps:
- task: NuGetCommand@2
inputs:
command: 'restore'
packDestination: $(nuget)
projects: WhiteLibrary
- script: |
pip install -r requirements.txt
displayName: Install needed compoments from PyPi
- task: VSBuild@1
inputs:
solution: UIAutomationTest\UIAutomationTest.csproj
- task: Bash@3
inputs:
targetType: inline
script: echo $BUILD_BUILDNUMBER >> build.info
- script: |
set DEPLOYMENT=$(deployment)
deploy.cmd
continueOnError: false
displayName: Copy build artifacts to deployment directory
- script: |
set DEPLOYMENT=$(deployment)
run_tests.cmd --exclude no_ci --xunit xunit.xml
continueOnError: true
displayName: Execute Acceptance tests
- task: PublishTestResults@2
inputs:
testRunner: JUnit
testResultsFiles: output/xunit.xml
- task: PublishBuildArtifacts@1
displayName: Publish Robot Framework logs and results
inputs:
PathtoPublish: output
ArtifactType: container
TargetPath: ""
ArtifactName: Robot Test results
- task: PublishBuildArtifacts@1
displayName: Publish WhiteLibrary
inputs:
PathToPublish: $(deployment)
ArtifactType: container
TargetPath: ""
ArtifactName: WhiteLibrary
| variables:
deployment: deploy
nuget: packages
steps:
- task: NuGetCommand@2
inputs:
command: 'restore'
packDestination: $(nuget)
projects: WhiteLibrary
- script: |
pip install -r requirements.txt
displayName: Install needed compoments from PyPi
- task: VSBuild@1
inputs:
solution: UIAutomationTest\UIAutomationTest.csproj
- task: Bash@3
inputs:
targetType: inline
script: echo $BUILD_BUILDNUMBER >> build.info
displayName: Generate build.info
- script: |
set DEPLOYMENT=$(deployment)
deploy.cmd
continueOnError: false
displayName: Copy build artifacts to deployment directory
- script: |
set DEPLOYMENT=$(deployment)
run_tests.cmd --exclude no_ci --xunit xunit.xml
continueOnError: true
displayName: Execute Acceptance tests
- task: PublishTestResults@2
inputs:
testRunner: JUnit
testResultsFiles: output/xunit.xml
- task: PublishBuildArtifacts@1
displayName: Publish Robot Framework logs and results
inputs:
PathtoPublish: output
ArtifactType: container
TargetPath: ""
ArtifactName: Robot Test results
- task: PublishBuildArtifacts@1
displayName: Publish WhiteLibrary
inputs:
PathToPublish: $(deployment)
ArtifactType: container
TargetPath: ""
ArtifactName: WhiteLibrary
| Add display name to bash step | Add display name to bash step
| YAML | apache-2.0 | Omenia/robotframework-whitelibrary,Omenia/robotframework-whitelibrary |
bde1ec3120116ff5a0f1299baeea751b08eb3530 | recipes/umi_tools/meta.yaml | recipes/umi_tools/meta.yaml | package:
name: "umi_tools"
version: "0.2.6"
source:
fn: v0.2.6.tar.gz
url: https://github.com/CGATOxford/UMI-tools/archive/v0.2.6.tar.gz
build:
number: 0
requirements:
build:
- python
- setuptools
- cython
- numpy
- pandas
- pysam
- future
run:
- python
- cython
- numpy
- pandas
- pysam
- future
test:
commands:
- umi_tools --help
about:
home: https://github.com/CGATOxford/UMI-tools
license: MIT License
summary: 'Tools for dealing with Unique Molecular Identifiers'
| package:
name: "umi_tools"
version: "0.2.3"
source:
fn: v0.2.3.tar.gz
url: https://github.com/CGATOxford/UMI-tools/archive/v0.2.3.tar.gz
md5: ebf34023a9fedb0ce87a0b74e3e655a8
build:
number: 2
skip: True # [py2k]
requirements:
build:
- python
- setuptools >=1.1
- cython >=0.19
- numpy >=1.7
- pandas >=0.12.0
- pysam >=0.8.4
- future
run:
- python
- cython >=0.19
- numpy >=1.7
- pandas >=0.12.0
- pysam >=0.8.4
- future
test:
commands:
- umi_tools --help
about:
home: https://github.com/CGATOxford/UMI-tools
license: MIT License
summary: 'Tools for dealing with Unique Molecular Identifiers'
| Revert " bumped UMI-tools version" | Revert " bumped UMI-tools version"
This reverts commit 399eef73392d14ec4290627b64e3e3aef2873d39.
| YAML | mit | blankenberg/bioconda-recipes,dkoppstein/recipes,dmaticzka/bioconda-recipes,jasper1918/bioconda-recipes,joachimwolff/bioconda-recipes,chapmanb/bioconda-recipes,pinguinkiste/bioconda-recipes,jfallmann/bioconda-recipes,peterjc/bioconda-recipes,JenCabral/bioconda-recipes,rvalieris/bioconda-recipes,chapmanb/bioconda-recipes,lpantano/recipes,shenwei356/bioconda-recipes,CGATOxford/bioconda-recipes,bioconda/recipes,rob-p/bioconda-recipes,gregvonkuster/bioconda-recipes,joachimwolff/bioconda-recipes,JenCabral/bioconda-recipes,ThomasWollmann/bioconda-recipes,xguse/bioconda-recipes,matthdsm/bioconda-recipes,rvalieris/bioconda-recipes,omicsnut/bioconda-recipes,keuv-grvl/bioconda-recipes,gvlproject/bioconda-recipes,dmaticzka/bioconda-recipes,mcornwell1957/bioconda-recipes,Luobiny/bioconda-recipes,cokelaer/bioconda-recipes,bioconda/bioconda-recipes,instituteofpathologyheidelberg/bioconda-recipes,phac-nml/bioconda-recipes,acaprez/recipes,joachimwolff/bioconda-recipes,npavlovikj/bioconda-recipes,joachimwolff/bioconda-recipes,hardingnj/bioconda-recipes,rob-p/bioconda-recipes,chapmanb/bioconda-recipes,dkoppstein/recipes,instituteofpathologyheidelberg/bioconda-recipes,npavlovikj/bioconda-recipes,bow/bioconda-recipes,HassanAmr/bioconda-recipes,dmaticzka/bioconda-recipes,ostrokach/bioconda-recipes,blankenberg/bioconda-recipes,guowei-he/bioconda-recipes,ThomasWollmann/bioconda-recipes,bioconda/bioconda-recipes,ostrokach/bioconda-recipes,zachcp/bioconda-recipes,JenCabral/bioconda-recipes,bow/bioconda-recipes,peterjc/bioconda-recipes,gvlproject/bioconda-recipes,CGATOxford/bioconda-recipes,cokelaer/bioconda-recipes,pinguinkiste/bioconda-recipes,instituteofpathologyheidelberg/bioconda-recipes,oena/bioconda-recipes,bebatut/bioconda-recipes,bebatut/bioconda-recipes,jasper1918/bioconda-recipes,oena/bioconda-recipes,ostrokach/bioconda-recipes,abims-sbr/bioconda-recipes,lpantano/recipes,martin-mann/bioconda-recipes,mdehollander/bioconda-recipes,Luobiny/bioconda-recipes,joachimwolff/bioconda-recipes,matthdsm/bioconda-recipes,saketkc/bioconda-recipes,HassanAmr/bioconda-recipes,daler/bioconda-recipes,rvalieris/bioconda-recipes,phac-nml/bioconda-recipes,colinbrislawn/bioconda-recipes,bioconda/recipes,saketkc/bioconda-recipes,dkoppstein/recipes,JenCabral/bioconda-recipes,abims-sbr/bioconda-recipes,phac-nml/bioconda-recipes,ostrokach/bioconda-recipes,bebatut/bioconda-recipes,ivirshup/bioconda-recipes,zwanli/bioconda-recipes,gregvonkuster/bioconda-recipes,daler/bioconda-recipes,colinbrislawn/bioconda-recipes,JenCabral/bioconda-recipes,gvlproject/bioconda-recipes,gvlproject/bioconda-recipes,hardingnj/bioconda-recipes,bow/bioconda-recipes,ivirshup/bioconda-recipes,keuv-grvl/bioconda-recipes,mdehollander/bioconda-recipes,zwanli/bioconda-recipes,bow/bioconda-recipes,rvalieris/bioconda-recipes,npavlovikj/bioconda-recipes,ThomasWollmann/bioconda-recipes,saketkc/bioconda-recipes,bow/bioconda-recipes,zachcp/bioconda-recipes,ThomasWollmann/bioconda-recipes,hardingnj/bioconda-recipes,HassanAmr/bioconda-recipes,CGATOxford/bioconda-recipes,matthdsm/bioconda-recipes,dmaticzka/bioconda-recipes,BIMSBbioinfo/bioconda-recipes,omicsnut/bioconda-recipes,roryk/recipes,gvlproject/bioconda-recipes,mcornwell1957/bioconda-recipes,xguse/bioconda-recipes,ostrokach/bioconda-recipes,abims-sbr/bioconda-recipes,blankenberg/bioconda-recipes,colinbrislawn/bioconda-recipes,daler/bioconda-recipes,jfallmann/bioconda-recipes,guowei-he/bioconda-recipes,phac-nml/bioconda-recipes,BIMSBbioinfo/bioconda-recipes,bow/bioconda-recipes,shenwei356/bioconda-recipes,zwanli/bioconda-recipes,zwanli/bioconda-recipes,cokelaer/bioconda-recipes,martin-mann/bioconda-recipes,oena/bioconda-recipes,pinguinkiste/bioconda-recipes,jasper1918/bioconda-recipes,rvalieris/bioconda-recipes,guowei-he/bioconda-recipes,pinguinkiste/bioconda-recipes,jasper1918/bioconda-recipes,saketkc/bioconda-recipes,keuv-grvl/bioconda-recipes,blankenberg/bioconda-recipes,dmaticzka/bioconda-recipes,CGATOxford/bioconda-recipes,BIMSBbioinfo/bioconda-recipes,Luobiny/bioconda-recipes,roryk/recipes,daler/bioconda-recipes,omicsnut/bioconda-recipes,guowei-he/bioconda-recipes,instituteofpathologyheidelberg/bioconda-recipes,mdehollander/bioconda-recipes,ivirshup/bioconda-recipes,mcornwell1957/bioconda-recipes,colinbrislawn/bioconda-recipes,xguse/bioconda-recipes,acaprez/recipes,jfallmann/bioconda-recipes,gregvonkuster/bioconda-recipes,keuv-grvl/bioconda-recipes,mdehollander/bioconda-recipes,lpantano/recipes,HassanAmr/bioconda-recipes,colinbrislawn/bioconda-recipes,saketkc/bioconda-recipes,rob-p/bioconda-recipes,daler/bioconda-recipes,BIMSBbioinfo/bioconda-recipes,CGATOxford/bioconda-recipes,peterjc/bioconda-recipes,oena/bioconda-recipes,abims-sbr/bioconda-recipes,acaprez/recipes,martin-mann/bioconda-recipes,omicsnut/bioconda-recipes,lpantano/recipes,cokelaer/bioconda-recipes,joachimwolff/bioconda-recipes,shenwei356/bioconda-recipes,mcornwell1957/bioconda-recipes,ThomasWollmann/bioconda-recipes,pinguinkiste/bioconda-recipes,ostrokach/bioconda-recipes,gvlproject/bioconda-recipes,keuv-grvl/bioconda-recipes,peterjc/bioconda-recipes,mdehollander/bioconda-recipes,rob-p/bioconda-recipes,mcornwell1957/bioconda-recipes,JenCabral/bioconda-recipes,dmaticzka/bioconda-recipes,BIMSBbioinfo/bioconda-recipes,bioconda/bioconda-recipes,HassanAmr/bioconda-recipes,zwanli/bioconda-recipes,hardingnj/bioconda-recipes,matthdsm/bioconda-recipes,abims-sbr/bioconda-recipes,pinguinkiste/bioconda-recipes,instituteofpathologyheidelberg/bioconda-recipes,bioconda/recipes,xguse/bioconda-recipes,shenwei356/bioconda-recipes,npavlovikj/bioconda-recipes,phac-nml/bioconda-recipes,chapmanb/bioconda-recipes,mdehollander/bioconda-recipes,hardingnj/bioconda-recipes,matthdsm/bioconda-recipes,martin-mann/bioconda-recipes,colinbrislawn/bioconda-recipes,HassanAmr/bioconda-recipes,martin-mann/bioconda-recipes,roryk/recipes,rvalieris/bioconda-recipes,ThomasWollmann/bioconda-recipes,chapmanb/bioconda-recipes,jasper1918/bioconda-recipes,ivirshup/bioconda-recipes,daler/bioconda-recipes,matthdsm/bioconda-recipes,ivirshup/bioconda-recipes,zachcp/bioconda-recipes,oena/bioconda-recipes,Luobiny/bioconda-recipes,keuv-grvl/bioconda-recipes,gregvonkuster/bioconda-recipes,BIMSBbioinfo/bioconda-recipes,jfallmann/bioconda-recipes,peterjc/bioconda-recipes,zwanli/bioconda-recipes,xguse/bioconda-recipes,zachcp/bioconda-recipes,peterjc/bioconda-recipes,omicsnut/bioconda-recipes,bioconda/bioconda-recipes,bebatut/bioconda-recipes,acaprez/recipes,abims-sbr/bioconda-recipes,CGATOxford/bioconda-recipes,saketkc/bioconda-recipes,guowei-he/bioconda-recipes,ivirshup/bioconda-recipes |
e55b89096f9f30c5a80f1fb8252c7239d236d24a | .github/FUNDING.yml | .github/FUNDING.yml | # These are supported funding model platforms
github: # Replace with up to 4 GitHub Sponsors-enabled usernames e.g., [user1, user2]
patreon: # Replace with a single Patreon username
open_collective: # Replace with a single Open Collective username
ko_fi: # Replace with a single Ko-fi username
tidelift: # Replace with a single Tidelift platform-name/package-name e.g., npm/babel
community_bridge: # Replace with a single Community Bridge project-name e.g., cloud-foundry
liberapay: # Replace with a single Liberapay username
issuehunt: # Replace with a single IssueHunt username
otechie: # Replace with a single Otechie username
custom: ['https://nlnetlabs.nl/funding/']
| github: [NLnetLabs]
custom: ['https://nlnetlabs.nl/funding/']
| Add GitHub Sponsors for Organisations | Add GitHub Sponsors for Organisations | YAML | bsd-3-clause | NLnetLabs/unbound,NLnetLabs/unbound,NLnetLabs/unbound,NLnetLabs/unbound |
fff54c506096fa54f8a442f1c139978039e7adce | .kitchen.travis.yml | .kitchen.travis.yml | ---
driver:
name: ec2
aws_ssh_key_id: travis
ssh_key: travis.pem
| ---
driver:
name: ec2
aws_ssh_key_id: travis
ssh_key: travis.pem
platforms:
- name: ubuntu-12.04
driver_config:
image_id: ami-4df9cf24
| Use an instance store AMI. | Use an instance store AMI. | YAML | apache-2.0 | balanced-cookbooks/balanced-omnibus |
479f95741ea222c6f659bb819e4755cc6551a73e | roles/common/tasks/main.yml | roles/common/tasks/main.yml | ---
- name: update apt index
apt: update_cache=yes
- name: python dependencies
apt: pkg={{ item }}
with_items:
- python-pip
- python-libxml2
- python-lxml
- python-greenlet
- python-openssl
- python-dev
- python-httplib2
- python-software-properties
- python-virtualenv
- python-mysqldb
- apt: pkg={{ item }}
with_items:
- ntp
- vlan
- name: set UTC timezone
copy: content='Etc/UTC' dest=/etc/timezone owner=root group=root mode=0644
notify:
- update timezone
- name: /opt/stack
file: dest=/opt/stack state=directory
- include: ssl.yml
- include: ssh.yml
- include: networking.yml
- include: system_tools.yml
# run this last so we only update if run was successful
- name: drop an motd with ursula metadata
action: template src=etc/motd.tail dest=/etc/motd.tail mode=0644
| ---
- name: update apt index
apt: update_cache=yes
- name: python dependencies
apt: pkg={{ item }}
with_items:
- python-pip
- python-libxml2
- python-lxml
- python-greenlet
- python-openssl
- python-dev
- python-httplib2
- python-software-properties
- python-virtualenv
- python-mysqldb
- apt: pkg={{ item }}
with_items:
- cryptsetup
- ntp
- vlan
- name: set UTC timezone
copy: content='Etc/UTC' dest=/etc/timezone owner=root group=root mode=0644
notify:
- update timezone
- name: /opt/stack
file: dest=/opt/stack state=directory
- include: ssl.yml
- include: ssh.yml
- include: networking.yml
- include: system_tools.yml
# run this last so we only update if run was successful
- name: drop an motd with ursula metadata
action: template src=etc/motd.tail dest=/etc/motd.tail mode=0644
| Add cryptsetup so that we can map encrypted vols | Add cryptsetup so that we can map encrypted vols
Adding cryptsetup so that we can easily add encrypted volumes and
images.
| YAML | mit | davidcusatis/ursula,knandya/ursula,knandya/ursula,dlundquist/ursula,zrs233/ursula,channus/ursula,EricCrosson/ursula,greghaynes/ursula,fancyhe/ursula,MaheshIBM/ursula,MaheshIBM/ursula,nirajdp76/ursula,masteinhauser/ursula,jwaibel/ursula,lihkin213/ursula,blueboxjesse/ursula,davidcusatis/ursula,davidcusatis/ursula,narengan/ursula,aldevigi/ursula,persistent-ursula/ursula,channus/ursula,wupeiran/ursula,blueboxgroup/ursula,pbannister/ursula,pgraziano/ursula,greghaynes/ursula,paulczar/ursula,nirajdp76/ursula,edtubillara/ursula,msambol/ursula,j2sol/ursula,jwaibel/ursula,ryshah/ursula,ddaskal/ursula,rongzhus/ursula,edtubillara/ursula,j2sol/ursula,fancyhe/ursula,masteinhauser/ursula,persistent-ursula/ursula,ryshah/ursula,aldevigi/ursula,dlundquist/ursula,ryshah/ursula,j2sol/ursula,zrs233/ursula,channus/ursula,edtubillara/ursula,masteinhauser/ursula,ddaskal/ursula,narengan/ursula,panxia6679/ursula,sivakom/ursula,jwaibel/ursula,edtubillara/ursula,EricCrosson/ursula,mjbrewer/ursula,wupeiran/ursula,mjbrewer/ursula,twaldrop/ursula,msambol/ursula,pgraziano/ursula,msambol/ursula,ddaskal/ursula,wupeiran/ursula,andrewrothstein/ursula,rongzhus/ursula,EricCrosson/ursula,zrs233/ursula,kennjason/ursula,lihkin213/ursula,fancyhe/ursula,mjbrewer/ursula,allomov/ursula,panxia6679/ursula,andrewrothstein/ursula,kennjason/ursula,knandya/ursula,sivakom/ursula,andrewrothstein/ursula,rongzhus/ursula,blueboxjesse/ursula,ryshah/ursula,blueboxjesse/ursula,sivakom/ursula,pbannister/ursula,rongzhus/ursula,lihkin213/ursula,dlundquist/ursula,twaldrop/ursula,allomov/ursula,nirajdp76/ursula,nirajdp76/ursula,paulczar/ursula,paulczar/ursula,MaheshIBM/ursula,pgraziano/ursula,wupeiran/ursula,kennjason/ursula,panxia6679/ursula,persistent-ursula/ursula,narengan/ursula,lihkin213/ursula,zrs233/ursula,blueboxgroup/ursula,channus/ursula,blueboxgroup/ursula,j2sol/ursula,narengan/ursula,pgraziano/ursula,panxia6679/ursula,greghaynes/ursula,aldevigi/ursula,fancyhe/ursula,masteinhauser/ursula,twaldrop/ursula,knandya/ursula,twaldrop/ursula,blueboxgroup/ursula,pbannister/ursula,persistent-ursula/ursula,blueboxjesse/ursula,allomov/ursula,ddaskal/ursula |
87d8e5fcef0a15599448042032d4ddba578223c4 | roles/packer/tasks/main.yml | roles/packer/tasks/main.yml | ---
- name: Download Packer
sudo: yes
get_url: url={{ packer_url }} dest={{ packer_install_dir }}
- name: Extract Packer
sudo: yes
unarchive: copy=no src={{ packer_install_dir }}/{{ packer_archive }}.zip dest={{ packer_install_dir }}
- name: Remove downloaded packer_archive
sudo: yes
shell: "rm -f {{ packer_install_dir }}/{{ packer_archive }}.zip"
| ---
- name: Download Packer
sudo: yes
shell: curl -o {{ packer_install_dir }}/{{ packer_archive }}.zip {{ packer_url }}
- name: Extract Packer
sudo: yes
unarchive: copy=no src={{ packer_install_dir }}/{{ packer_archive }}.zip dest={{ packer_install_dir }}
- name: Remove downloaded packer_archive
sudo: yes
shell: "rm -f {{ packer_install_dir }}/{{ packer_archive }}.zip"
| Remove get_url in favour of curl | Remove get_url in favour of curl | YAML | apache-2.0 | snowplow/ansible-playbooks |
260274f82cea6856664d0adce151c221b78c2385 | buildscripts/condarecipe.hsa/meta.yaml | buildscripts/condarecipe.hsa/meta.yaml | package:
name: numba-hsa
version: "0.20.0"
source:
path: ../..
build:
number: {{ environ.get('GIT_DESCRIBE_NUMBER', 0) }}
entry_points:
- pycc = numba.pycc:main
- numba = numba.numba_entry:main
requirements:
# build and run dependencies are duplicated to avoid setuptools issues
# when we also set install_requires in setup.py
build:
- python
- argparse [py26]
- numpy
# On channel https://binstar.org/numba/
- llvmlite 0.6*
- funcsigs [py26 or py27]
- singledispatch [py26 or py27 or py33]
run:
- python
- argparse [py26]
- numpy
# On channel https://binstar.org/numba/
- llvmlite 0.6*
- funcsigs [py26 or py27]
- singledispatch [py26 or py27 or py33]
- libhlc
test:
requires:
- argparse [py26]
- unittest2 [py26]
- jinja2
files:
- mandel.py
commands:
- pycc -h
- numba -h
| package:
name: numba-hsa
version: 0.21.0
source:
path: ../..
build:
number: 0
entry_points:
- pycc = numba.pycc:main
- numba = numba.numba_entry:main
requirements:
# build and run dependencies are duplicated to avoid setuptools issues
# when we also set install_requires in setup.py
build:
- python
- argparse [py26]
- numpy x.x
# On channel https://binstar.org/numba/
- llvmlite 0.8*
- funcsigs [py26 or py27]
- singledispatch [py26 or py27 or py33]
run:
- python
- argparse [py26]
- numpy x.x
# On channel https://binstar.org/numba/
- llvmlite 0.8*
- funcsigs [py26 or py27]
- singledispatch [py26 or py27 or py33]
- libhlc
test:
requires:
- argparse [py26]
- unittest2 [py26]
- jinja2
files:
- mandel.py
commands:
- pycc -h
- numba -h
| Update buildscript for custom HSA builds | Update buildscript for custom HSA builds
| YAML | bsd-2-clause | stefanseefeld/numba,jriehl/numba,pombredanne/numba,sklam/numba,seibert/numba,numba/numba,gmarkall/numba,numba/numba,stefanseefeld/numba,stonebig/numba,stonebig/numba,IntelLabs/numba,numba/numba,stuartarchibald/numba,gmarkall/numba,jriehl/numba,gmarkall/numba,stonebig/numba,sklam/numba,stuartarchibald/numba,numba/numba,sklam/numba,cpcloud/numba,jriehl/numba,sklam/numba,gmarkall/numba,seibert/numba,pombredanne/numba,IntelLabs/numba,cpcloud/numba,jriehl/numba,IntelLabs/numba,pombredanne/numba,IntelLabs/numba,stuartarchibald/numba,numba/numba,gmarkall/numba,stonebig/numba,stefanseefeld/numba,IntelLabs/numba,seibert/numba,stefanseefeld/numba,stefanseefeld/numba,sklam/numba,pombredanne/numba,seibert/numba,jriehl/numba,seibert/numba,stonebig/numba,cpcloud/numba,stuartarchibald/numba,pombredanne/numba,cpcloud/numba,stuartarchibald/numba,cpcloud/numba |
fce7a5af097120a4497317498d0099e0b878bc30 | .codeclimate.yml | .codeclimate.yml | engines:
csslint:
enabled: true
duplication:
enabled: true
config:
languages:
- python
- javascript
fixme:
enabled: true
pep8:
enabled: true
radon:
enabled: true
shellcheck:
enabled: true
ratings:
paths:
- app/**
- bin/**
exclude_paths:
- config/
| engines:
csslint:
enabled: true
duplication:
enabled: true
config:
languages:
- python
- javascript
fixme:
enabled: true
pep8:
enabled: true
radon:
enabled: true
shellcheck:
enabled: true
ratings:
paths:
- app/**
- bin/**
exclude_paths:
- config/
- app/static/css/bootstrap.css
- app/static/css/normalize.css
- app/static/tether.css
- app/static/js/bootstrap.js
- app/static/js/jquery.js
- app/static/js/tether.js
| Exclude third-party code from test coverage | Exclude third-party code from test coverage
| YAML | mit | albertyw/base-flask,albertyw/base-flask,albertyw/base-flask,albertyw/base-flask |
73859f3b540623eb2dd66a82314e9b73eb47195b | _data/projects.yml | _data/projects.yml | - name: cook
description: CLI written in Node.js which allows you to easily install softwares on Ubuntu.
url: https://github.com/mlcdf/cook
img: 1f4e6
img_alt: package emoji
color: 83dcff
- name: ikarauraga
description: 2D space shooter made with Unity for Ludum Dare 35 with Clément Stéphano.
url: https://github.com/mlcdf/ikarauraga
img: 1f3ae
img_alt: video game controller emoji
color: f51843
- name: dotfiles
description: 'My config files: zsh, oh-my-zsh, z, git, node.js, python, atom, hyper, ...'
url: https://github.com/mlcdf/dotfiles
img: 1f527
img_alt: wrench emoji
color: ffeb3b
- name: generator-pym
description: Scaffold out a Python module. Includes CI, license, Makefile, pylint, ...
url: https://github.com/mlcdf/generator-pym
img: 1f40d
img_alt: snake emoji
color: e9e9e9
- name: material-utilities
description: Set of CSS classes & custom properties for Material shadows and opacities.
url: https://github.com/mlcdf/material-utilities
img: '2728'
img_alt: sparkle emoji
color: '000000'
| - name: cook
description: Easily install popular softwares that are missing from APT.
url: https://github.com/cook-pkg/cook
img: 1f4e6
img_alt: package emoji
color: 83dcff
- name: ikarauraga
description: 2D space shooter made with Unity for Ludum Dare 35 with Clément Stéphano.
url: https://github.com/mlcdf/ikarauraga
img: 1f3ae
img_alt: video game controller emoji
color: f51843
- name: dotfiles
description: 'My config files: zsh, oh-my-zsh, z, git, node.js, python, atom, hyper, ...'
url: https://github.com/mlcdf/dotfiles
img: 1f527
img_alt: wrench emoji
color: ffeb3b
- name: generator-pym
description: Scaffold out a Python module. Includes CI, license, Makefile, pylint, ...
url: https://github.com/mlcdf/generator-pym
img: 1f40d
img_alt: snake emoji
color: e9e9e9
- name: material-utilities
description: Set of CSS classes & custom properties for Material shadows and opacities.
url: https://github.com/mlcdf/material-utilities
img: '2728'
img_alt: sparkle emoji
color: '000000'
| Update cook link & description | Update cook link & description
| YAML | mit | mlcdf/website,mlcdf/website |
49f58a7354800dfcdd231b54cc33642c3a2d95ef | .github/workflows/main.yml | .github/workflows/main.yml | name: dotfiles CI
on: [push]
jobs:
build:
runs-on: ${{ matrix.os }}
strategy:
matrix:
os: [ubuntu-latest, macos-latest]
node-version: [12.x]
steps:
- uses: actions/checkout@v1
- name: Use the Node.js ${{ matrix.node-version }}
uses: actions/setup-node@v1
with:
node-version: ${{ matrix.node-version }}
- name: Update packages list and install zsh on Ubuntu
if: runner.os == 'Linux'
run: |
sudo apt update -y
sudo apt install -y zsh
- name: Update packages list and install zsh on macOS
if: runner.os == 'macOS'
run: |
brew update
brew install zsh
- name: Check syntax of initialize.zsh file
run: zsh -n initialize.zsh
env:
CI: true
- name: Check syntax of initialize script files
run: "zsh -n ./init/*.zsh"
env:
CI: true
- name: Check syntax of bin/pull-dotfiles file
run: zsh -n bin/pull-dotfiles
env:
CI: true
- name: Slack Notification
uses: homoluctus/slatify@v1.5
if: always()
with:
job_name: '*dotfiles CI*'
type: ${{ job.status }}
icon_emoji: ":octocat:"
url: ${{ secrets.SLACK_WEBHOOK }}
| name: dotfiles CI
on: [push]
jobs:
build:
runs-on: ${{ matrix.os }}
strategy:
matrix:
os: [ubuntu-latest, macos-latest]
node-version: [12.x]
steps:
- uses: actions/checkout@v1
- name: Use the Node.js ${{ matrix.node-version }}
uses: actions/setup-node@v1
with:
node-version: ${{ matrix.node-version }}
- name: Update packages list and install zsh on Ubuntu
if: runner.os == 'Linux'
run: |
sudo apt update -y
sudo apt install -y zsh
- name: Update packages list and install zsh on macOS
if: runner.os == 'macOS'
run: |
brew update
brew install zsh
- name: Check syntax of initialize.zsh file
run: zsh -n initialize.zsh
env:
CI: true
- name: Check syntax of initialize script files
run: "zsh -n ./init/*.zsh"
env:
CI: true
- name: Check syntax of bin/pull-dotfiles file
run: zsh -n bin/pull-dotfiles
env:
CI: true
- name: Slack Notification
uses: homoluctus/slatify@v1.5
if: always()
with:
job_name: '*${{ github.workflow }}*'
type: ${{ job.status }}
icon_emoji: ":octocat:"
url: ${{ secrets.SLACK_WEBHOOK }}
| Use github.workflow value in slack notification | Use github.workflow value in slack notification
| YAML | mit | Hiroto-K/dotfiles,Hiroto-K/dotfiles |
930664471ca83855726957d5883a779c40db471d | .github/workflows/main.yml | .github/workflows/main.yml | # This is a basic workflow to help you get started with Actions
name: CI
on:
push:
branches: [ master ]
pull_request:
branches: [ master ]
workflow_dispatch:
jobs:
test:
runs-on: ubuntu-18.04
steps:
- uses: actions/checkout@v3
- name: Setup Node.js environment
uses: actions/setup-node@v3.5.0
with:
node-version: 16.5.0
- name: Reuse npm cache folder
uses: actions/cache@v3
env:
cache-name: cache-node-modules
with:
path: |
~/.npm
./node_modules
# invalidate cache when any package-lock.json changes
key: ${{ runner.os }}-npm-test-x1-${{ hashFiles('**/package.json') }}
restore-keys: |
${{ runner.os }}-npm-test-x1-
- name: install npm dependencies
# TODO remove legace peer deps once everything is up to date
run: npm install --legacy-peer-deps
- name: build
run: npm run build
- name: test:node
run: npm run test:node
- name: test:browser
uses: GabrielBB/xvfb-action@v1
with:
run: npm run test:browser
- name: lint
run: npm run lint
| # This is a basic workflow to help you get started with Actions
name: CI
on:
push:
branches: [ master ]
pull_request:
branches: [ master ]
workflow_dispatch:
jobs:
test:
runs-on: ubuntu-18.04
steps:
- uses: actions/checkout@v3
- name: Setup Node.js environment
uses: actions/setup-node@v3.5.1
with:
node-version: 16.5.0
- name: Reuse npm cache folder
uses: actions/cache@v3
env:
cache-name: cache-node-modules
with:
path: |
~/.npm
./node_modules
# invalidate cache when any package-lock.json changes
key: ${{ runner.os }}-npm-test-x1-${{ hashFiles('**/package.json') }}
restore-keys: |
${{ runner.os }}-npm-test-x1-
- name: install npm dependencies
# TODO remove legace peer deps once everything is up to date
run: npm install --legacy-peer-deps
- name: build
run: npm run build
- name: test:node
run: npm run test:node
- name: test:browser
uses: GabrielBB/xvfb-action@v1
with:
run: npm run test:browser
- name: lint
run: npm run lint
| Update actions/setup-node action to v3.5.1 | Update actions/setup-node action to v3.5.1 | YAML | apache-2.0 | pubkey/async-test-util |
9c67323b4bc94da3f7a7ce75fa56296f6d10a2c9 | .github/workflows/main.yml | .github/workflows/main.yml | name: E-Mail Release Notification
on:
release:
types: [released]
jobs:
notify:
runs-on: ubuntu-latest
steps:
- name: Notify about a new release
uses: daysofwonder/sendgrid-notification-action@v1.4
env:
SENDGRID_API_TOKEN: ${{ secrets.SENDGRID_API_TOKEN }}
RECIPIENTS_URL: ${{ secrets.RECIPIENTS_URL }}
SENDER_EMAIL: ${{ secrets.SENDER_EMAIL }}
DISTRIBUTION_LISTS: ${{ secrets.DISTRIBUTION_LISTS }}
| name: E-Mail Release Notification
on:
release:
types: [released]
jobs:
notify:
runs-on: ubuntu-latest
steps:
- name: Notify about a new release
uses: daysofwonder/sendgrid-notification-action@v1.5
env:
SENDGRID_API_TOKEN: ${{ secrets.SENDGRID_API_TOKEN }}
RECIPIENTS_URL: ${{ secrets.RECIPIENTS_URL }}
SENDER_EMAIL: ${{ secrets.SENDER_EMAIL }}
DISTRIBUTION_LISTS: ${{ secrets.DISTRIBUTION_LISTS }}
| Update sendgrid notif uses version in action notify | Update sendgrid notif uses version in action notify
| YAML | mit | daysofwonder/asmodeenet_sso_js,daysofwonder/asmodeenet_sso_js,daysofwonder/asmodeenet_sso_js |
780b1a9326e55d4524c2052aa2bb4564fd251bf4 | build/docker-compose.yml | build/docker-compose.yml | version: '2'
services:
# the web container
web-server:
image: venveo/baseimage:7.0-apache
volumes:
- "./php:/var/www/public"
ports:
- "80:80"
- "443:443"
links:
- mysql-server
- redis-server
- memcached-server
depends_on:
- mysql-server
- redis-server
- memcached-server
environment:
- APP_ENV=local
- APP_CPTRIGGER=manage
# mysql container
mysql-server:
image: venveo/baseimage:5.7-mysql
volumes:
- "./mysql/data:/var/lib/mysql"
ports:
- "3306:3306"
environment:
- MYSQL_ROOT_PASSWORD=TestingRootPassword
- MYSQL_DATABASE=testing_db
- MYSQL_USER=testing_user
- MYSQL_PASSWORD=testing_password
- MYSQL_ALLOW_EMPTY_PASSWORD=no
# redis container
redis-server:
image: venveo/baseimage:3.2-redis
ports:
- "6379:6379"
# memcached container
memcached-server:
image: venveo/baseimage:1.4-memcached
ports:
- "11211:11211"
| version: '2'
services:
# the web container
web-server:
image: venveo/baseimage:1.10-nginx
# image: venveo/baseimage:7.0-apache
volumes:
- "./php:/var/www/public"
# - "./php:/usr/share/nginx/html"
ports:
- "80:80"
- "443:443"
links:
- mysql-server
- redis-server
- memcached-server
- php-fpm
depends_on:
- mysql-server
- redis-server
- memcached-server
- php-fpm
environment:
- APP_ENV=local
- APP_CPTRIGGER=manage
# mysql container
mysql-server:
image: venveo/baseimage:5.7-mysql
volumes:
- "./mysql/data:/var/lib/mysql"
ports:
- "3306:3306"
environment:
- MYSQL_ROOT_PASSWORD=TestingRootPassword
- MYSQL_DATABASE=testing_db
- MYSQL_USER=testing_user
- MYSQL_PASSWORD=testing_password
- MYSQL_ALLOW_EMPTY_PASSWORD=no
# redis container
redis-server:
image: venveo/baseimage:3.2-redis
ports:
- "6379:6379"
# memcached container
memcached-server:
image: venveo/baseimage:1.4-memcached
ports:
- "11211:11211"
# php-fpm container
php-fpm:
image: venveo/baseimage:7.0-php-fpm
ports:
- "9000:9000"
| Update compose to test nginx and php-fpm | Update compose to test nginx and php-fpm
| YAML | mit | venveo/docker-baseimage,venveo/docker-baseimage |
a5fcc710a799e63e7c2163f7d9b12c3e1c93294d | appveyor.yml | appveyor.yml | version: 1.0.{build}
configuration:
- Debug
- Release
clone_folder: C:\Fruit
environment:
BOOST_DIR: C:\Libraries\boost
matrix:
- image: Visual Studio 2017
CMAKE_PATH: '"C:\Program Files (x86)\CMake\bin"'
CMAKE_GENERATOR: '"MinGW Makefiles"'
- image: Visual Studio 2015
CMAKE_PATH: '"C:\Program Files (x86)\CMake\bin"'
CMAKE_GENERATOR: '"Visual Studio 14 2015 Win64"'
VCVARSALL_SCRIPT: '"C:\Program Files (x86)\Microsoft Visual Studio 14.0\VC\vcvarsall.bat"'
- image: Visual Studio 2017
CMAKE_PATH: '"C:\Program Files (x86)\Microsoft Visual Studio\2017\Community\Common7\IDE\CommonExtensions\Microsoft\CMake\CMake\bin"'
CMAKE_GENERATOR: '"Visual Studio 15 2017 Win64"'
VCVARSALL_SCRIPT: '"C:\Program Files (x86)\Microsoft Visual Studio\2017\Community\VC\Auxiliary\Build\vcvarsall.bat"'
install:
- cmd: setx PATH %PATH:C:\Program Files\Git\usr\bin;=%
build_script:
- cmd: C:\Fruit\extras\scripts\postsubmit.bat
| version: 1.0.{build}
configuration:
- Debug
- Release
clone_folder: C:\Fruit
environment:
BOOST_DIR: C:\Libraries\boost
matrix:
- image: Visual Studio 2017
CMAKE_PATH: '"C:\Program Files (x86)\CMake\bin"'
CMAKE_GENERATOR: '"MinGW Makefiles"'
- image: Visual Studio 2015
CMAKE_PATH: '"C:\Program Files (x86)\CMake\bin"'
CMAKE_GENERATOR: '"Visual Studio 14 2015 Win64"'
VCVARSALL_SCRIPT: '"C:\Program Files (x86)\Microsoft Visual Studio 14.0\VC\vcvarsall.bat"'
- image: Visual Studio 2017
CMAKE_PATH: '"C:\Program Files (x86)\Microsoft Visual Studio\2017\Community\Common7\IDE\CommonExtensions\Microsoft\CMake\CMake\bin"'
CMAKE_GENERATOR: '"Visual Studio 15 2017 Win64"'
VCVARSALL_SCRIPT: '"C:\Program Files (x86)\Microsoft Visual Studio\2017\Community\VC\Auxiliary\Build\vcvarsall.bat"'
install:
- cmd: setx PATH "%PATH:C:\Program Files\Git\usr\bin;=%"
build_script:
- cmd: C:\Fruit\extras\scripts\postsubmit.bat
| Fix quoting in a setx command used when testing Fruit on Windows in AppVeyor. | Fix quoting in a setx command used when testing Fruit on Windows in AppVeyor.
| YAML | apache-2.0 | google/fruit,google/fruit,google/fruit |
25fbf5304019458c30835ae8ab933610a055c3f0 | appveyor.yml | appveyor.yml | version: 0.1.{build}
configuration: Release
platform: Any CPU
build_script:
- cmd: .\build.bat Package
test: off
artifacts:
- path: '**\CertiPay.*.nupkg' | version: 0.1.{build}
configuration: Release
platform: Any CPU
build_script:
- cmd: .\build.bat Package
test: off
artifacts:
- path: '**\CertiPay.Services.PDF.*.nupkg' | Tweak publish artifcats to only catch the service | Tweak publish artifcats to only catch the service
| YAML | mit | kylebjones/CertiPay.Services.PDF |
58155cefe86ccc7b2082d40f669fec2a7dc1b9e5 | appveyor.yml | appveyor.yml | version: 2.1.18.{build}
pull_requests:
do_not_increment_build_number: true
branches:
only:
- master
image: Visual Studio 2017
configuration: Release
init:
- -cmd: git config --global core.autocrlf true
dotnet_csproj:
patch: true
file: '**\*.csproj'
version: '{version}'
package_version: '{version}'
assembly_version: '{version}'
file_version: '{version}'
informational_version: '{version}'
before_build:
- cmd: choco install codecov
- cmd: dotnet restore
build:
publish_nuget: true
publish_nuget_symbols: true
verbosity: minimal
test_script:
- cmd: dotnet test /p:CollectCoverage=true /p:CoverletOutputFormat=opencover ./test/ArgSentry.Test/ArgSentry.Test.csproj --logger AppVeyor
- cmd: codecov -f "./test/ArgSentry.Test/coverage.opencover.xml"
artifacts:
- path: '**\*.nupkg'
deploy:
- provider: NuGet
api_key:
secure: 92tF47KjBR7PvZKHSvOyde7y5i17Ciwai1+X4gVtNKqGEK1ZrVl8EHQkatu/BjcH
skip_symbols: false
artifact: /.*\.nupkg/
on:
branch: master | version: 2.1.{build}
pull_requests:
do_not_increment_build_number: true
branches:
only:
- master
image: Visual Studio 2017
configuration: Release
init:
- cmd: git config --global core.autocrlf true
install:
- cmd: choco install codecov
dotnet_csproj:
patch: true
file: '**\*.csproj'
version: '{version}'
package_version: '{version}'
assembly_version: '{version}'
file_version: '{version}'
informational_version: '{version}'
before_build:
- cmd: dotnet restore
build_script:
- cmd: dotnet build ./src/ArgSentry/ArgSentry.csproj --configuration Release --no-restore
test_script:
- cmd: dotnet test /p:CollectCoverage=true /p:CoverletOutputFormat=opencover ./test/ArgSentry.Test/ArgSentry.Test.csproj --logger AppVeyor
- cmd: codecov -f "./test/ArgSentry.Test/coverage.opencover.xml"
after_test:
- cmd: dotnet pack ./src/ArgSentry/ArgSentry.csproj --configuration Release --no-build --no-restore --output ../../dist
artifacts:
- path: './dist/*.nupkg'
deploy:
- provider: NuGet
api_key:
secure: 92tF47KjBR7PvZKHSvOyde7y5i17Ciwai1+X4gVtNKqGEK1ZrVl8EHQkatu/BjcH
skip_symbols: true
artifact: /.*\.nupkg/
on:
branch: master | Build & deploy process cleanup | Build & deploy process cleanup
| YAML | mit | eric-davis/ArgSentry |
4b471f61de3d5c4876a4e5e10b6c6fb3296ed416 | tasks/main.yml | tasks/main.yml | ---
- name: "++++++++++ install and config dehydrated and all challenge hooks ++++++++++"
include_tasks: setup.yml
become: true
when:
- le_certificates is defined
- le_deploy == false
- ansible_os_family == "Debian"
tags:
- setup
- name: "++++++++++ sign/renew non-existant/changed/expiring certificates ++++++++++"
include_tasks: certificate.yml
become: true
when:
- le_certificates is defined
- le_deploy == false
- ansible_os_family == "Debian"
with_items:
- "{{ le_certificates }}"
loop_control:
loop_var: le_certificate_loop
tags:
- certificate
- name: "++++++++++ deploy signed/renewed certificates to other webservers +++++++++"
include_tasks: deploy.yml
when: (le_deploy == true or le_deploy == "force") and le_deploy_certificates is defined
with_items:
- "{{ le_deploy_certificates }}"
loop_control:
loop_var: le_deploy_certificate_loop
tags:
- deploy | ---
- name: "++++++++++ install and config dehydrated and all challenge hooks ++++++++++"
include_tasks: setup.yml
when:
- le_certificates is defined
- le_deploy == false
- ansible_os_family == "Debian"
tags:
- setup
- name: "++++++++++ sign/renew non-existant/changed/expiring certificates ++++++++++"
include_tasks: certificate.yml
when:
- le_certificates is defined
- le_deploy == false
- ansible_os_family == "Debian"
with_items:
- "{{ le_certificates }}"
loop_control:
loop_var: le_certificate_loop
tags:
- certificate
- name: "++++++++++ deploy signed/renewed certificates to other webservers +++++++++"
include_tasks: deploy.yml
when: (le_deploy == true or le_deploy == "force") and le_deploy_certificates is defined
with_items:
- "{{ le_deploy_certificates }}"
loop_control:
loop_var: le_deploy_certificate_loop
tags:
- deploy | Remove become statement in dynamic tasks include | Remove become statement in dynamic tasks include
| YAML | mit | rembik/ansible-role-letsencrypt,rembik/ansible-role-letsencrypt |
9cf2ef41a908a1a126b5418806a2ba4de31ebb86 | .conda/meta.yaml | .conda/meta.yaml | package:
name: csdms-dakota
version: {{"0.2" + environ.get('BUILD_STR', '')}}
source:
path: ..
requirements:
build:
- python
- numpy
- pyyaml
- basic-modeling-interface
run:
- python
- numpy
- pyyaml
- basic-modeling-interface
- nose
| package:
name: csdms-dakota
version: {{"0.2" + environ.get('BUILD_STR', '')}}
source:
path: ..
requirements:
build:
- python
- numpy
- pyyaml
- basic-modeling-interface
- nose
run:
- python
- numpy
- pyyaml
- basic-modeling-interface
- nose
| Include nose in build requirements | Include nose in build requirements
| YAML | mit | csdms/dakota,csdms/dakota |
4ad385f3bb4504cc156485ed941a67aab4133423 | deploy/kubernetes/elasticsearch/snapshot-cronjob.yaml | deploy/kubernetes/elasticsearch/snapshot-cronjob.yaml | apiVersion: batch/v1beta1
kind: CronJob
metadata:
name: es-monthly-snapshotter
labels:
name: es-monthly-snapshotter
spec:
schedule: "@monthly"
jobTemplate:
spec:
template:
spec:
containers:
- name: es-monthly-snapshotter
image: {{ DOCKER_IMAGE_PREFIX }}/curl{{ DOCKER_IMAGE_TAG }}
imagePullPolicy: {{ IMAGE_PULL_POLICY }}
env:
- name: ES_PASSWORD
valueFrom:
secretKeyRef:
name: kibana-secrets
key: elasticsearch.password
- name: ELASTICSEARCH_SERVICE_HOSTNAME
value: "{{ ELASTICSEARCH_SERVICE_HOSTNAME }}"
args:
- /bin/sh
- -c
- 'curl -s -i -u "kibana:${ES_PASSWORD}" -XPUT "http://${ELASTICSEARCH_SERVICE_HOSTNAME}:9200/_snapshot/snapshot_storage/%3Csnapshot-%7Bnow%2Fd%7D%3E" | tee /dev/stderr | grep "200 OK"'
restartPolicy: OnFailure
| apiVersion: batch/v1beta1
kind: CronJob
metadata:
name: es-monthly-snapshotter
labels:
name: es-monthly-snapshotter
spec:
schedule: "@monthly"
concurrencyPolicy: Forbid
jobTemplate:
spec:
template:
metadata:
labels:
app: es-monthly-snapshotter
spec:
containers:
- name: es-monthly-snapshotter
image: {{ DOCKER_IMAGE_PREFIX }}/curl{{ DOCKER_IMAGE_TAG }}
imagePullPolicy: {{ IMAGE_PULL_POLICY }}
env:
- name: ES_PASSWORD
valueFrom:
secretKeyRef:
name: kibana-secrets
key: elasticsearch.password
- name: ELASTICSEARCH_SERVICE_HOSTNAME
value: "{{ ELASTICSEARCH_SERVICE_HOSTNAME }}"
args:
- /bin/sh
- -c
- 'curl -s -i -u "kibana:${ES_PASSWORD}" -XPUT "http://${ELASTICSEARCH_SERVICE_HOSTNAME}:9200/_snapshot/snapshot_storage/%3Csnapshot-%7Bnow%2Fd%7D%3E" | tee /dev/stderr | grep "200 OK"'
restartPolicy: OnFailure
| Add tagging for easier log alerting, forbid concurrent pods | Add tagging for easier log alerting, forbid concurrent pods
| YAML | agpl-3.0 | ssadedin/seqr,ssadedin/seqr,ssadedin/seqr,ssadedin/seqr,ssadedin/seqr |
5ca82302472740c4bc577889e5b2fda96bd93d2d | .forestry/settings.yml | .forestry/settings.yml | ---
new_page_extension: md
auto_deploy: false
admin_path: "/admin"
webhook_url:
sections:
- type: jekyll-pages
label: Pages
create: all
- type: jekyll-posts
label: Posts
create: all
upload_dir: uploads
public_path: "/uploads"
front_matter_path: ''
use_front_matter_path:
file_template: ":filename:"
build:
preview_command: bundle exec jekyll build --drafts --unpublished --future -d _site
publish_command: bundle exec jekyll build -d _site
preview_env:
- JEKYLL_ENV=staging
publish_env:
- JEKYLL_ENV=production
preview_output_directory: _site
output_directory: _site
| ---
new_page_extension: md
auto_deploy: true
admin_path: "/admin"
webhook_url:
sections:
- type: jekyll-pages
label: Pages
create: all
- type: jekyll-posts
label: Posts
create: all
upload_dir: uploads
public_path: "/uploads"
front_matter_path: ''
use_front_matter_path:
file_template: ":filename:"
build:
preview_command: bundle exec jekyll build --drafts --unpublished --future -d _site
publish_command: bundle exec jekyll build -d _site
preview_env:
- JEKYLL_ENV=staging
publish_env:
- JEKYLL_ENV=production
preview_output_directory: _site
output_directory: _site
| Update from Forestry.io - Updated Forestry configuration | Update from Forestry.io - Updated Forestry configuration
| YAML | mit | digitalfort/goulburnspeedway |
bc91abff849d5085025bd202fd77911b6e5e358e | metadata/me.lucky.duress.yml | metadata/me.lucky.duress.yml | Categories:
- Security
License: GPL-3.0-only
AuthorName: lucky
AuthorEmail: 44uaanjm0@relay.firefox.com
AuthorWebSite: https://github.com/x13a
SourceCode: https://github.com/x13a/Duress
IssueTracker: https://github.com/x13a/Duress/issues
AutoName: Duress
RepoType: git
Repo: https://github.com/x13a/Duress.git
Builds:
- versionName: 1.0.1
versionCode: 2
commit: cd358b81b57ac772ccf1b723e7a9a7625c17545e
subdir: app
sudo:
- apt-get update || apt-get update
- apt-get install -y openjdk-11-jdk-headless
- update-alternatives --auto java
gradle:
- yes
AutoUpdateMode: Version v%v
UpdateCheckMode: Tags
CurrentVersion: 1.0.1
CurrentVersionCode: 2
| Categories:
- Security
License: GPL-3.0-only
AuthorName: lucky
AuthorEmail: 44uaanjm0@relay.firefox.com
AuthorWebSite: https://github.com/x13a
SourceCode: https://github.com/x13a/Duress
IssueTracker: https://github.com/x13a/Duress/issues
AutoName: Duress
RepoType: git
Repo: https://github.com/x13a/Duress.git
Builds:
- versionName: 1.0.1
versionCode: 2
commit: cd358b81b57ac772ccf1b723e7a9a7625c17545e
subdir: app
sudo:
- apt-get update || apt-get update
- apt-get install -y openjdk-11-jdk-headless
- update-alternatives --auto java
gradle:
- yes
- versionName: 1.0.2
versionCode: 3
commit: 326e3c7a073a4ba48ef69dfaa327556ffd2343c8
subdir: app
sudo:
- apt-get update || apt-get update
- apt-get install -y openjdk-11-jdk-headless
- update-alternatives --auto java
gradle:
- yes
AutoUpdateMode: Version v%v
UpdateCheckMode: Tags
CurrentVersion: 1.0.2
CurrentVersionCode: 3
| Update Duress to 1.0.2 (3) | Update Duress to 1.0.2 (3)
| YAML | agpl-3.0 | f-droid/fdroiddata,f-droid/fdroiddata |
d6e138cd384aad02f37851c8d748fc81262e7af4 | devtools/conda-recipe/meta.yaml | devtools/conda-recipe/meta.yaml | package:
name: openmoltools-dev
version: !!str dev
source:
path: ../../
build:
preserve_egg_dir: True
number: 0
requirements:
build:
- python
- setuptools
- pandas
- six
- mdtraj
- numpy
- numpydoc
- scipy
- openmm
- ambermini
- pytables
- parmed
# - rdkit # rdkit is an optional dependency, may want to comment this out for the release version.
run:
- python
- setuptools
- pandas
- six
- mdtraj
- numpy
- numpydoc
- scipy
- openmm
- ambermini
- pytables
- parmed
# - rdkit # rdkit is an optional dependency, may want to comment this out for the release version.
test:
requires:
- nose
imports:
- openmoltools
#commands:
# - nosetests openmoltools -v -a '!slow'
about:
home: https://github.com/choderalab/openmoltools
license: GNU Lesser General Public License v2 or later (LGPLv2+)
| package:
name: openmoltools-dev
version: !!str dev
source:
path: ../../
build:
preserve_egg_dir: True
number: 0
requirements:
build:
- python
- setuptools
- pandas
- six
- mdtraj
- numpy
- numpydoc
- scipy
- openmm-dev
- ambermini
- pytables
- parmed
# - rdkit # rdkit is an optional dependency, may want to comment this out for the release version.
run:
- python
- setuptools
- pandas
- six
- mdtraj
- numpy
- numpydoc
- scipy
- openmm-dev
- ambermini
- pytables
- parmed
# - rdkit # rdkit is an optional dependency, may want to comment this out for the release version.
test:
requires:
- nose
imports:
- openmoltools
#commands:
# - nosetests openmoltools -v -a '!slow'
about:
home: https://github.com/choderalab/openmoltools
license: GNU Lesser General Public License v2 or later (LGPLv2+)
| Change openmm test dependency to openmm-dev | Change openmm test dependency to openmm-dev
| YAML | mit | choderalab/openmoltools,choderalab/openmoltools |
448fe2dd64b52e41e647320ee002c5a6ff84e23b | devtools/conda-recipe/meta.yaml | devtools/conda-recipe/meta.yaml | package:
name: perses-dev
version: !!str 0.0.0
source:
path: ../../
build:
preserve_egg_dir: True
number: 0
requirements:
build:
- python
- setuptools
- openmm-dev
- parmed-dev
- numpy
- scipy
- openmoltools-dev
- alchemy
- numexpr
- openmmtools
- pymbar
- autograd
run:
- python
- openmm-dev
- parmed-dev
- numpy
- scipy
- openmoltools-dev
- alchemy
- numexpr
- openmmtools
- pymbar
- autograd
test:
requires:
- nose
- openmmtools
- pymbar
imports:
- perses
about:
home: https://github.com/choderalab/perses
license: GNU Lesser General Public License v2 or later (LGPLv2+)
| package:
name: perses-dev
version: !!str 0.0.0
source:
path: ../../
build:
preserve_egg_dir: True
number: 0
requirements:
build:
- python
- setuptools
- numpy
- scipy
- numexpr
- autograd
- pymbar-dev
- openmm-dev
- parmed-dev
- openmoltools-dev
- alchemy-dev
- openmmtools-dev
run:
- python
- setuptools
- numpy
- scipy
- numexpr
- autograd
- pymbar-dev
- openmm-dev
- parmed-dev
- openmoltools-dev
- alchemy-dev
- openmmtools-dev
test:
requires:
- nose
imports:
- perses
about:
home: https://github.com/choderalab/perses
license: GNU Lesser General Public License v2 or later (LGPLv2+)
| Fix conda recipe dependencies to ensure we get openmm-dev and parmed-dev | Fix conda recipe dependencies to ensure we get openmm-dev and parmed-dev
| YAML | mit | choderalab/perses,choderalab/perses |
9c6e2ec6a2fb4a457b3cd3e7ca08be468110fdc4 | .github/dependabot.yml | .github/dependabot.yml | # https://help.github.com/github/administering-a-repository/configuration-options-for-dependency-updates
version: 2
updates:
- package-ecosystem: "maven"
directory: "/"
schedule:
interval: "weekly"
- package-ecosystem: "github-actions"
directory: "/"
schedule:
interval: "weekly"
| # https://help.github.com/github/administering-a-repository/configuration-options-for-dependency-updates
version: 2
updates:
- package-ecosystem: "maven"
open-pull-requests-limit: 10
directory: "/"
schedule:
interval: "weekly"
- package-ecosystem: "github-actions"
directory: "/"
schedule:
interval: "weekly"
| Increase open pull requests limit | Increase open pull requests limit
| YAML | mit | jan-molak/jenkins-build-monitor-plugin,jan-molak/jenkins-build-monitor-plugin,jan-molak/jenkins-build-monitor-plugin,jan-molak/jenkins-build-monitor-plugin |
b224c285e684ced74f7ba332f97bf6c9aa3f306b | stacks/resource-groups.yml | stacks/resource-groups.yml | # stacks/resource-groups.yml
AWSTemplateFormatVersion: "2010-09-09"
Description: >-
Creates resource groups for resources in the stack, based on various sets of
tag and resource type filters.
Parameters:
EnvironmentType: { Type: String }
RootStackName: { Type: String }
RootStackId: { Type: String }
Resources:
StackResourceGroup:
Type: AWS::ResourceGroups::Group
Properties:
Description: >-
Inlcudes all supported resources that belong to the root stack
Name: !Sub ${RootStackName}
ResourceQuery:
Query:
ResourceTypeFilters:
- AWS::AllSupported
TagFilters:
- Key: prx:cloudformation:root-stack-name
Values:
- !Ref RootStackName
Type: TAG_FILTERS_1_0
Tags:
- { Key: prx:meta:tagging-version, Value: "2021-04-07" }
- { Key: prx:cloudformation:stack-name, Value: !Ref AWS::StackName }
- { Key: prx:cloudformation:stack-id, Value: !Ref AWS::StackId }
# - { Key: prx:cloudformation:root-stack-name, Value: !Ref RootStackName }
# - { Key: prx:cloudformation:root-stack-id, Value: !Ref RootStackId }
- { Key: prx:ops:environment, Value: !Ref EnvironmentType }
- { Key: prx:dev:application, Value: Infrastructure }
Outputs:
StackResourceGroupName:
Value: !Ref StackResourceGroup
| # stacks/resource-groups.yml
AWSTemplateFormatVersion: "2010-09-09"
Description: >-
Creates resource groups for resources in the stack, based on various sets of
tag and resource type filters.
Parameters:
EnvironmentType: { Type: String }
RootStackName: { Type: String }
RootStackId: { Type: String }
Resources:
StackResourceGroup:
Type: AWS::ResourceGroups::Group
Properties:
Description: >-
Inlcudes all supported resources that belong to the root stack
Name: !Sub ${RootStackName}
ResourceQuery:
Query:
ResourceTypeFilters:
- AWS::AllSupported
TagFilters:
- Key: prx:cloudformation:root-stack-name
Values:
- !Ref RootStackName
Type: TAG_FILTERS_1_0
# Tags:
# - { Key: prx:meta:tagging-version, Value: "2021-04-07" }
# - { Key: prx:cloudformation:stack-name, Value: !Ref AWS::StackName }
# - { Key: prx:cloudformation:stack-id, Value: !Ref AWS::StackId }
# # - { Key: prx:cloudformation:root-stack-name, Value: !Ref RootStackName }
# # - { Key: prx:cloudformation:root-stack-id, Value: !Ref RootStackId }
# - { Key: prx:ops:environment, Value: !Ref EnvironmentType }
# - { Key: prx:dev:application, Value: Infrastructure }
Outputs:
StackResourceGroupName:
Value: !Ref StackResourceGroup
| Remove all resource group tags? | Remove all resource group tags?
| YAML | mit | PRX/Infrastructure,PRX/Infrastructure,PRX/Infrastructure,PRX/Infrastructure,PRX/Infrastructure |
d3f7b177e0e90eed6e153d582544660cf743e828 | .github/dependabot.yml | .github/dependabot.yml | # To get started with Dependabot version updates, you'll need to specify which
# package ecosystems to update and where the package manifests are located.
# Please see the documentation for all configuration options:
# https://help.github.com/github/administering-a-repository/configuration-options-for-dependency-updates
version: 2
updates:
- package-ecosystem: "docker"
directory: "/"
schedule:
interval: "daily"
open-pull-requests-limit: 10
- package-ecosystem: "npm"
directory: "/"
schedule:
interval: "daily"
open-pull-requests-limit: 10
- package-ecosystem: "pip"
directory: "/"
schedule:
interval: "daily"
open-pull-requests-limit: 10
- package-ecosystem: "github-actions"
directory: "/"
schedule:
interval: "daily"
open-pull-requests-limit: 10
| # To get started with Dependabot version updates, you'll need to specify which
# package ecosystems to update and where the package manifests are located.
# Please see the documentation for all configuration options:
# https://help.github.com/github/administering-a-repository/configuration-options-for-dependency-updates
version: 2
updates:
- package-ecosystem: "docker"
directory: "/"
schedule:
interval: "daily"
open-pull-requests-limit: 10
- package-ecosystem: "npm"
directory: "/csunplugged/"
schedule:
interval: "daily"
open-pull-requests-limit: 10
- package-ecosystem: "pip"
directory: "/"
schedule:
interval: "daily"
open-pull-requests-limit: 10
- package-ecosystem: "github-actions"
directory: "/"
schedule:
interval: "daily"
open-pull-requests-limit: 10
| Update location of JavaScript dependency files | Update location of JavaScript dependency files
| YAML | mit | uccser/cs-unplugged,uccser/cs-unplugged,uccser/cs-unplugged,uccser/cs-unplugged |
a645b82cddadce0f0a702a01774b12d0d477748b | config/outgoing_mail.yml | config/outgoing_mail.yml | development:
address: "smtp.example.com"
port: "25"
user_name: "user"
password: "password"
authentication: "plain" # plain, login, or cram_md5
domain: "example.com"
outgoing_address: "canvas@example.com"
default_name: "Instructure Canvas"
# or to use sendmail --
#
# development:
# delivery_method: "sendmail"
test:
delivery_method: "test"
production:
address: "localhost"
port: "25"
# user_name: "user"
# password: "password"
# authentication: "plain" # plain, login, or cram_md5
domain: "canvas.localdomain"
outgoing_address: "canvas@canvas.localdomain"
default_name: "Instructure Canvas"
# If receiving mail from multiple inboxes (see incoming_mail.yml.example),
# you'll want to include those addresses in a reply_to_addresses array so
# Canvas will select the Reply-To field of outgoing messages from all of the
# incoming mailboxes.
multiple_inboxes:
address: "smtp.example.com"
port: "25"
user_name: "user"
password: "password"
authentication: "plain" # plain, login, or cram_md5
domain: "example.com"
outgoing_address: "canvas@example.com"
default_name: "Instructure Canvas"
reply_to_addresses:
- "canvas1@example.com"
- "canvas2@example.com"
- "canvas3@example.com"
- "canvas4@example.com"
| development:
delivery_method: "sendmail"
# or to use sendmail --
#
# development:
# delivery_method: "sendmail"
test:
delivery_method: "test"
production:
address: "localhost"
port: "25"
# user_name: "user"
# password: "password"
# authentication: "plain" # plain, login, or cram_md5
domain: "canvas.localdomain"
outgoing_address: "canvas@canvas.localdomain"
default_name: "Instructure Canvas"
# If receiving mail from multiple inboxes (see incoming_mail.yml.example),
# you'll want to include those addresses in a reply_to_addresses array so
# Canvas will select the Reply-To field of outgoing messages from all of the
# incoming mailboxes.
multiple_inboxes:
address: "smtp.example.com"
port: "25"
user_name: "user"
password: "password"
authentication: "plain" # plain, login, or cram_md5
domain: "example.com"
outgoing_address: "canvas@example.com"
default_name: "Instructure Canvas"
reply_to_addresses:
- "canvas1@example.com"
- "canvas2@example.com"
- "canvas3@example.com"
- "canvas4@example.com"
| Use sendmail in dev mode | Use sendmail in dev mode
| YAML | agpl-3.0 | HotChalk/canvas-lms,HotChalk/canvas-lms,HotChalk/canvas-lms,HotChalk/canvas-lms |
301c82a49bbdd9c21de2c8b03bc75db41bdef971 | metadata/t20kdc.offlinepuzzlesolver.yml | metadata/t20kdc.offlinepuzzlesolver.yml | Categories:
- Games
License: GPL-3.0-or-later
AuthorName: 20kdc
SourceCode: https://gitlab.com/20kdc/offline-puzzle-solver
IssueTracker: https://gitlab.com/20kdc/offline-puzzle-solver/-/issues
AutoName: Offline Puzzle Solver
RepoType: git
Repo: https://gitlab.com/20kdc/offline-puzzle-solver.git
Builds:
- versionName: '1.0'
versionCode: 1
commit: v1
subdir: app
gradle:
- yes
- versionName: '2.0'
versionCode: 2
commit: 49d9c3f4f28071b8ee406f487e6f00ee2b8eb0cf
subdir: app
gradle:
- yes
AutoUpdateMode: Version v%c
UpdateCheckMode: Tags
CurrentVersion: '2.0'
CurrentVersionCode: 2
| Categories:
- Games
License: GPL-3.0-or-later
AuthorName: 20kdc
SourceCode: https://gitlab.com/20kdc/offline-puzzle-solver
IssueTracker: https://gitlab.com/20kdc/offline-puzzle-solver/-/issues
AutoName: Offline Puzzle Solver
RepoType: git
Repo: https://gitlab.com/20kdc/offline-puzzle-solver.git
Builds:
- versionName: '1.0'
versionCode: 1
commit: v1
subdir: app
gradle:
- yes
- versionName: '2.0'
versionCode: 2
commit: 49d9c3f4f28071b8ee406f487e6f00ee2b8eb0cf
subdir: app
gradle:
- yes
- versionName: '4.0'
versionCode: 4
commit: 06320725744f782eac6bfc2fcd504a01e5a801b7
subdir: app
gradle:
- yes
AutoUpdateMode: Version v%c
UpdateCheckMode: Tags
CurrentVersion: '4.0'
CurrentVersionCode: 4
| Update Offline Puzzle Solver to 4.0 (4) | Update Offline Puzzle Solver to 4.0 (4)
| YAML | agpl-3.0 | f-droid/fdroiddata,f-droid/fdroiddata |
e16d451f5b7dd56f9047268fe0a2be23f608dd5c | config/jobs/image-pushing/k8s-staging-dns.yaml | config/jobs/image-pushing/k8s-staging-dns.yaml | postsubmits:
# this is the github repo we'll build from; this block needs to be repeated for each repo.
kubernetes/dns:
- name: dns-push-images
cluster: k8s-infra-prow-build-trusted
annotations:
# this is the name of some testgrid dashboard to report to.
testgrid-dashboards: sig-network-dns
decorate: true
branches:
- ^master$
spec:
serviceAccountName: gcb-builder
containers:
- image: gcr.io/k8s-testimages/image-builder:v20200901-ab141a0
command:
- /run.sh
args:
# this is the project GCB will run in, which is the same as the GCR images are pushed to.
- --project=k8s-staging-dns
- --scratch-bucket=gs://k8s-staging-dns-gcb
- --env-passthrough=PULL_BASE_REF
- .
| postsubmits:
# this is the github repo we'll build from; this block needs to be repeated for each repo.
kubernetes/dns:
- name: dns-push-images
cluster: k8s-infra-prow-build-trusted
annotations:
# this is the name of some testgrid dashboard to report to.
testgrid-dashboards: sig-network-dns
decorate: true
# match on regex '^\d+\.\d+\.\d+$' to trigger job on new tags
branches:
- ^master$
- ^\d+\.\d+\.\d+$
spec:
serviceAccountName: gcb-builder
containers:
- image: gcr.io/k8s-testimages/image-builder:v20200901-ab141a0
command:
- /run.sh
args:
# this is the project GCB will run in, which is the same as the GCR images are pushed to.
- --project=k8s-staging-dns
- --scratch-bucket=gs://k8s-staging-dns-gcb
- --env-passthrough=PULL_BASE_REF
- .
| Update DNS image push yaml to trigger on new tags | Update DNS image push yaml to trigger on new tags | YAML | apache-2.0 | cblecker/test-infra,cjwagner/test-infra,fejta/test-infra,brahmaroutu/test-infra,michelle192837/test-infra,BenTheElder/test-infra,kubernetes/test-infra,cblecker/test-infra,pwittrock/test-infra,jessfraz/test-infra,pwittrock/test-infra,kubernetes/test-infra,BenTheElder/test-infra,dims/test-infra,cjwagner/test-infra,dims/test-infra,BenTheElder/test-infra,michelle192837/test-infra,BenTheElder/test-infra,cjwagner/test-infra,cblecker/test-infra,michelle192837/test-infra,cjwagner/test-infra,monopole/test-infra,brahmaroutu/test-infra,kubernetes/test-infra,BenTheElder/test-infra,BenTheElder/test-infra,cblecker/test-infra,dims/test-infra,dims/test-infra,monopole/test-infra,fejta/test-infra,michelle192837/test-infra,monopole/test-infra,fejta/test-infra,jessfraz/test-infra,jessfraz/test-infra,fejta/test-infra,cblecker/test-infra,pwittrock/test-infra,dims/test-infra,fejta/test-infra,dims/test-infra,monopole/test-infra,jessfraz/test-infra,jessfraz/test-infra,kubernetes/test-infra,michelle192837/test-infra,brahmaroutu/test-infra,monopole/test-infra,pwittrock/test-infra,kubernetes/test-infra,brahmaroutu/test-infra,michelle192837/test-infra,brahmaroutu/test-infra,pwittrock/test-infra,jessfraz/test-infra,fejta/test-infra,kubernetes/test-infra,cjwagner/test-infra,monopole/test-infra,brahmaroutu/test-infra,cblecker/test-infra,cjwagner/test-infra |
2625b41203d796312cfd9639f3bf8e25235e32bb | .bazelci/presubmit.yml | .bazelci/presubmit.yml | ---
# NOTE: To avoid listing the same things for build_flags/test_flags for each
# of these tasks, they are listed in the .bazelrc instead.
tasks:
macos_latest:
name: "Latest Bazel"
platform: macos
bazel: latest
build_targets:
- "tools/..."
- "test/..."
test_targets:
- "tools/..."
- "test/..."
- "examples/..."
macos_last_green:
name: "Last Green Bazel"
platform: macos
bazel: last_green
build_targets:
- "tools/..."
- "test/..."
test_targets:
- "tools/..."
- "test/..."
- "examples/..."
macos_latest_head_deps:
name: "Latest Bazel with Head Deps"
platform: macos
bazel: latest
shell_commands:
# Update the WORKSPACE to use head versions of some deps to ensure nothing
# has landed on them breaking this project.
- .bazelci/update_workspace_to_deps_heads.sh
build_targets:
- "tools/..."
- "test/..."
test_targets:
- "tools/..."
- "test/..."
- "examples/..."
buildifier:
version: latest
# Function doc string requirements are a little hard to follow, they did get
# split into multiple warnings, so there is now just one left that causes
# warnings. For the history of some of the issues, see:
# https://github.com/bazelbuild/buildtools/issues/576
warnings: -function-docstring-args
| ---
x_defaults:
# YAML has a feature for "repeated nodes", BazelCI is fine with extra nodes
# it doesn't know about; so that is used to avoid repeating common subparts.
common: &common
platform: macos
build_targets:
- "tools/..."
- "test/..."
test_targets:
- "tools/..."
- "test/..."
- "examples/..."
# NOTE: To avoid listing the same things for build_flags/test_flags for each
# of these tasks, they are listed in the .bazelrc instead.
tasks:
macos_latest:
name: "Latest Bazel"
bazel: latest
<<: *common
macos_last_green:
name: "Last Green Bazel"
bazel: last_green
<<: *common
macos_latest_head_deps:
name: "Latest Bazel with Head Deps"
bazel: latest
shell_commands:
# Update the WORKSPACE to use head versions of some deps to ensure nothing
# has landed on them breaking this project.
- .bazelci/update_workspace_to_deps_heads.sh
<<: *common
buildifier:
version: latest
# Function doc string requirements are a little hard to follow, they did get
# split into multiple warnings, so there is now just one left that causes
# warnings. For the history of some of the issues, see:
# https://github.com/bazelbuild/buildtools/issues/576
warnings: -function-docstring-args
| Use yaml anchors, aliases, and merge to avoid repetition. | Use yaml anchors, aliases, and merge to avoid repetition.
RELNOTES: None.
PiperOrigin-RevId: 252430973
| YAML | apache-2.0 | bazelbuild/rules_apple,bazelbuild/rules_apple,bazelbuild/rules_apple,bazelbuild/rules_apple |
4fea6d16e0904fc4970e637cb0ce89d98afb73ab | packages/bi/bizzlelude.yaml | packages/bi/bizzlelude.yaml | homepage: http://github.com/TheBizzle
changelog-type: ''
hash: 81d83dde07d0a8164793272dfd76f84e821111c1f31c8f78577e55c598974587
test-bench-deps: {}
maintainer: jason.bertsche@gmail.com
synopsis: A lousy Prelude replacement by a lousy dude
changelog: ''
basic-deps:
text: ! '>=1.2.2 && <1.3.0'
base-noprelude: ! '>=4.8 && <5'
containers: ! '>=0.5.7 && <0.6.0'
directory: ! '>=1.3.0 && <1.4.0'
all-versions:
- '0.0.3'
- '0.0.4'
- '0.1.0'
- '1.0.1'
- '1.0.2'
- '1.0.3'
- '1.0.4'
- '1.1.0'
author: Jason Bertsche
latest: '1.1.0'
description-type: haddock
description: A lousy Prelude replacement by a lousy dude. Imports common things while
trying to avoid importing things could cause naming conflicts. Favors `Text` over
`String`. Wants using `IO` to be painful.
license-name: BSD3
| homepage: http://github.com/TheBizzle
changelog-type: ''
hash: 20595839189ef17dd0d6992654dfb7fdb33af25e4bd467f792fd7e01160ee357
test-bench-deps: {}
maintainer: jason.bertsche@gmail.com
synopsis: A lousy Prelude replacement by a lousy dude
changelog: ''
basic-deps:
text: ! '>=1.2.3 && <1.3.0'
base-noprelude: ! '>=4.8 && <5'
containers: ! '>=0.5.11 && <0.6.0'
directory: ! '>=1.3.2 && <1.4.0'
all-versions:
- '0.0.3'
- '0.0.4'
- '0.1.0'
- '1.0.1'
- '1.0.2'
- '1.0.3'
- '1.0.4'
- '1.1.0'
- '1.2.0'
author: Jason Bertsche
latest: '1.2.0'
description-type: haddock
description: A lousy Prelude replacement by a lousy dude. Imports common things while
trying to avoid importing things could cause naming conflicts. Favors `Text` over
`String`. Wants using `IO` to be painful.
license-name: BSD3
| Update from Hackage at 2018-03-05T00:55:55Z | Update from Hackage at 2018-03-05T00:55:55Z
| YAML | mit | commercialhaskell/all-cabal-metadata |
8b780c59f065339cb1c07cb4e88413e2ae067109 | .forestry/settings.yml | .forestry/settings.yml | ---
new_page_extension: md
auto_deploy: false
admin_path:
webhook_url:
sections:
- type: directory
path: journal
label: Journal
create: documents
match: "**/*.md"
- type: directory
path: projects
label: Projects
create: documents
match: "**/*.md"
upload_dir: uploads
public_path: "/uploads"
front_matter_path: ''
use_front_matter_path: false
file_template: ":filename:"
| ---
new_page_extension: md
auto_deploy: false
admin_path:
webhook_url:
sections:
- type: directory
path: journal
label: Journal
create: documents
match: "**/*.md"
- type: directory
path: projects
label: Projects
create: documents
match: "**/*.md"
upload_dir: uploads
public_path: "/uploads"
front_matter_path: ''
use_front_matter_path: false
file_template: ":filename:"
build:
publish_command: gridsome build
output_directory: dist
| Update from Forestry.io - Updated Forestry configuration | Update from Forestry.io - Updated Forestry configuration | YAML | mit | wkadwaikar/wkadwaikar.github.io,gatsbimantico/blog,gatsbimantico/blog,wkadwaikar/wkadwaikar.github.io |
0c58c7f24800d0bd20c5e93f3ce05cb10e9195d2 | .github/dependabot.yml | .github/dependabot.yml | version: 2
updates:
- package-ecosystem: maven
directory: "/"
schedule:
interval: daily
open-pull-requests-limit: 10
| version: 2
updates:
- package-ecosystem: maven
directory: "/"
schedule:
interval: daily
open-pull-requests-limit: 10
- package-ecosystem: "github-actions"
directory: "/"
schedule:
interval: daily
| Use Dependabot to update GitHub Actions | Use Dependabot to update GitHub Actions
| YAML | apache-2.0 | maxmind/minfraud-api-java,maxmind/minfraud-api-java |
1f6ab8b04f5926e7fc715d0fcfc7f1234723e28f | packages/co/constraint.yaml | packages/co/constraint.yaml | homepage: ''
changelog-type: ''
hash: f199db7d27874e311233788db8d4462d3857f0060def6e1ba989a12dabae2445
test-bench-deps: {}
maintainer: m.farkasdyck@gmail.com
synopsis: Reified constraints
changelog: ''
basic-deps:
base: ! '>=4.10 && <5'
category: ! '>=0.1.2 && <0.3'
all-versions:
- 0.1.0.0
- 0.1.0.1
- 0.1.1.0
- 0.1.1.1
- 0.1.2.0
- 0.1.2.1
author: M Farkas-Dyck
latest: 0.1.2.1
description-type: haddock
description: ''
license-name: BSD-3-Clause
| homepage: ''
changelog-type: ''
hash: 0467251601d0c10639d774052f3953f96c4f453501bb96a48d36f72dc039db98
test-bench-deps: {}
maintainer: m.farkasdyck@gmail.com
synopsis: Reified constraints
changelog: ''
basic-deps:
base: ! '>=4.10 && <5'
category: ! '>=0.1.2 && <0.3'
unconstrained: ! '>=0.1 && <0.2'
all-versions:
- 0.1.0.0
- 0.1.0.1
- 0.1.1.0
- 0.1.1.1
- 0.1.2.0
- 0.1.2.1
- 0.1.3.0
author: M Farkas-Dyck
latest: 0.1.3.0
description-type: haddock
description: ''
license-name: BSD-3-Clause
| Update from Hackage at 2018-12-30T18:20:23Z | Update from Hackage at 2018-12-30T18:20:23Z
| YAML | mit | commercialhaskell/all-cabal-metadata |
3e8cb619b8aeb7d7d858081ac6681cb65034e2bb | tasks/main.yml | tasks/main.yml | # SPDX-License-Identifier: BSD-3-Clause
- name: Set version specific variables
include_vars: "{{ item }}"
with_first_found:
- "{{ ansible_distribution }}-{{ ansible_distribution_major_version }}.yml"
- "{{ ansible_distribution }}.yml"
- "{{ ansible_os_family }}.yml"
- "default.yml"
- name: Install packages
package:
name: "{{ network_packages }}"
state: present
- name: Enable network service
service:
name: "{{ network_service_name }}"
state: started
enabled: yes
- name: Configure networking connection profiles
network_connections:
provider: "{{ network_provider | mandatory }}"
ignore_errors: "{{ network_ignore_errors | default(omit) }}"
force_state_change: "{{ network_force_state_change | default(omit) }}"
connections: "{{ network_connections | default([]) }}"
- name: Re-test connectivity
ping:
| # SPDX-License-Identifier: BSD-3-Clause
- name: Set version specific variables
include_vars: "{{ item }}"
with_first_found:
- "{{ ansible_distribution }}-{{ ansible_distribution_major_version }}.yml"
- "{{ ansible_distribution }}.yml"
- "{{ ansible_os_family }}.yml"
- "default.yml"
# needed for ansible_facst.packages
- name: Check which packages are installed
package_facts:
# Depending on the plugins, checking installed packages might be slow
# for example subscription manager might slow this down
# Therefore install packages only when rpm does not find them
- name: Install packages
package:
name: "{{ network_packages }}"
state: present
when:
- not network_packages is subset(ansible_facts.packages.keys())
- name: Enable network service
service:
name: "{{ network_service_name }}"
state: started
enabled: yes
- name: Configure networking connection profiles
network_connections:
provider: "{{ network_provider | mandatory }}"
ignore_errors: "{{ network_ignore_errors | default(omit) }}"
force_state_change: "{{ network_force_state_change | default(omit) }}"
connections: "{{ network_connections | default([]) }}"
- name: Re-test connectivity
ping:
| Work around slow package manager | Work around slow package manager
| YAML | bsd-3-clause | linux-system-roles/network,linux-system-roles/network,linux-system-roles/network |
e2cfe082a35a1c34db40cc088a63fbfadc295f46 | tasks/main.yml | tasks/main.yml | ---
# tasks file for ndenv
- name: Run specified tasks on current platform
include: "{{ ansible_os_family }}.yml"
- block:
- name: Check whether ndenv is installed
shell: "{{ ndenv_login_shell }} -lc 'ndenv versions'"
register: result_ndenv_versions
ignore_errors: yes
changed_when: no
- name: Install ndenv
shell: "{{ ndenv_login_shell }} -lc 'anyenv install ndenv'"
when: "result_ndenv_versions.rc != 0"
- name: Install nodejs
shell: "{{ ndenv_login_shell }} -lc 'ndenv install {{ item }}'"
when: "item not in result_ndenv_versions.stdout"
with_items: "{{ ndenv_node_versions }}"
- name: Check default version
shell: "{{ ndenv_login_shell }} -lc 'ndenv global'"
register: result_ndenv_global
changed_when: no
when: "ndenv_global_version is defined"
- name: Set default version
shell: "{{ ndenv_login_shell }} -lc 'ndenv global {{ ndenv_global_version }}'"
when: "ndenv_global_version is defined and ndenv_global_version not in result_ndenv_global.stdout"
become: yes
become_user: "{{ anyenv_owner }}"
# Ref. https://github.com/ansible/ansible/pull/13424
remote_user: "{{ anyenv_owner }}"
| ---
# tasks file for ndenv
- name: Run specified tasks on current platform
include: "{{ ansible_os_family }}.yml"
- block:
- name: Check whether ndenv is installed
shell: "su -s {{ ndenv_login_shell }} -c 'ndenv versions' - {{ anyenv_owner }}"
register: result_ndenv_versions
ignore_errors: yes
changed_when: no
- name: Install ndenv
shell: "su -s {{ ndenv_login_shell }} -c 'anyenv install ndenv' - {{ anyenv_owner }}"
when: "result_ndenv_versions.rc != 0"
- name: Install nodejs
shell: "su -s {{ ndenv_login_shell }} -c 'ndenv install {{ item }}' - {{ anyenv_owner }}"
when: "item not in result_ndenv_versions.stdout"
with_items: "{{ ndenv_node_versions }}"
- name: Check default version
shell: "su -s {{ ndenv_login_shell }} -c 'ndenv global' - {{ anyenv_owner }}"
register: result_ndenv_global
changed_when: no
when: "ndenv_global_version is defined"
- name: Set default version
shell: "su -s {{ ndenv_login_shell }} -c 'ndenv global {{ ndenv_global_version }}' - {{ anyenv_owner }}"
when: "ndenv_global_version is defined and ndenv_global_version not in result_ndenv_global.stdout"
become: yes
| Refactor tasks to switch user without remote_user. | Refactor tasks to switch user without remote_user.
- In several cases, remote_user can't be overwritten in this role.
- Following warning occurs, but become_method doesn't work well on
container.
- When become_method=sudo(default), almost containers don't have sudo
command.
- When become_method=su, docker connection plugin of Ansible doesn't
allow 'su' as become_method.
```
[WARNING]: Consider using 'become', 'become_method', and 'become_user' rather than running su
```
| YAML | mit | FGtatsuro/ansible-ndenv |
73a1e8e3f9fd6ee04becde0b0650de43d36a23c4 | tasks/main.yml | tasks/main.yml | ---
# tasks file for consul-template
- name: Write install script
template:
src=install.j2
dest={{consul_template_install_script}}
mode=0755
tags:
- consul-template
- name: Check consul-template version
shell: consul-template --version 2>&1 | grep {{consul_template_ver}}
failed_when: false
changed_when: false
register: consul_template_ver_matches
tags:
- consul-template
- name: Install consul-template
shell: "{{consul_template_install_script}}"
when: consul_template_ver_matches.rc != 0
tags:
- consul-template
- name: Verify binary
stat:
path: "{{consul_template_bin_path}}//{{consul_template_bin_name}}"
register: consul_template_verify_binary
- name: Fail if binary is not present
fail:
msg: consul-template binary is not present
when: not consul_template_verify_binary.stat.exists
- name: Make templates dir
file:
path={{consul_template_templates_dir}}
state=directory
tags:
- consul-template
| ---
# tasks file for consul-template
- name: Write install script
template:
src=install.j2
dest={{consul_template_install_script}}
mode=0755
tags:
- consul-template
- name: Check consul-template version
shell: consul-template --version 2>&1 | grep {{consul_template_ver}}
failed_when: false
changed_when: false
register: consul_template_ver_matches
tags:
- consul-template
- name: Install consul-template
shell: "{{consul_template_install_script}}"
when: consul_template_ver_matches.rc != 0
tags:
- consul-template
- name: Verify binary
stat:
path: "{{consul_template_bin_path}}//{{consul_template_bin_name}}"
register: consul_template_verify_binary
tags:
- consul-template
- name: Fail if binary is not present
fail:
msg: consul-template binary is not present
when: not consul_template_verify_binary.stat.exists
tags:
- consul-template
- name: Make templates dir
file:
path={{consul_template_templates_dir}}
state=directory
tags:
- consul-template
| Add tag to all tasks | Add tag to all tasks
| YAML | mit | mtchavez/ansible-consul-template |
458de30c4f40050bbe52248f711bd6b955a30a32 | packages/gl/glob-posix.yaml | packages/gl/glob-posix.yaml | homepage: https://github.com/rdnetto/glob-posix#readme
changelog-type: ''
hash: 8c87efe365f523a59422bcf80f157a2b59887d085d94021ce0579a3d1679ec23
test-bench-deps:
MissingH: -any
unix: -any
base: -any
tasty-expected-failure: -any
filepath: -any
criterion: -any
glob-posix: -any
tasty-hunit: -any
tasty: -any
Glob: -any
directory: -any
maintainer: rdnetto@gmail.com
synopsis: Haskell bindings for POSIX glob library.
changelog: ''
basic-deps:
base: ! '>=4.7 && <5'
all-versions:
- 0.1.0.0
- 0.1.0.1
- 0.2.0.0
author: R D'Netto
latest: 0.2.0.0
description-type: haddock
description: |-
Wrapper for the glob(3) function. The key functions are glob and globMany.
GNU extensions are supported but contained in a different module to encourage portability.
license-name: LicenseRef-Apache
| homepage: https://github.com/rdnetto/glob-posix#readme
changelog-type: ''
hash: 438df35fc614ea9d355ec1390046d7abcc8afabf23f8e589abc2c9ef54b705be
test-bench-deps:
MissingH: -any
unix: -any
base: -any
tasty-expected-failure: -any
filepath: -any
criterion: -any
glob-posix: -any
tasty-hunit: -any
tasty: -any
Glob: -any
directory: -any
maintainer: rdnetto@gmail.com
synopsis: Haskell bindings for POSIX glob library.
changelog: ''
basic-deps:
base: ! '>=4.7 && <5'
all-versions:
- 0.1.0.0
- 0.1.0.1
- 0.2.0.0
- 0.2.0.1
author: R D'Netto
latest: 0.2.0.1
description-type: haddock
description: |-
Wrapper for the glob(3) function. The key functions are glob and globMany.
GNU extensions are supported but contained in a different module to encourage portability.
license-name: LicenseRef-Apache
| Update from Hackage at 2019-03-03T05:43:42Z | Update from Hackage at 2019-03-03T05:43:42Z
| YAML | mit | commercialhaskell/all-cabal-metadata |
4ad78cf014261e4ded2bf85799ce39466e46c5d4 | build/ci/macos.yml | build/ci/macos.yml | parameters:
mode: 'debug'
sanitize: ''
jobs:
- job: 'macOS_${{ parameters.mode }}'
displayName: 'macOS (${{ parameters.mode }})'
variables:
- name: CCACHE_DIR
value: $(Build.SourcesDirectory)/ccache
pool:
vmImage: macOS-11
steps:
- script: |
brew install ccache
displayName: 'Dependencies'
- template: setup.yml
- template: cache.yml
parameters:
mode: ${{ parameters.mode }}
- script: |
echo "##vso[task.setvariable variable=config]mode=${{ parameters.mode }} sanitize=${{ parameters.sanitize }} XCODE=/Applications/Xcode_13.0.app/Contents/Developer"
displayName: 'Configure'
- ${{ if eq(parameters.mode, 'debug') }}:
- script: |
make -j $(sysctl -n hw.ncpu) $(config) tests
displayName: 'Test'
- ${{ if eq(parameters.mode, 'release') }}:
- script: |
make -j $(sysctl -n hw.ncpu) $(config) libfly
displayName: 'Build'
- template: package.yml
parameters:
contents: 'build/release/clang/x64/etc/libfly-*.tar.bz2'
| parameters:
mode: 'debug'
sanitize: ''
jobs:
- job: 'macOS_${{ parameters.mode }}'
displayName: 'macOS (${{ parameters.mode }})'
variables:
- name: CCACHE_DIR
value: $(Build.SourcesDirectory)/ccache
pool:
vmImage: macOS-11
steps:
- script: |
brew install ccache
displayName: 'Dependencies'
- template: setup.yml
- template: cache.yml
parameters:
mode: ${{ parameters.mode }}
- script: |
echo "##vso[task.setvariable variable=config]mode=${{ parameters.mode }} sanitize=${{ parameters.sanitize }}"
displayName: 'Configure'
- ${{ if eq(parameters.mode, 'debug') }}:
- script: |
make -j $(sysctl -n hw.ncpu) $(config) tests
displayName: 'Test'
- ${{ if eq(parameters.mode, 'release') }}:
- script: |
make -j $(sysctl -n hw.ncpu) $(config) libfly
displayName: 'Build'
- template: package.yml
parameters:
contents: 'build/release/clang/x64/etc/libfly-*.tar.bz2'
| Use default Xcode version on macOS CI | Use default Xcode version on macOS CI
At one point, Xcode 13 wasn't the default.
| YAML | mit | trflynn89/libfly,trflynn89/libfly |
1cbdd28e57c4ff15785520dbfdc60a1fb6664fb3 | tasks/main.yml | tasks/main.yml | ---
# Tasks file for jenkins
- name: Load the OS specific varibles
include_vars: "{{ role_path }}/vars/{{ ansible_os_family }}.yml"
- include: install_debian.yml
tags:
- install
- jenkins
when: ansible_os_family == "Debian"
- name : Restart jenkins if new version installed
include : "{{ role_path }}/handlers/restart_jenkins.yml"
when : jenkins_task_package_install.changed
- name : Waiting the application is available
include : "{{ role_path }}/handlers/waiting_jenkins.yml"
- include: configure.yml
tags:
- configure
- jenkins
- include: manage_cli.yml
tags:
- jenkins
- install
- include: plugins.yml
tags:
- jenkins
- install
| ---
# Tasks file for jenkins
- name: Load the OS specific varibles
include_vars: "{{ role_path }}/vars/{{ ansible_os_family }}.yml"
- include: install_debian.yml
tags:
- install
- jenkins
when: ansible_os_family == "Debian"
- name : Restart jenkins if new version installed
include : "{{ role_path }}/handlers/restart_jenkins.yml"
when : jenkins_task_package_install.changed
- name : Waiting the application is available
include : "{{ role_path }}/handlers/waiting_jenkins.yml"
- name : Configure jenkins launching settings
include: configure.yml
tags:
- configure
- jenkins
- name : Install and configure Jenkins CLI
include: manage_cli.yml
tags:
- jenkins
- install
- name : Manage plugins installations and upgrades
include: plugins.yml
tags:
- jenkins
- install
| Add name key for included files | Add name key for included files
| YAML | mit | infOpen/ansible-role-jenkins |
f46121a22ab430ff0aa65361daede339a6391850 | config/application.example.yml | config/application.example.yml | development: &base
session_secret: xxx
database_url: 'postgres://localhost/app_prototype_development'
assets_server_url: "http://localhost:1234"
precompiled_assets: false
precompiled_assets_host:
app_mailer_from_email:
bugsnag_api_key:
test:
<<: *base
database_url: 'postgres://localhost/app_prototype_test'
production:
<<: *base
| development: &base
session_secret: xxx
database_url: 'postgres://localhost/app_prototype_development'
assets_server_url: "http://localhost:1234"
precompiled_assets: false
precompiled_assets_host:
app_mailer_from_email:
bugsnag_api_key:
postmark_api_key:
test:
<<: *base
database_url: 'postgres://localhost/app_prototype_test'
postmark_api_key: 'POSTMARK_API_TEST'
production:
<<: *base
| Use a dummy Postmark API key in the test environment | Use a dummy Postmark API key in the test environment
| YAML | mit | icelab/rodakase-skeleton,icelab/dry-web-skeleton,icelab/rodakase-skeleton,icelab/dry-web-skeleton,icelab/rodakase-skeleton,icelab/dry-web-skeleton |
0f58621ca23ee51d0e02253917993d2c745e6b51 | packages/ve/vec-optics.yaml | packages/ve/vec-optics.yaml | homepage: https://github.com/phadej/vec
changelog-type: markdown
hash: 1962a9943f8168ccee95c954c378b10eee884248ed9b05db5fa4e7e0c491853a
test-bench-deps: {}
maintainer: Oleg.Grenrus <oleg.grenrus@iki.fi>
synopsis: ! 'Vec: length-indexed (sized) list: optics support'
changelog: |
# Revision history for vec
## 0.3
- Split out of `vec`
basic-deps:
optics-core: ^>=0.2
base: ! '>=4.9 && <4.14'
vec: ^>=0.3
fin: ^>=0.1.1
all-versions:
- '0.3'
author: Oleg Grenrus <oleg.grenrus@iki.fi>
latest: '0.3'
description-type: haddock
description: |-
This package provides [optics](https://hackage.haskell.org/package/optics) and instances
for data types in [vec](https://hackage.haskell.org/package/vec) package.
"Data.Vec.Optics.Instances" provides all the instances.
the other modules provide some named optics too.
license-name: BSD-3-Clause
| homepage: https://github.com/phadej/vec
changelog-type: markdown
hash: 60d786949e94b2812b4a5cc89080166dc8c5362939748a349a71414cc6b04b81
test-bench-deps: {}
maintainer: Oleg.Grenrus <oleg.grenrus@iki.fi>
synopsis: 'Vec: length-indexed (sized) list: optics support'
changelog: |
# Revision history for vec
## 0.3
- Split out of `vec`
basic-deps:
optics-core: '>=0.2 && <0.4'
base: '>=4.9 && <4.15'
vec: ^>=0.3
fin: ^>=0.1.1
all-versions:
- '0.3'
author: Oleg Grenrus <oleg.grenrus@iki.fi>
latest: '0.3'
description-type: haddock
description: |-
This package provides [optics](https://hackage.haskell.org/package/optics) and instances
for data types in [vec](https://hackage.haskell.org/package/vec) package.
"Data.Vec.Optics.Instances" provides all the instances.
the other modules provide some named optics too.
license-name: BSD-3-Clause
| Update from Hackage at 2020-04-16T07:50:15Z | Update from Hackage at 2020-04-16T07:50:15Z
| YAML | mit | commercialhaskell/all-cabal-metadata |
e226cade79cdd40afec89e8c540339476661af76 | packages/wa/wai-thrift.yaml | packages/wa/wai-thrift.yaml | homepage: https://github.com/yogeshsajanikar/wai-thrift
changelog-type: ''
hash: 5310da25fa2299cb667a796abc6176292236057193c49dcf60c04f515f371d12
test-bench-deps: {}
maintainer: yogesh_sajanikar@yahoo.com
synopsis: Thrift transport layer for Wai
changelog: ''
basic-deps:
bytestring: -any
wai: -any
stm: -any
base: ! '>=4.8 && <4.9'
blaze-builder: -any
thrift: -any
http-types: -any
all-versions:
- '0.0.1.0'
- '0.0.1.1'
author: Yogesh Sajanikar
latest: '0.0.1.1'
description-type: haddock
description: ! 'Implements a read-only transport layer for Wai
Request, and write-only transport layer for Wai
streaming body.'
license-name: MIT
| homepage: https://github.com/yogeshsajanikar/wai-thrift
changelog-type: ''
hash: f776d8cac8165b0099030013a75000c52d1f86fc8788b29f19fdfa99f927f0b8
test-bench-deps: {}
maintainer: yogesh_sajanikar@yahoo.com
synopsis: Thrift transport layer for Wai
changelog: ''
basic-deps:
bytestring: ! '>=0.10'
wai: ! '>=3.0'
base: ! '>=4.8 && <4.9'
blaze-builder: ! '>=0.4'
thrift: ==0.9.3
http-types: ! '>=0.8'
all-versions:
- '0.0.1.0'
- '0.0.1.1'
- '0.0.1.2'
- '0.0.1.3'
author: Yogesh Sajanikar
latest: '0.0.1.3'
description-type: haddock
description: ! 'Implements a read-only transport layer for Wai
Request, and write-only transport layer for Wai
streaming body.'
license-name: MIT
| Update from Hackage at 2015-10-21T16:37:19+0000 | Update from Hackage at 2015-10-21T16:37:19+0000
| YAML | mit | commercialhaskell/all-cabal-metadata |
8d6835987baa0fba2ea50de5d297ef3adebc7d79 | packages/ye/yesod-auth-hmac-keccak.yaml | packages/ye/yesod-auth-hmac-keccak.yaml | homepage: ''
changelog-type: markdown
hash: cbb31fadf64e20903d026b879e30ee273392b3634b6838626969e258cab8e45d
test-bench-deps: {}
maintainer: nek0@chelnok.de
synopsis: ! 'An account authentication plugin for yesod with encrypted
token transfer.'
changelog: ! '# Revision history for yesod-auth-hmac-keccak
## 0.0.0.0 -- YYYY-mm-dd
* First version. Released on an unsuspecting world.
'
basic-deps:
shakespeare: -any
bytestring: -any
yesod-core: -any
base: ! '>=4.9 && <4.11'
persistent: -any
text: -any
yesod-persistent: -any
yesod-form: -any
cryptonite: -any
mtl: -any
random: -any
aeson: -any
yesod-static: -any
yesod-auth: -any
all-versions:
- '0.0.0.1'
- '0.0.0.2'
- '0.0.0.3'
author: nek0
latest: '0.0.0.3'
description-type: haddock
description: ! 'This authentication plugin for Yesod uses a challenge-response
authentication mechanism, where the password is encrypted via HMAC-
Keccak on the client side.'
license-name: MIT
| homepage: ''
changelog-type: markdown
hash: a6424cb27dbdfce09e6d7563ea7d72bc81c2dd3eba633d5a2c5756a494cc2c12
test-bench-deps: {}
maintainer: nek0@chelnok.de
synopsis: ! 'An account authentication plugin for yesod with encrypted
token transfer.'
changelog: ! '# Revision history for yesod-auth-hmac-keccak
## 0.0.0.0 -- YYYY-mm-dd
* First version. Released on an unsuspecting world.
'
basic-deps:
shakespeare: -any
bytestring: -any
yesod-core: -any
base: ! '>=4.9 && <4.11'
persistent: -any
text: -any
yesod-persistent: -any
yesod-form: -any
cryptonite: -any
mtl: -any
random: -any
aeson: -any
yesod-static: -any
yesod-auth: -any
all-versions:
- '0.0.0.1'
- '0.0.0.2'
- '0.0.0.3'
- '0.0.0.4'
author: nek0
latest: '0.0.0.4'
description-type: haddock
description: ! 'This authentication plugin for Yesod uses a challenge-response
authentication mechanism, where the password is encrypted via HMAC-
Keccak on the client side.'
license-name: MIT
| Update from Hackage at 2017-12-24T07:01:21Z | Update from Hackage at 2017-12-24T07:01:21Z
| YAML | mit | commercialhaskell/all-cabal-metadata |
98ad481fbf5eb925fe6569719a4c32ce1febf22f | packages/pa/parsec-free.yaml | packages/pa/parsec-free.yaml | homepage: https://github.com/jwiegley/parsec-free
changelog-type: ''
hash: b6ecc93a03f98d6e062ac00a50d3616cbd9bff40d3b899d9f2b7c45710b19552
test-bench-deps: {}
maintainer: johnw@newartisans.com
synopsis: Parsec API encoded as a deeply-embedded DSL, for debugging and analysis
changelog: ''
basic-deps:
free: -any
bytestring: -any
base: ! '>=4 && <5'
text: ! '>=0.2 && <1.3'
parsec: ! '>=3.1.11'
mtl: -any
transformers: ! '>=0.4 && <0.6'
all-versions:
- '3.1.11.1'
- '3.1.11.2'
- '3.1.11.3'
author: John Wiegley
latest: '3.1.11.3'
description-type: haddock
description: Parsec API encoded as a deeply-embedded DSL, for debugging and analysis
license-name: MIT
| homepage: https://github.com/jwiegley/parsec-free
changelog-type: ''
hash: 900acd53ec3d2610c3d8ad8d5cbcb6f61a458bd597c48cc9d0dcbc02ed7d33d5
test-bench-deps: {}
maintainer: johnw@newartisans.com
synopsis: Parsec API encoded as a deeply-embedded DSL, for debugging and analysis
changelog: ''
basic-deps:
free: -any
bytestring: -any
base: ! '>=4 && <5'
text: ! '>=0.2 && <1.3'
parsec: ! '>=3.1.11'
containers: -any
lens: -any
mtl: -any
transformers: ! '>=0.4 && <0.6'
all-versions:
- '3.1.11.1'
- '3.1.11.2'
- '3.1.11.3'
- '3.1.11.4'
author: John Wiegley
latest: '3.1.11.4'
description-type: haddock
description: Parsec API encoded as a deeply-embedded DSL, for debugging and analysis
license-name: MIT
| Update from Hackage at 2016-10-11T00:02:03+00:00 | Update from Hackage at 2016-10-11T00:02:03+00:00
| YAML | mit | commercialhaskell/all-cabal-metadata |
08e8d79d689fb393468fcabb63f7dcee25efa0e3 | packages/pr/primes-type.yaml | packages/pr/primes-type.yaml | homepage: https://github.com/kindaro/primes-type#readme
changelog-type: ''
hash: 71f9a403061697291463bb459197e8e797a9fc7dc1a18b69eee1be4035b6d4f0
test-bench-deps:
HTF: ! '>=0.13 && <0.14'
base: ! '>=4.8 && <5'
primes: ! '>=0.2 && <0.3'
primes-type: -any
maintainer: kindaro@gmail.com
synopsis: Type-safe prime numbers.
changelog: ''
basic-deps:
base: ! '>=4.8 && <4.11'
primes: ! '>=0.2 && <0.3'
all-versions:
- '0.2.0.0'
- '0.2.0.1'
author: Ignat Insarov
latest: '0.2.0.1'
description-type: haddock
description: This library provides type safe prime numbers. The idea is based upon
the concept of a predicate type from type theory.
license-name: ISC
| homepage: https://github.com/kindaro/primes-type#readme
changelog-type: ''
hash: 1c389071fbe4bd78d0bb2ac205b69f88425ebd859bbdf79482ffa08dcd439d6c
test-bench-deps:
HTF: ! '>=0.13 && <0.14'
base: ! '>=4.8 && <5'
primes: ==0.2.*
criterion: -any
primes-type: -any
maintainer: kindaro@gmail.com
synopsis: Type-safe prime numbers.
changelog: ''
basic-deps:
base: ! '>=4.8 && <4.11'
primes: ! '>=0.2 && <0.3'
deepseq: -any
all-versions:
- '0.2.0.0'
- '0.2.0.1'
- '0.2.0.2'
author: Ignat Insarov
latest: '0.2.0.2'
description-type: haddock
description: This library provides type safe prime numbers. The idea is based upon
the concept of a predicate type from type theory.
license-name: ISC
| Update from Hackage at 2017-10-06T00:12:43Z | Update from Hackage at 2017-10-06T00:12:43Z
| YAML | mit | commercialhaskell/all-cabal-metadata |
b2b5fc1c897736c83ecc18bf59188b42ac887243 | packages/sn/snap-stream.yaml | packages/sn/snap-stream.yaml | homepage: ''
changelog-type: ''
hash: acae39c1ad385b96de141a5264ef7c335e3b34d3d718d2cdaf348494b8f02448
test-bench-deps: {}
maintainer: maintainer@obsidian.systems
synopsis: Streaming Snap handlers
changelog: ''
basic-deps:
bytestring: ==0.10.*
base: ! '>=4.9 && <4.12'
io-streams: ! '>=1.4 && <1.6'
snap-core: ==1.0.*
attoparsec: ==0.13.*
all-versions:
- '0.1'
author: Obsidian Systems LLC
latest: '0.1'
description-type: haddock
description: Snap handlers for streaming access with range requests
license-name: BSD-3-Clause
| homepage: ''
changelog-type: markdown
hash: 2096c688612cb1bef40a51d6dabfc1cd3304dcff546f1d423706d4dc7999234e
test-bench-deps: {}
maintainer: maintainer@obsidian.systems
synopsis: Streaming Snap handlers
changelog: |
# Changelog for snap-stream
## 0.1.1.0
* Loosen `base` version bounds
## 0.1
* Initial release: Snap handlers for streaming access with range requests
basic-deps:
bytestring: '>=0.10 && <0.11'
base: '>=4.9 && <4.15'
io-streams: '>=1.4 && <1.6'
snap-core: '>=1.0 && <1.1'
attoparsec: '>=0.13 && <0.14'
all-versions:
- '0.1'
- 0.1.1.0
author: Obsidian Systems LLC
latest: 0.1.1.0
description-type: markdown
description: |
# snap-stream
[Snap](https://hackage.haskell.org/package/snap) handlers for HTTP [range requests](https://developer.mozilla.org/en-US/docs/Web/HTTP/Range_requests).
This package provides a way to respond to an HTTP request for a particular range of bytes. Range requests include a header that specifies which bytes the client is requesting (e.g., `Range: bytes=0-1023`). See the documentation for `serveStreamAs`.
license-name: BSD-3-Clause
| Update from Hackage at 2021-11-16T13:26:18Z | Update from Hackage at 2021-11-16T13:26:18Z
| YAML | mit | commercialhaskell/all-cabal-metadata |
eeced700efef1d5b43b3513d67f41cdab93f9b1b | packages/ve/vector-text.yaml | packages/ve/vector-text.yaml | homepage: https://github.com/luna/vector-text
changelog-type: ''
hash: 56eb6d61eb0b8b7ea8529c80f04faea5dae5922f7790b17ebf1b087cf2763f1d
test-bench-deps: {}
maintainer: Wojciech Danilo <wojciech.danilo@luna-lang.org>
synopsis: Text implementation based on unboxed char vector.
changelog: ''
basic-deps:
base: ! '>=4.10 && <4.12'
text: -any
prologue: -any
vector-binary-instances: -any
binary: -any
vector: -any
all-versions:
- '1.1'
- '1.1.1'
- '1.1.2'
- '1.1.3'
author: Luna Team
latest: '1.1.3'
description-type: haddock
description: ''
license-name: Apache-2.0
| homepage: https://github.com/luna/vector-text
changelog-type: ''
hash: 1db139021dae0cf901e2d920f59ad8321019bda7718f5e1245c968bb9169f7fa
test-bench-deps: {}
maintainer: Wojciech Danilo <wojciech.danilo@luna-lang.org>
synopsis: Text implementation based on unboxed char vector.
changelog: ''
basic-deps:
base: ! '>=4.10 && <4.12'
text: -any
prologue: -any
vector-binary-instances: -any
binary: -any
vector: -any
all-versions:
- '1.1'
- '1.1.1'
- '1.1.2'
- '1.1.3'
- '1.1.5'
author: Luna Team
latest: '1.1.5'
description-type: haddock
description: ''
license-name: Apache-2.0
| Update from Hackage at 2018-08-24T10:06:01Z | Update from Hackage at 2018-08-24T10:06:01Z
| YAML | mit | commercialhaskell/all-cabal-metadata |
166e2fc5c03dd8e263dd2db32dd1739e25a27ddf | packages/in/interpolate.yaml | packages/in/interpolate.yaml | homepage: https://github.com/sol/interpolate#readme
changelog-type: ''
hash: bc243859cb5489b5ccd1e771714f79fc801bbd209c6c69b2059156c8f3dcb1ae
test-bench-deps:
bytestring: -any
base: ==4.*
hspec: ! '>=1.5'
text: -any
quickcheck-instances: -any
base-compat: -any
QuickCheck: -any
haskell-src-meta: ! '>=0.8'
template-haskell: -any
maintainer: Simon Hengel <sol@typeful.net>
synopsis: String interpolation done right
changelog: ''
basic-deps:
base: ==4.*
haskell-src-meta: ! '>=0.8'
template-haskell: -any
all-versions:
- 0.0.0
- 0.0.1
- 0.0.2
- 0.0.3
- 0.0.4
- 0.1.0
- 0.1.1
- 0.2.0
author: Simon Hengel <sol@typeful.net>
latest: 0.2.0
description-type: haddock
description: String interpolation done right
license-name: MIT
| homepage: https://github.com/sol/interpolate#readme
changelog-type: ''
hash: e102d1870811afce16a339fc47f75cef6fc093bf931453569f326669186fb1db
test-bench-deps:
bytestring: -any
base: ==4.*
hspec: '>=1.5'
text: -any
quickcheck-instances: -any
base-compat: -any
QuickCheck: -any
haskell-src-meta: '>=0.8'
template-haskell: -any
maintainer: Simon Hengel <sol@typeful.net>
synopsis: String interpolation done right
changelog: ''
basic-deps:
base: ==4.*
haskell-src-meta: '>=0.8'
template-haskell: -any
all-versions:
- 0.0.0
- 0.0.1
- 0.0.2
- 0.0.3
- 0.0.4
- 0.1.0
- 0.1.1
- 0.2.0
- 0.2.1
author: Simon Hengel <sol@typeful.net>
latest: 0.2.1
description-type: haddock
description: String interpolation done right
license-name: MIT
| Update from Hackage at 2020-06-05T11:52:20Z | Update from Hackage at 2020-06-05T11:52:20Z
| YAML | mit | commercialhaskell/all-cabal-metadata |
5fa02622ff1d83109715a856d5ecb7f6d9faa3be | fig.yml | fig.yml | db:
image: orchardup/postgresql
ports:
- "5432"
web:
build: .
command: gin
volumes:
- .:/go/src/github.com/hackedu/backend
ports:
- "3000:3000"
links:
- db
| db:
image: orchardup/postgresql
ports:
- "5432"
redis:
image: orchardup/redis
web:
build: .
command: gin
volumes:
- .:/go/src/github.com/hackedu/backend
ports:
- "3000:3000"
links:
- db
| Add Redis service to Fig. | Add Redis service to Fig.
| YAML | mit | zachlatta/website,zachlatta/website,hackedu/website,zachlatta/website,hackedu/website,hackedu/website |
743d6782e1ef8b8585b1daaabab82742c28577fb | playbooks/group_vars/all.yml | playbooks/group_vars/all.yml |
# Copyright (c) 2015-2018 Agalmic Ventures LLC (www.agalmicventures.com)
#
# Permission is hereby granted, free of charge, to any person obtaining a copy
# of this software and associated documentation files (the "Software"), to deal
# in the Software without restriction, including without limitation the rights
# to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
# copies of the Software, and to permit persons to whom the Software is
# furnished to do so, subject to the following conditions:
#
# The above copyright notice and this permission notice shall be included in
# all copies or substantial portions of the Software.
#
# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
# SOFTWARE.
is_digital_ocean: ansible_system_vendor == 'DigitalOcean'
#NOTE: This check is to work around issues where ansible_virtualization_role is
#incorrectly identified
is_virtual_machine: ansible_virtualization_role == 'guest' or is_digital_ocean
is_physical_machine: not is_virtual_machine
|
# Copyright (c) 2015-2018 Agalmic Ventures LLC (www.agalmicventures.com)
#
# Permission is hereby granted, free of charge, to any person obtaining a copy
# of this software and associated documentation files (the "Software"), to deal
# in the Software without restriction, including without limitation the rights
# to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
# copies of the Software, and to permit persons to whom the Software is
# furnished to do so, subject to the following conditions:
#
# The above copyright notice and this permission notice shall be included in
# all copies or substantial portions of the Software.
#
# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
# SOFTWARE.
is_digital_ocean: "{{ ansible_system_vendor == 'DigitalOcean' }}"
#NOTE: This check is to work around issues where ansible_virtualization_role is
#incorrectly identified
is_virtual_machine: "{{ ansible_virtualization_role == 'guest' or is_digital_ocean }}"
is_physical_machine: "{{ not is_virtual_machine }}"
| Fix missing interpolation in Ansible group_vars | Fix missing interpolation in Ansible group_vars
| YAML | mit | AgalmicVentures/Environment,AgalmicVentures/Environment,AgalmicVentures/Environment |
6ae8e9cc8e715eb3fd7dacb4478ef4066e76b4a7 | readthedocs-environment.yml | readthedocs-environment.yml | channels:
- conda-forge
dependencies:
- python=3.7
- ipython
- numpy
- numba
- scipy
- sparse
- matplotlib
- pandoc
- ipykernel
- nbsphinx
- nbconvert=5.4.1
| channels:
- conda-forge
dependencies:
- python=3.7
- ipython
- numpy
- numba
- scipy
- sparse
- matplotlib
- pandoc
- ipykernel
- nbsphinx
- nbconvert=5.4.1
- pip:
- pyganja
| Add pyganja dependency to ReadTheDocs yml config | Add pyganja dependency to ReadTheDocs yml config
This was only added to the pip config, so didn't work when built in conda
| YAML | bsd-3-clause | arsenovic/clifford,arsenovic/clifford |
6b153ec01b1c27b258bcab9b89ba239ec2c46766 | recipes/deeptools/meta.yaml | recipes/deeptools/meta.yaml | package:
name: deeptools
version: '2.0.0'
source:
fn: deepTools-2.0.0.tar.gz
url: https://pypi.python.org/packages/source/d/deepTools/deepTools-2.0.0.tar.gz
md5: 09821e73f56ddb744a35d01173c6eb29
build:
number: 0
skip: True # [not py27]
requirements:
build:
- python
- setuptools
- numpy x.x
- scipy >=0.15.0
- matplotlib >=1.4.0
- pysam >=0.8.2
- bx-python >=0.7.2
run:
- python
- pybigwig >=0.2.3
- numpy x.x
- scipy >=0.15.0
- matplotlib >=1.4.0
- pysam >=0.8.2
- bx-python >=0.7.2
test:
imports:
- deeptools
commands:
- nosetests --with-doctest deeptools
requires:
- nose
about:
home: https://github.com/fidelram/deepTools
license: GPL3
summary: A set of user-friendly tools for normalization and visualzation of deep-sequencing data
| package:
name: deeptools
version: '2.0.0'
source:
fn: deepTools-2.0.0.tar.gz
url: https://pypi.python.org/packages/source/d/deepTools/deepTools-2.0.0.tar.gz
md5: 09821e73f56ddb744a35d01173c6eb29
build:
number: 0
skip: True # [not py27]
requirements:
build:
- python
- setuptools
- numpy >=1.8.0
- scipy >=0.15.0
- matplotlib >=1.4.0
- pysam >=0.8.2
- bx-python >=0.7.2
run:
- python
- pybigwig >=0.2.3
- numpy >=1.8.0
- scipy >=0.15.0
- matplotlib >=1.4.0
- pysam >=0.8.2
- bx-python >=0.7.2
test:
imports:
- deeptools
commands:
- bamCompare --version
about:
home: https://github.com/fidelram/deepTools
license: GPL3
summary: A set of user-friendly tools for normalization and visualzation of deep-sequencing data
| Remove x.x from numpy and drop the nosetests in favor of just bamCompare --version | Remove x.x from numpy and drop the nosetests in favor of just bamCompare --version
| YAML | mit | bebatut/bioconda-recipes,JenCabral/bioconda-recipes,bow/bioconda-recipes,joachimwolff/bioconda-recipes,JingchaoZhang/bioconda-recipes,acaprez/recipes,matthdsm/bioconda-recipes,cokelaer/bioconda-recipes,colinbrislawn/bioconda-recipes,roryk/recipes,saketkc/bioconda-recipes,keuv-grvl/bioconda-recipes,mdehollander/bioconda-recipes,blankenberg/bioconda-recipes,colinbrislawn/bioconda-recipes,HassanAmr/bioconda-recipes,zwanli/bioconda-recipes,ThomasWollmann/bioconda-recipes,CGATOxford/bioconda-recipes,guowei-he/bioconda-recipes,mcornwell1957/bioconda-recipes,chapmanb/bioconda-recipes,keuv-grvl/bioconda-recipes,hardingnj/bioconda-recipes,zwanli/bioconda-recipes,ivirshup/bioconda-recipes,daler/bioconda-recipes,gvlproject/bioconda-recipes,hardingnj/bioconda-recipes,mdehollander/bioconda-recipes,jasper1918/bioconda-recipes,abims-sbr/bioconda-recipes,ostrokach/bioconda-recipes,shenwei356/bioconda-recipes,CGATOxford/bioconda-recipes,gregvonkuster/bioconda-recipes,mcornwell1957/bioconda-recipes,BIMSBbioinfo/bioconda-recipes,ostrokach/bioconda-recipes,ivirshup/bioconda-recipes,instituteofpathologyheidelberg/bioconda-recipes,shenwei356/bioconda-recipes,xguse/bioconda-recipes,gvlproject/bioconda-recipes,BIMSBbioinfo/bioconda-recipes,abims-sbr/bioconda-recipes,lpantano/recipes,chapmanb/bioconda-recipes,chapmanb/bioconda-recipes,pinguinkiste/bioconda-recipes,jasper1918/bioconda-recipes,gvlproject/bioconda-recipes,acaprez/recipes,rob-p/bioconda-recipes,dmaticzka/bioconda-recipes,oena/bioconda-recipes,gvlproject/bioconda-recipes,rvalieris/bioconda-recipes,bow/bioconda-recipes,BIMSBbioinfo/bioconda-recipes,mdehollander/bioconda-recipes,mdehollander/bioconda-recipes,joachimwolff/bioconda-recipes,instituteofpathologyheidelberg/bioconda-recipes,joachimwolff/bioconda-recipes,Luobiny/bioconda-recipes,ivirshup/bioconda-recipes,bow/bioconda-recipes,gvlproject/bioconda-recipes,BIMSBbioinfo/bioconda-recipes,zachcp/bioconda-recipes,oena/bioconda-recipes,xguse/bioconda-recipes,saketkc/bioconda-recipes,zwanli/bioconda-recipes,keuv-grvl/bioconda-recipes,martin-mann/bioconda-recipes,dkoppstein/recipes,martin-mann/bioconda-recipes,chapmanb/bioconda-recipes,hardingnj/bioconda-recipes,rob-p/bioconda-recipes,HassanAmr/bioconda-recipes,pinguinkiste/bioconda-recipes,ostrokach/bioconda-recipes,ThomasWollmann/bioconda-recipes,oena/bioconda-recipes,guowei-he/bioconda-recipes,acaprez/recipes,CGATOxford/bioconda-recipes,saketkc/bioconda-recipes,matthdsm/bioconda-recipes,roryk/recipes,bow/bioconda-recipes,mcornwell1957/bioconda-recipes,keuv-grvl/bioconda-recipes,martin-mann/bioconda-recipes,pinguinkiste/bioconda-recipes,peterjc/bioconda-recipes,instituteofpathologyheidelberg/bioconda-recipes,phac-nml/bioconda-recipes,BIMSBbioinfo/bioconda-recipes,rvalieris/bioconda-recipes,peterjc/bioconda-recipes,BIMSBbioinfo/bioconda-recipes,HassanAmr/bioconda-recipes,gvlproject/bioconda-recipes,rvalieris/bioconda-recipes,dkoppstein/recipes,martin-mann/bioconda-recipes,xguse/bioconda-recipes,daler/bioconda-recipes,omicsnut/bioconda-recipes,blankenberg/bioconda-recipes,npavlovikj/bioconda-recipes,gregvonkuster/bioconda-recipes,mcornwell1957/bioconda-recipes,gregvonkuster/bioconda-recipes,saketkc/bioconda-recipes,joachimwolff/bioconda-recipes,bow/bioconda-recipes,JenCabral/bioconda-recipes,npavlovikj/bioconda-recipes,blankenberg/bioconda-recipes,JingchaoZhang/bioconda-recipes,chapmanb/bioconda-recipes,CGATOxford/bioconda-recipes,Luobiny/bioconda-recipes,cokelaer/bioconda-recipes,lpantano/recipes,phac-nml/bioconda-recipes,hardingnj/bioconda-recipes,JingchaoZhang/bioconda-recipes,cokelaer/bioconda-recipes,daler/bioconda-recipes,rvalieris/bioconda-recipes,bioconda/recipes,bioconda/bioconda-recipes,bioconda/bioconda-recipes,oena/bioconda-recipes,ivirshup/bioconda-recipes,zachcp/bioconda-recipes,abims-sbr/bioconda-recipes,omicsnut/bioconda-recipes,colinbrislawn/bioconda-recipes,xguse/bioconda-recipes,bioconda/bioconda-recipes,bioconda/recipes,ThomasWollmann/bioconda-recipes,colinbrislawn/bioconda-recipes,jasper1918/bioconda-recipes,acaprez/recipes,abims-sbr/bioconda-recipes,phac-nml/bioconda-recipes,HassanAmr/bioconda-recipes,pinguinkiste/bioconda-recipes,jfallmann/bioconda-recipes,pinguinkiste/bioconda-recipes,peterjc/bioconda-recipes,peterjc/bioconda-recipes,ostrokach/bioconda-recipes,npavlovikj/bioconda-recipes,dmaticzka/bioconda-recipes,zwanli/bioconda-recipes,zwanli/bioconda-recipes,yesimon/bioconda-recipes,saketkc/bioconda-recipes,colinbrislawn/bioconda-recipes,matthdsm/bioconda-recipes,oena/bioconda-recipes,zachcp/bioconda-recipes,bebatut/bioconda-recipes,rob-p/bioconda-recipes,joachimwolff/bioconda-recipes,rvalieris/bioconda-recipes,hardingnj/bioconda-recipes,xguse/bioconda-recipes,lpantano/recipes,pinguinkiste/bioconda-recipes,keuv-grvl/bioconda-recipes,cokelaer/bioconda-recipes,daler/bioconda-recipes,ThomasWollmann/bioconda-recipes,instituteofpathologyheidelberg/bioconda-recipes,yesimon/bioconda-recipes,keuv-grvl/bioconda-recipes,JingchaoZhang/bioconda-recipes,HassanAmr/bioconda-recipes,yesimon/bioconda-recipes,bioconda/recipes,dmaticzka/bioconda-recipes,jfallmann/bioconda-recipes,bebatut/bioconda-recipes,dmaticzka/bioconda-recipes,guowei-he/bioconda-recipes,shenwei356/bioconda-recipes,phac-nml/bioconda-recipes,zwanli/bioconda-recipes,guowei-he/bioconda-recipes,matthdsm/bioconda-recipes,blankenberg/bioconda-recipes,mdehollander/bioconda-recipes,bioconda/bioconda-recipes,phac-nml/bioconda-recipes,ivirshup/bioconda-recipes,jfallmann/bioconda-recipes,dmaticzka/bioconda-recipes,CGATOxford/bioconda-recipes,jasper1918/bioconda-recipes,guowei-he/bioconda-recipes,matthdsm/bioconda-recipes,ThomasWollmann/bioconda-recipes,saketkc/bioconda-recipes,npavlovikj/bioconda-recipes,dkoppstein/recipes,rob-p/bioconda-recipes,yesimon/bioconda-recipes,abims-sbr/bioconda-recipes,colinbrislawn/bioconda-recipes,dmaticzka/bioconda-recipes,ThomasWollmann/bioconda-recipes,JenCabral/bioconda-recipes,peterjc/bioconda-recipes,Luobiny/bioconda-recipes,peterjc/bioconda-recipes,daler/bioconda-recipes,zachcp/bioconda-recipes,lpantano/recipes,martin-mann/bioconda-recipes,bow/bioconda-recipes,bebatut/bioconda-recipes,ostrokach/bioconda-recipes,mcornwell1957/bioconda-recipes,instituteofpathologyheidelberg/bioconda-recipes,matthdsm/bioconda-recipes,mdehollander/bioconda-recipes,rvalieris/bioconda-recipes,jasper1918/bioconda-recipes,ivirshup/bioconda-recipes,Luobiny/bioconda-recipes,omicsnut/bioconda-recipes,joachimwolff/bioconda-recipes,roryk/recipes,HassanAmr/bioconda-recipes,ostrokach/bioconda-recipes,omicsnut/bioconda-recipes,daler/bioconda-recipes,shenwei356/bioconda-recipes,abims-sbr/bioconda-recipes,jfallmann/bioconda-recipes,JenCabral/bioconda-recipes,JenCabral/bioconda-recipes,CGATOxford/bioconda-recipes,gregvonkuster/bioconda-recipes,JenCabral/bioconda-recipes,omicsnut/bioconda-recipes |
2f0e7257f62d84d1849a1d957e5c8cd8e5c8aa49 | tests/acceptance.suite.yml | tests/acceptance.suite.yml | # Codeception Test Suite Configuration
#
# Suite for acceptance tests.
# Perform tests in browser using the WebDriver or PhpBrowser.
# If you need both WebDriver and PHPBrowser tests - create a separate suite.
class_name: AcceptanceTester
modules:
enabled:
- PhpBrowser:
url: 'https://registry.dev/' # put your local url
- Laravel5:
part: ORM
cleanup: false # can't wrap into transaction
- \Helper\Acceptance
| # Codeception Test Suite Configuration
#
# Suite for acceptance tests.
# Perform tests in browser using the WebDriver or PhpBrowser.
# If you need both WebDriver and PHPBrowser tests - create a separate suite.
class_name: AcceptanceTester
modules:
enabled:
- PhpBrowser:
url: 'http://registry.dev/' # put your local url
- Laravel5:
part: ORM
environment_file: .env.testing
- \Helper\Acceptance
| Update the codecept acceptance config | Update the codecept acceptance config
| YAML | agpl-3.0 | jonphipps/Metadata-Registry,jonphipps/Metadata-Registry,jonphipps/Metadata-Registry,jonphipps/Metadata-Registry,jonphipps/Metadata-Registry,jonphipps/Metadata-Registry,jonphipps/Metadata-Registry,jonphipps/Metadata-Registry,jonphipps/Metadata-Registry,jonphipps/Metadata-Registry |
2866b7b10457946bb8575f744a79ed655102df0e | codeship-steps.yml | codeship-steps.yml | - name: download_cache
service: cache
command: true
- name: trigger_build
service: herokudeployment
command: true
- service: integrationtest
command: deploy_to_heroku_app codeship-heroku-deployment
- service: integrationtest
command: heroku run --app codeship-heroku-deployment -- ls
- service: herokudeployment
tag: master
type: push
image_name: codeship/heroku-deployment
registry: https://index.docker.io/v1/
encrypted_dockercfg_path: dockercfg.encrypted
| - name: download_cache
service: cache
command: true
- name: trigger_build
service: herokudeployment
command: true
- service: integrationtest
command: deploy_to_heroku_app codeship-heroku-deployment
- service: integrationtest
command: heroku run --app codeship-heroku-deployment -- true
- service: herokudeployment
tag: master
type: push
image_name: codeship/heroku-deployment
registry: https://index.docker.io/v1/
encrypted_dockercfg_path: dockercfg.encrypted
| Use true instead of ls as test command | Use true instead of ls as test command
| YAML | mit | codeship-library/heroku-deployment,codeship-library/heroku-deployment |
d109924b8700b582aba6ad8fe01ee6d9b9b7c6ca | release/tag/goreleaser.yaml | release/tag/goreleaser.yaml | # Copyright 2019 Google LLC
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
builds:
- env:
- CGO_ENABLED=0
- GO111MODULE=on
goos:
- darwin
- linux
- windows
goarch:
- amd64
ldflags: -s -w -X github.com/GoogleContainerTools/kpt/run.version={{.Version}}
archives:
- files:
- LICENSE*
- README*
- CHANGELOG*
- lib.zip*
name_template: "{{ .ProjectName }}_{{ .Os }}_{{ .Arch }}_{{ .Version }}"
checksum:
name_template: 'checksums.txt'
snapshot:
name_template: "master"
changelog:
sort: asc
filters:
exclude:
- '(?i)^docs:'
- '(?i)^test:'
- '(?i)^regenerate docs'
- Merge pull request
- Merge branch
release:
github:
owner: GoogleContainerTools
name: kpt
| # Copyright 2019 Google LLC
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
builds:
- env:
- CGO_ENABLED=0
- GO111MODULE=on
goos:
- darwin
- linux
- windows
goarch:
- amd64
ldflags: -s -w -X github.com/GoogleContainerTools/kpt/run.version={{.Version}}
archives:
- files:
- LICENSES*
- lib.zip*
name_template: "{{ .ProjectName }}_{{ .Os }}_{{ .Arch }}-{{ .Version }}"
checksum:
name_template: 'checksums.txt'
snapshot:
name_template: "master"
changelog:
sort: asc
filters:
exclude:
- '(?i)^docs:'
- '(?i)^test:'
- '(?i)^regenerate docs'
- Merge pull request
- Merge branch
release:
github:
owner: GoogleContainerTools
name: kpt
| Remove files from release artifacts and improve naming | Remove files from release artifacts and improve naming
| YAML | apache-2.0 | GoogleContainerTools/kpt,GoogleContainerTools/kpt,GoogleContainerTools/kpt,GoogleContainerTools/kpt,GoogleContainerTools/kpt |
d7dcb41b271e6b5d5e74988a979a286185a0093e | tests/model/SiteTreeBrokenLinksTest.yml | tests/model/SiteTreeBrokenLinksTest.yml | Page:
content:
Title: ContentPage
Content: 'This is some partially happy content. It has one missing a skiplink, but does have another <a name="yes-anchor-here">skiplink here</a>.'
about:
Title: About
URLSegment: about
Content: 'about us here <a name="yes-anchor-here">about skiplinks here</a>.'
brokenInternalRedirector:
RedirectionType: Internal
Title: RedirectorPageToBrokenInteralPage
LinkToID: 0
workingInternalRedirector:
RedirectionType: Internal
Title: RedirectorPageToBrokenInteralPage
LinkTo: =>Page.content
File:
privacypolicy:
Name: privacypolicy.pdf
Title: privacypolicy.pdf
Filename: assets/privacypolicy.pdf
ErrorPage:
404:
Title: Page not Found
ErrorCode: 404 | Page:
content:
Title: ContentPage
Content: 'This is some partially happy content. It has one missing a skiplink, but does have another <a name="yes-anchor-here">skiplink here</a>.'
about:
Title: About
URLSegment: about
Content: 'about us here <a name="yes-anchor-here">about skiplinks here</a>.'
brokenInternalRedirector:
RedirectionType: Internal
Title: RedirectorPageToBrokenInteralPage
LinkToID: 0
workingInternalRedirector:
RedirectionType: Internal
Title: RedirectorPageToBrokenInteralPage
LinkTo: =>Page.content
File:
privacypolicy:
Name: privacypolicy.pdf
Title: privacypolicy.pdf
Filename: assets/privacypolicy.pdf
ErrorPage:
404:
Title: Page not Found
ErrorCode: 404 | Remove tab - invalid YAML | FIX: Remove tab - invalid YAML
| YAML | bsd-3-clause | jonom/silverstripe-cms,jonom/silverstripe-cms,silverstripe/silverstripe-cms,silverstripe/silverstripe-cms |
30899b1e9d8c213d4d10d352aee242a3d632fd14 | .github/workflows/generate-documentation.yml | .github/workflows/generate-documentation.yml | name: Manually generate documentation
on:
workflow_dispatch
permissions:
contents: read
pages: write
id-token: write
packages: read
concurrency:
group: "update-standard-library-documentation"
cancel-in-progress: true
jobs:
generate-documentation:
name: Generate documentation for release
environment:
name: github-pages
url: ${{ steps.deployment.outputs.page_url }}
runs-on: ubuntu-latest
container:
image: ghcr.io/ponylang/ponyc-ci-stdlib-builder:latest
credentials:
username: ${{ github.repository_owner }}
password: ${{ secrets.GITHUB_TOKEN }}
steps:
- name: Checkout
uses: actions/checkout@v3
- name: Build
run: "bash .ci-scripts/build-stdlib-documentation.bash"
- name: Setup Pages
uses: actions/configure-pages@v2
- name: Upload artifact
uses: actions/upload-pages-artifact@v1
with:
path: 'stdlib-docs/site/'
- name: Deploy to GitHub Pages
id: deployment
uses: actions/deploy-pages@v1
| name: Manually generate documentation
on:
workflow_dispatch
permissions:
contents: read
pages: write
id-token: write
packages: read
concurrency:
group: "update-standard-library-documentation"
cancel-in-progress: true
jobs:
generate-documentation:
name: Generate documentation for release
environment:
name: github-pages
url: ${{ steps.deployment.outputs.page_url }}
runs-on: ubuntu-latest
container:
image: ghcr.io/ponylang/ponyc-ci-stdlib-builder:release
credentials:
username: ${{ github.repository_owner }}
password: ${{ secrets.GITHUB_TOKEN }}
steps:
- name: Checkout
uses: actions/checkout@v3
- name: Build
run: "bash .ci-scripts/build-stdlib-documentation.bash"
- name: Setup Pages
uses: actions/configure-pages@v2
- name: Upload artifact
uses: actions/upload-pages-artifact@v1
with:
path: 'stdlib-docs/site/'
- name: Deploy to GitHub Pages
id: deployment
uses: actions/deploy-pages@v1
| Put generated documentation step back to using release | Put generated documentation step back to using release
| YAML | bsd-2-clause | ponylang/ponyc,ponylang/ponyc,ponylang/ponyc |
2e886f6c2c8483c673b329acd8052bc0dcabf449 | image-yaml/overcloud-images.yaml | image-yaml/overcloud-images.yaml | disk_images:
-
imagename: overcloud
arch: amd64
type: qcow2
elements:
- hosts baremetal dhcp-all-interfaces os-collect-config overcloud-full overcloud-controller overcloud-compute overcloud-ceph-storage heat-config-puppet heat-config-script puppet-modules hiera os-net-config stable-interface-names grub2 element-manifest network-gateway dynamic-login enable-packages-install pip-and-virtualenv-override
packages:
- python-psutil
- python-debtcollector
- plotnetcfg
- sos
- python-networking-cisco
- python-UcsSdk
- device-mapper-multipath
- python-networking-bigswitch
- openstack-neutron-bigswitch-lldp
- openstack-neutron-bigswitch-agent
options:
- "--min-tmpfs 5"
-
imagename: ramdisk_agent
arch: amd64
type: qcow2
elements:
- ironic-agent dynamic-login element-manifest network-gateway epel enable-packages-install pip-and-virtualenv-override
packages:
- python-hardware-detect
options:
- "--min-tmpfs=5"
| disk_images:
-
imagename: overcloud
arch: amd64
type: qcow2
elements:
- hosts baremetal dhcp-all-interfaces os-collect-config overcloud-full overcloud-controller overcloud-compute overcloud-ceph-storage heat-config-puppet heat-config-script puppet-modules hiera os-net-config stable-interface-names grub2 element-manifest network-gateway dynamic-login enable-packages-install pip-and-virtualenv-override
packages:
- python-psutil
- python-debtcollector
- plotnetcfg
- sos
- python-networking-cisco
- python-UcsSdk
- device-mapper-multipath
- python-networking-bigswitch
- openstack-neutron-bigswitch-lldp
- openstack-neutron-bigswitch-agent
options:
- "--min-tmpfs 5"
-
imagename: ramdisk_agent
arch: amd64
type: qcow2
elements:
- ironic-agent dynamic-login element-manifest network-gateway enable-packages-install pip-and-virtualenv-override
packages:
- python-hardware-detect
options:
- "--min-tmpfs=5"
| Remove epel element from base yaml ramdisk agent image | Remove epel element from base yaml ramdisk agent image
The epel element should not be in the base element.
Change-Id: I843e9ed3bd027dd0b975d6b423ba3475631b7bd9
| YAML | apache-2.0 | openstack/tripleo-common,openstack/tripleo-common |
dd03e75be462746978c910690947e84fc340e7b3 | .goreleaser.yml | .goreleaser.yml | builds:
- ldflags:
- -s -w -X github.com/khlieng/dispatch/version.Tag=v{{.Version}} -X github.com/khlieng/dispatch/version.Commit={{.ShortCommit}} -X github.com/khlieng/dispatch/version.Date={{.Date}}
env:
- CGO_ENABLED=0
goos:
- linux
- darwin
- windows
goarch:
- amd64
- arm
- arm64
goarm:
- 6
- 7
archive:
files:
- none*
format_overrides:
- goos: windows
format: zip
checksum:
name_template: "checksums.txt"
changelog:
filters:
exclude:
- ^v.*
- "(?i)^update.*dep"
- Merge pull request
- Merge branch
release:
name_template: "{{.Version}}"
| builds:
- ldflags:
- -s -w -X github.com/khlieng/dispatch/version.Tag=v{{.Version}} -X github.com/khlieng/dispatch/version.Commit={{.ShortCommit}} -X github.com/khlieng/dispatch/version.Date={{.Date}}
env:
- CGO_ENABLED=0
goos:
- linux
- darwin
- windows
goarch:
- amd64
- arm
- arm64
goarm:
- 6
- 7
archive:
files:
- none*
format_overrides:
- goos: windows
format: zip
replacements:
amd64: x64
darwin: mac
checksum:
name_template: "checksums.txt"
changelog:
filters:
exclude:
- ^v.*
- "(?i)^update.*dep"
- Merge pull request
- Merge branch
release:
name_template: "{{.Version}}"
| Add amd64 and darwin archive name replacements | Add amd64 and darwin archive name replacements
| YAML | mit | khlieng/dispatch,khlieng/dispatch,khlieng/name_pending,khlieng/name_pending,khlieng/dispatch,khlieng/name_pending,khlieng/name_pending |
cb3d490b395df11d40cab7d00d40842e3be093df | util/ansible/roles/azuracast-fixtures/tasks/main.yml | util/ansible/roles/azuracast-fixtures/tasks/main.yml | ---
- name: Find fixture ZIP files
find:
paths: "{{ util_base }}/fixtures"
patterns: "*.zip"
recurse: yes
register: fixture_zip_files
- name: Extract fixture ZIP files
unarchive:
src: "{{ item.path }}"
dest: "{{ app_base }}/stations"
owner: azuracast
with_items: "{{ fixture_zip_files.files }}"
when: fixture_zip_files.matched > 0
- name: Find fixture SQL files
find:
paths: "{{ util_base }}/fixtures"
patterns: "*.sql"
recurse: yes
register: fixture_sql_files
- name: Import fixture SQL files
become: true
become_user: azuracast
shell: "php {{ util_base }}/cli.php dbal:import {{ item.path }}"
with_items: "{{ fixture_sql_files.files }}"
when: fixture_sql_files.matched > 0
- name: Clear AzuraCast Cache and Restart Stations
become: true
become_user: azuracast
shell: php {{ util_base }}/cli.php cache:clear && php {{ util_base }}/cli.php azuracast:radio:restart
when: fixture_zip_files.matched > 0 or fixture_sql_files.matched > 0 | ---
- name: Find fixture ZIP files
find:
paths: "{{ util_base }}/fixtures"
patterns: "*.zip"
recurse: yes
register: fixture_zip_files
- name: Extract fixture ZIP files
unarchive:
src: "{{ item.path }}"
dest: "{{ app_base }}/stations"
owner: azuracast
with_items: "{{ fixture_zip_files.files }}"
when: fixture_zip_files.matched > 0
- name: Find fixture SQL files
find:
paths: "{{ util_base }}/fixtures"
patterns: "*.sql"
recurse: yes
register: fixture_sql_files
- name: Import fixture SQL files
become: true
become_user: azuracast
shell: "php {{ util_base }}/cli.php dbal:import {{ item.path }}"
with_items: "{{ fixture_sql_files.files }}"
when: fixture_sql_files.matched > 0
register: fixture_sql_command_result
failed_when: "'exception' in command_result.stdout"
- name: Clear AzuraCast Cache and Restart Stations
become: true
become_user: azuracast
shell: php {{ util_base }}/cli.php cache:clear && php {{ util_base }}/cli.php azuracast:radio:restart
when: fixture_zip_files.matched > 0 or fixture_sql_files.matched > 0 | Add failure criteria to fixture install. | Add failure criteria to fixture install.
| YAML | agpl-3.0 | AzuraCast/AzuraCast,AzuraCast/AzuraCast,SlvrEagle23/AzuraCast,SlvrEagle23/AzuraCast,SlvrEagle23/AzuraCast,SlvrEagle23/AzuraCast,SlvrEagle23/AzuraCast,AzuraCast/AzuraCast,AzuraCast/AzuraCast |
859fee51a0d121842a42fd12da7e617258cc8011 | recipes/dotty-dict/meta.yaml | recipes/dotty-dict/meta.yaml | {% set name = "dotty-dict" %}
{% set version = "1.3.0" %}
package:
name: {{ name|lower }}
version: {{ version }}
pypi_name: "dotty_dict"
source:
url: https://pypi.io/packages/source/{{ pypi_name[0] }}/{{ pypi_name }}/{{ pypi_name }}-{{ version }}.tar.gz
sha256: eb0035a3629ecd84397a68f1f42f1e94abd1c34577a19cd3eacad331ee7cbaf0
build:
number: 0
noarch: python
script: {{ PYTHON }} -m pip install . -vv
requirements:
host:
- pip
- python >=3.5
- setuptools_scm
run:
- python >=3.5
test:
imports:
- dotty_dict
commands:
- pip check
requires:
- pip
about:
home: https://pypi.org/project/dotty-dict/
summary: Dictionary wrapper for quick access to deeply nested keys.
license: MIT
license_file: LICENSE
dev_url: https://github.com/pawelzny/dotty_dict
doc_url: https://dotty-dict.readthedocs.io
extra:
recipe-maintainers:
- KingDarBoja
| {% set name = "dotty-dict" %}
{% set version = "1.3.0" %}
{% set pypi_name = "dotty_dict" %}
package:
name: {{ name|lower }}
version: {{ version }}
pypi_name: {{ pypi_name }}
source:
url: https://pypi.io/packages/source/{{ pypi_name[0] }}/{{ pypi_name }}/{{ pypi_name }}-{{ version }}.tar.gz
sha256: eb0035a3629ecd84397a68f1f42f1e94abd1c34577a19cd3eacad331ee7cbaf0
build:
number: 0
noarch: python
script: {{ PYTHON }} -m pip install . -vv
requirements:
host:
- pip
- python >=3.5
- setuptools_scm
run:
- python >=3.5
test:
imports:
- dotty_dict
commands:
- pip check
requires:
- pip
about:
home: https://pypi.org/project/dotty-dict/
summary: Dictionary wrapper for quick access to deeply nested keys.
license: MIT
license_file: LICENSE
dev_url: https://github.com/pawelzny/dotty_dict
doc_url: https://dotty-dict.readthedocs.io
extra:
recipe-maintainers:
- KingDarBoja
| Use jinja template variable at the top | Use jinja template variable at the top
| YAML | bsd-3-clause | igortg/staged-recipes,scopatz/staged-recipes,mariusvniekerk/staged-recipes,goanpeca/staged-recipes,hadim/staged-recipes,goanpeca/staged-recipes,ReimarBauer/staged-recipes,jochym/staged-recipes,conda-forge/staged-recipes,jochym/staged-recipes,jakirkham/staged-recipes,igortg/staged-recipes,kwilcox/staged-recipes,jakirkham/staged-recipes,patricksnape/staged-recipes,stuertz/staged-recipes,ocefpaf/staged-recipes,johanneskoester/staged-recipes,ocefpaf/staged-recipes,stuertz/staged-recipes,kwilcox/staged-recipes,mariusvniekerk/staged-recipes,patricksnape/staged-recipes,hadim/staged-recipes,scopatz/staged-recipes,johanneskoester/staged-recipes,ReimarBauer/staged-recipes,conda-forge/staged-recipes |
20eaa43d18b9dbaaa66f9bfb26453d4d872e910c | roles/dev/defaults/main.yml | roles/dev/defaults/main.yml | ---
projects:
- { name: idea, version: '2017.3.5', url_prefix: 'https://download.jetbrains.com/idea/ideaIC-' }
- { name: pycharm, version: '2017.3.4', url_prefix: 'https://download.jetbrains.com/python/pycharm-professional-' }
bashrc_path: "{{ ansible_env.HOME }}/.bashrc"
| ---
projects:
- { name: idea, version: '2018.1', url_prefix: 'https://download.jetbrains.com/idea/ideaIC-' }
- { name: pycharm, version: '2018.1', url_prefix: 'https://download.jetbrains.com/python/pycharm-professional-' }
bashrc_path: "{{ ansible_env.HOME }}/.bashrc"
| Upgrade pycharm and intellij idea | Upgrade pycharm and intellij idea
| YAML | mit | andornaut/ansible-workstation |
334627535ea1d9987bfa6aca762230af66cd8618 | roles/fidimag/vars/main.yml | roles/fidimag/vars/main.yml | ---
FIDIMAG_REPO_URL: https://github.com/computationalmodelling/fidimag.git
FIDIMAG_INSTALL_PATH: /opt/fidimag
FIDIMAG_VERSION: v0.0.1
XPRA_DISPLAYS:
- 25
- 30
- 36
- 49
- 50
- 64
- 75
- 81
- 90
- 99
| ---
FIDIMAG_REPO_URL: https://github.com/computationalmodelling/fidimag.git
FIDIMAG_INSTALL_PATH: /opt/fidimag
FIDIMAG_VERSION: v0.0.2
XPRA_DISPLAYS:
- 25
- 30
- 36
- 49
- 50
- 64
- 75
- 81
- 90
- 99
| Build fidimag with the somewhat stable 0.0.2 version for consistency, which fixes a container installation issue. | Build fidimag with the somewhat stable 0.0.2 version for consistency, which fixes a container installation issue.
| YAML | bsd-3-clause | fangohr/virtualmicromagnetics,computationalmodelling/virtualmicromagnetics,computationalmodelling/virtualmicromagnetics,fangohr/virtualmicromagnetics |
461718186328d1b705636e57215a47e3fb59d27e | config/ansible/tasks/gerrit_replication_update.yaml | config/ansible/tasks/gerrit_replication_update.yaml | ---
- include: update_configrepo.yaml
- stat: path=/root/config/gerrit/replication.config
register: rstat
- name: Check if replication.config changed
command: diff /root/config/gerrit/replication.config /home/gerrit/site_path/etc/replication.config
register: diffstat
no_log: True
failed_when: diffstat.rc > 1
- name: Update Gerrit replication.config
command: cp /root/config/gerrit/replication.config /home/gerrit/site_path/etc/
when: rstat.stat.exists == True and diffstat.rc == 1
- name: Reload the replication plugin
command: ssh -p 29418 admin@{{ fqdn }} gerrit plugin reload replication
when: rstat.stat.exists == True and diffstat.rc == 1
- name: Restart the replication
command: ssh -p 29418 admin@{{ fqdn }} replication start --all
when: rstat.stat.exists == True and diffstat.rc == 1
| ---
- include: update_configrepo.yaml
- stat: path=/root/config/gerrit/replication.config
register: rstat
- name: Check if replication.config changed
command: diff /root/config/gerrit/replication.config /home/gerrit/site_path/etc/replication.config
register: diffstat
no_log: True
when: rstat.stat.exists == True
failed_when: diffstat.rc > 1
- name: Update Gerrit replication.config
command: cp /root/config/gerrit/replication.config /home/gerrit/site_path/etc/
when: rstat.stat.exists == True and diffstat.rc == 1
- name: Reload the replication plugin
command: ssh -p 29418 admin@{{ fqdn }} gerrit plugin reload replication
when: rstat.stat.exists == True and diffstat.rc == 1
- name: Restart the replication
command: ssh -p 29418 admin@{{ fqdn }} replication start --all
when: rstat.stat.exists == True and diffstat.rc == 1
| Fix failing diff if /root/config/gerrit/replication.config missing | Fix failing diff if /root/config/gerrit/replication.config missing
If /root/config/gerrit/replication.config is missing the diff will
return with an exit code of 2, and therefore the task will fail.
Due to the earlier stat it's already known if the file exists, so just
checking this using a "when" condition.
Change-Id: I22627f4af1189fdddbef5f8ee8bea5489dd7012f
| YAML | apache-2.0 | enovance/software-factory,enovance/software-factory,enovance/software-factory,enovance/software-factory,enovance/software-factory |
58b151577bfbceb41fed5ba485cadd5ef28905be | roles/scripts/tasks/main.yml | roles/scripts/tasks/main.yml | - name: Install custom script helpers
copy: src={{ item }} dest=/usr/local/bin/{{ item | basename }} mode=0755
with_fileglob:
- ../files/*
- name: Get wp-phptidy for html coding standards
command: curl -L -o /usr/local/bin/wp-phptidy https://raw.githubusercontent.com/scribu/wp-phptidy/master/wp-phptidy.php
- name: Give wp-phptidy execute permissions
file:
path: /usr/local/bin/wp-phptidy
mode: 0755 | - name: Install custom script helpers
copy: src={{ item }} dest=/usr/local/bin/{{ item | basename }} mode=0755
with_fileglob:
- ../files/*
| Remove undocumented and unused wp-phptidy in favor of phpcs | Remove undocumented and unused wp-phptidy in favor of phpcs
| YAML | mit | Seravo/wp-vagrant,Seravo/wp-palvelu-vagrant |
f3a6b2c0fb62823da59eea3986dac6ad9aa7322b | .octorunner.yml | .octorunner.yml | image: golang:latest
script:
- go get -u github.com/kardianos/govendor
- cd /go
- mkdir -p src/github.com/boyvanduuren/
- ln -s /var/run/octorunner /go/src/github.com/boyvanduuren/octorunner
- cd /go/src/github.com/boyvanduuren/octorunner
- govendor sync
- go test ./... | image: golang:latest
script:
- go get -u github.com/kardianos/govendor
- cd /go
- mkdir -p src/github.com/boyvanduuren/
- ln -s /var/run/octorunner /go/src/github.com/boyvanduuren/octorunner
- cd /go/src/github.com/boyvanduuren/octorunner
- govendor sync
- go get
- go test ./...
| Make sure ql dep is downloaded before running tests | Make sure ql dep is downloaded before running tests
| YAML | mit | boyvanduuren/octorunner |
158a871e7d94638e368d190ca8e05a8656b5fc63 | .overcommit.yml | .overcommit.yml | #
# Select version of overcommit and the other tools from Gemfile
#
gemfile: Gemfile
#
# Hooks that are run against every commit message after a user has written it.
#
CommitMsg:
HardTabs:
enabled: true
SingleLineSubject:
enabled: true
#
# Hooks that are run after `git commit` is executed, before the commit message
# editor is displayed.
#
PreCommit:
BundleCheck:
enabled: true
LocalPathsInGemfile:
enabled: true
ExecutePermissions:
enabled: true
exclude:
- bin/*
RuboCop:
enabled: true
TrailingWhitespace:
enabled: true
YamlSyntax:
enabled: true
#
# Hooks that are run before pushing changes
#
PrePush:
RSpec:
enabled: true
| #
# Select version of overcommit and the other tools from Gemfile
#
gemfile: Gemfile
#
# Hooks that are run against every commit message after a user has written it.
#
CommitMsg:
HardTabs:
enabled: true
SingleLineSubject:
enabled: true
#
# Hooks that are run after `git commit` is executed, before the commit message
# editor is displayed.
#
PreCommit:
BundleCheck:
enabled: true
LocalPathsInGemfile:
enabled: true
ExecutePermissions:
enabled: true
exclude:
- bin/*
RuboCop:
enabled: true
exclude:
- 'db/schema.rb'
- 'db/migrate/*.rb'
TrailingWhitespace:
enabled: true
YamlSyntax:
enabled: true
#
# Hooks that are run before pushing changes
#
PrePush:
RSpec:
enabled: true
| Exclude migration related files from rubocop hook | Exclude migration related files from rubocop hook
| YAML | apache-2.0 | CredentialEngine/CredentialRegistry,coffeejunk/CredentialRegistry,learningtapestry/learningregistry,learningtapestry/learningregistry,CredentialEngine/CredentialRegistry,CredentialEngine/CredentialRegistry,coffeejunk/CredentialRegistry,coffeejunk/CredentialRegistry,learningtapestry/learningregistry |
431723522dce7fb6fdfa9a3599a2d374724d3479 | .pep8speaks.yml | .pep8speaks.yml | scanner:
linter: flake8
flake8:
max-line-length: 79 # Using the default of 79 in PEP8
ignore:
- E741 # l and b are valid variable names for the galactic frame
- E226 # Don't force "missing whitespace around arithmetic operator"
- E402 # .conf has to be set in the __init__.py modules imports
- W503 # line break before binary operator
- W504 # we've been perpetually annoyed by W504 "line break after binary operator", since there's often no real alternative
exclude:
- _astropy_init.py
- version.py
| scanner:
linter: flake8
flake8:
max-line-length: 100
ignore:
- E741 # l and b are valid variable names for the galactic frame
- E226 # Don't force "missing whitespace around arithmetic operator"
- E402 # .conf has to be set in the __init__.py modules imports
- W503 # line break before binary operator
- W504 # we've been perpetually annoyed by W504 "line break after binary operator", since there's often no real alternative
exclude:
- _astropy_init.py
- version.py
| Make PEP8speaks max line length agree with tox flake8 codestyle test | Make PEP8speaks max line length agree with tox flake8 codestyle test
| YAML | bsd-3-clause | larrybradley/photutils,astropy/photutils |
7f4727cfa3b307f7500ef08da2ce5907d4c57215 | config/mongoid.yml | config/mongoid.yml | development:
host: localhost
database: cartoque_development
test:
host: localhost
database: cartoque_test
# set these environment variables on your prod server
production:
host: <%= ENV['MONGOID_HOST'] || 'localhost' %>
port: <%= ENV['MONGOID_PORT'] || '27017' %>
username: <%= ENV['MONGOID_USERNAME'] %>
password: <%= ENV['MONGOID_PASSWORD'] %>
database: <%= ENV['MONGOID_DATABASE'] || 'cartoque' %>
# slaves:
# - host: slave1.local
# port: 27018
# - host: slave2.local
# port: 27019
| development:
host: localhost
database: cartoque
test:
host: localhost
database: cartoque_test
# set these environment variables on your prod server
production:
host: <%= ENV['MONGOID_HOST'] || 'localhost' %>
port: <%= ENV['MONGOID_PORT'] || '27017' %>
username: <%= ENV['MONGOID_USERNAME'] %>
password: <%= ENV['MONGOID_PASSWORD'] %>
database: <%= ENV['MONGOID_DATABASE'] || 'cartoque' %>
# slaves:
# - host: slave1.local
# port: 27018
# - host: slave2.local
# port: 27019
| Use the same mongodb database name in development and production modes for simplicity | Use the same mongodb database name in development and production modes for simplicity
| YAML | mit | cartoque/cartoque,jbbarth/cartoque,skylost/cartoque,jbbarth/cartoque,skylost/cartoque,cartoque/cartoque,jbbarth/cartoque,cartoque/cartoque |
e30604311f71e4ac84500b15f54b661165a5287f | test/config/database.yml | test/config/database.yml | <% raise "ENV['DB'] not specified!" unless ENV['DB'] %>
# a bit hackish - tests would require a refactoring instead
mysql2: &mysql2
host: localhost
adapter: mysql2
database: activewarehouse_etl_test
username: root
encoding: utf8
local_infile: true
# the tests would require a rework: disabling casting for now
cast: false
postgresql: &postgresql
adapter: postgresql
database: activewarehouse_etl_test
username: postgres
# TODO - refactor test to avoid using 2 databases maybe?
operational_database:
<<: *<%= ENV['DB'] %>
data_warehouse:
<<: *<%= ENV['DB'] %>
etl_execution:
adapter: mysql2
database: etl_execution
| <% raise "ENV['DB'] not specified!" unless ENV['DB'] %>
# a bit hackish - tests would require a refactoring instead
mysql2: &mysql2
host: 127.0.0.1
adapter: mysql2
database: activewarehouse_etl_test
username: root
encoding: utf8
local_infile: true
# the tests would require a rework: disabling casting for now
cast: false
postgresql: &postgresql
adapter: postgresql
database: activewarehouse_etl_test
username: postgres
# TODO - refactor test to avoid using 2 databases maybe?
operational_database:
<<: *<%= ENV['DB'] %>
data_warehouse:
<<: *<%= ENV['DB'] %>
etl_execution:
<<: *mysql2
database: etl_execution
| Make the test suite start on my machine | Make the test suite start on my machine
| YAML | mit | activewarehouse/activewarehouse-etl |
62c0abcb38c5c836b24c06b78c01bfd655c0bfdb | metadata/bluepie.ad_silence.yml | metadata/bluepie.ad_silence.yml | AntiFeatures:
- NonFreeAdd
Categories:
- System
License: GPL-3.0-only
AuthorName: BluePie
AuthorEmail: iambluepie@gmail.com
AuthorWebSite: https://github.com/aghontpi
WebSite: https://github.com/aghontpi/ad-silence
SourceCode: https://github.com/aghontpi/ad-silence
IssueTracker: https://github.com/aghontpi/ad-silence/issues
AutoName: Ad-silence
RepoType: git
Repo: https://github.com/aghontpi/ad-silence.git
Builds:
- versionName: 0.3.0
versionCode: 7
commit: a28e8eb5751431da5dcf22d185674073fc0e1f6f
subdir: app
sudo:
- apt-get update || apt-get update
- apt-get install -y openjdk-11-jdk
- update-alternatives --auto java
gradle:
- yes
AutoUpdateMode: Version %v
UpdateCheckMode: Tags
CurrentVersion: 0.3.0
CurrentVersionCode: 7
| AntiFeatures:
- NonFreeAdd
Categories:
- System
License: GPL-3.0-only
AuthorName: BluePie
AuthorEmail: iambluepie@gmail.com
AuthorWebSite: https://github.com/aghontpi
WebSite: https://github.com/aghontpi/ad-silence
SourceCode: https://github.com/aghontpi/ad-silence
IssueTracker: https://github.com/aghontpi/ad-silence/issues
AutoName: Ad-silence
RepoType: git
Repo: https://github.com/aghontpi/ad-silence.git
Builds:
- versionName: 0.3.0
versionCode: 7
commit: a28e8eb5751431da5dcf22d185674073fc0e1f6f
subdir: app
sudo:
- apt-get update || apt-get update
- apt-get install -y openjdk-11-jdk
- update-alternatives --auto java
gradle:
- yes
- versionName: 0.4.0
versionCode: 11
commit: 11f63efe95a71fa93770b23d98ae059e984b1e6b
subdir: app
sudo:
- apt-get update || apt-get update
- apt-get install -y openjdk-11-jdk
- update-alternatives --auto java
gradle:
- yes
AutoUpdateMode: Version %v
UpdateCheckMode: Tags
CurrentVersion: 0.4.0
CurrentVersionCode: 11
| Update Ad-silence to 0.4.0 (11) | Update Ad-silence to 0.4.0 (11)
| YAML | agpl-3.0 | f-droid/fdroiddata,f-droid/fdroiddata |
f5532e8520212b55dbb46bbc7736aa1d34fe656b | environment/external.yaml | environment/external.yaml | resources:
httpproxy:
number: 1
flavor: large
image: trusty
networks: default
assign_floating_ip: true
| resources:
httpproxy:
number: 1
flavor: large
image: trusty
networks: default
boot_volume: default_volume
assign_floating_ip: true
| Make httpproxy boot from volume | Make httpproxy boot from volume
so that we can create a new external layout
| YAML | apache-2.0 | devendermishrajio/puppet-rjil,amar266/puppet-rjil,JioCloud/puppet-rjil-keystone,swamireddy/puppet-rjil,bafna/puppet-rjil,swamireddy/puppet-rjil,rakeshmi/puppet-rjil,ajayaa/puppet-rjil,sajuptpm/puppet-rjil-temp,amar266/puppet-rjil,d4devops/puppet-openstack_cloud,sajuptpm/puppet-rjil-temp,anshprat/puppet-rjil,anshprat/puppet-rjil,roopali8/puppet-rjil,vpramo/puppet-rjil,upadhyay-prashant/puppet-rjil,bafna/puppet-rjil,JioCloud/puppet-rjil,upadhyay-prashant/puppet-rjil,vpramo/puppet-rjil,sajuptpm/puppet-rjil,JioCloud/puppet-rjil,d4devops/puppet-openstack_cloud,alokjani/puppet-rjil,rakeshmi/puppet-rjil,sajuptpm/puppet-rjil-temp,rakeshmi/puppet-rjil,JioCloud/puppet-rjil,JioCloud/puppet-rjil-keystone,JioCloud/puppet-rjil-keystone,jiocloudDSS/puppet-dss,akash1808/puppet-rjil,devendermishrajio/puppet-rjil,punituee/puppet-rjil,rahul1aggarwal/puppet-rjil,rahul1aggarwal/puppet-rjil,punituee/puppet-rjil,rakeshmi/puppet-rjil,devendermishrajio/puppet-rjil,jiocloudDSS/puppet-dss,saneax/puppet-rjil,sajuptpm/puppet-rjil,ajayaa/puppet-rjil,rahul1aggarwal/puppet-rjil,bafna/try-dss,bafna/try-dss,rohit-k/puppet-rjil,JioCloud/puppet-rjil,roopali8/puppet-rjil,d4devops/puppet-openstack_cloud,rohit-k/puppet-rjil,anshprat/puppet-rjil,bafna/puppet-rjil,alokjani/puppet-rjil,akash1808/puppet-rjil,devendermishrajio/puppet-rjil,vpramo/puppet-rjil,JioCloud/puppet-rjil-keystone,alokjani/puppet-rjil,upadhyay-prashant/puppet-rjil,ajayaa/puppet-rjil,akash1808/puppet-rjil,sajuptpm/puppet-rjil,upadhyay-prashant/puppet-rjil,swamireddy/puppet-rjil,punituee/puppet-rjil,bafna/puppet-rjil,roopali8/puppet-rjil,jiocloudDSS/puppet-dss,sajuptpm/puppet-rjil-temp,saneax/puppet-rjil,ajayaa/puppet-rjil,bafna/try-dss,saneax/puppet-rjil,rahul1aggarwal/puppet-rjil,alokjani/puppet-rjil,amar266/puppet-rjil,roopali8/puppet-rjil,rohit-k/puppet-rjil,saneax/puppet-rjil,swamireddy/puppet-rjil,jiocloudDSS/puppet-dss,punituee/puppet-rjil,akash1808/puppet-rjil,sajuptpm/puppet-rjil,bafna/try-dss,anshprat/puppet-rjil,vpramo/puppet-rjil,amar266/puppet-rjil,d4devops/puppet-openstack_cloud,rohit-k/puppet-rjil |
239fa847ca7d091b5a8d8d9341e433a211c43de4 | actions/workflows/st2_integration_tests.yaml | actions/workflows/st2_integration_tests.yaml | ---
chain:
-
name: "core_local_date"
ref: "st2cd.action_run"
params:
name: "core_local_date"
action: "core.local"
params: "date"
hosts: "{{hostname}}"
on-success: "core_http_google"
on-failure: "core_http_google"
-
name: "core_http_google"
ref: "st2cd.action_run"
params:
name: "core_http_google"
action: "core.http"
params: "url=http://www.google.com"
hosts: "{{hostname}}"
on-success: "core_remote_single_host"
on-failure: "core_remote_single_host"
-
name: "core_remote_single_host"
ref: "st2cd.action_run"
params:
name: "core_remote_single_host"
action: "core.remote"
params: "hosts={{hostname}} hostname"
hosts: "{{hostname}}"
on-success: "linux_diag_loadavg"
on-failure: "linux_diag_loadavg"
-
name: "linux_diag_loadavg"
ref: "st2cd.action_run"
params:
name: "linux_diag_loadavg"
action: "linux.diag_loadavg"
params: "hostname={{hostname}}"
hosts: "{{hostname}}"
on-success: "aggregator"
on-failure: "aggregator"
-
name: "aggregator"
ref: "st2cd.result_aggregator"
params:
results: "{{__results}}"
default: "core_local_date"
| ---
chain:
-
name: "core_local_date"
ref: "st2cd.action_run"
params:
name: "core_local_date"
action: "core.local"
params: "date"
hosts: "{{hostname}}"
on-success: "core_http_google"
-
name: "core_http_google"
ref: "st2cd.action_run"
params:
name: "core_http_google"
action: "core.http"
params: "url=http://www.google.com"
hosts: "{{hostname}}"
on-success: "core_remote_single_host"
-
name: "core_remote_single_host"
ref: "st2cd.action_run"
params:
name: "core_remote_single_host"
action: "core.remote"
params: "hosts={{hostname}} hostname"
hosts: "{{hostname}}"
on-success: "action_chain_load"
-
name: "action_chain_load"
ref: "st2cd.action_run"
params:
name: "linux_diag_loadavg"
action: "linux.diag_loadavg"
params: "hostname={{hostname}}"
hosts: "{{hostname}}"
default: "core_local_date"
| Make this actually fail if there is a problem | Make this actually fail if there is a problem
| YAML | apache-2.0 | StackStorm/st2cd,StackStorm/st2cd |
0f1ca99984de1b8d421475e7ac5714f954aa4945 | .github/release-drafter.yml | .github/release-drafter.yml | name-template: "v$RESOLVED_VERSION 🌈"
tag-template: "v$RESOLVED_VERSION"
categories:
- title: "🚀 Features"
labels:
- "feature"
- "enhancement"
- title: "🐛 Bug Fixes"
labels:
- "fix"
- "bugfix"
- "bug"
- title: "🧰 Maintenance"
label: "chore"
change-template: "- $TITLE (#$NUMBER)"
version-resolver:
major:
labels:
- "major"
minor:
labels:
- "minor"
patch:
labels:
- "patch"
default: patch
template: |
## Installation
The Third Coast `vendordeps` file is at: http://maven.strykeforce.org/thirdcoast.json
To install, use **Install new libraries (online)** in VS Code or download manually to your project `vendordeps` directory.
## Changes
$CHANGES
| name-template: "v$RESOLVED_VERSION 🌈"
tag-template: "v$RESOLVED_VERSION"
categories:
- title: "🚀 Features"
labels:
- "feature"
- "enhancement"
- title: "🐛 Bug Fixes"
labels:
- "fix"
- "bugfix"
- "bug"
- title: "🧰 Maintenance"
label: "chore"
change-template: "- $TITLE (#$NUMBER)"
version-resolver:
major:
labels:
- "major"
minor:
labels:
- "minor"
patch:
labels:
- "patch"
default: patch
template: |
## Installation
The Third Coast `vendordeps` file is at: http://maven.strykeforce.org/thirdcoast.json
To install, use **Install new libraries (online)** in VS Code or download manually to your project `vendordeps` directory.
```
$ ./gradlew vendordep --url=http://maven.strykeforce.org/thirdcoast.json
```
## Changes
$CHANGES
| Add GradleRIO command to download vendordep | Add GradleRIO command to download vendordep
| YAML | mit | strykeforce/thirdcoast,strykeforce/thirdcoast,strykeforce/thirdcoast |
e7143eface23933a10215f59e50325f5d264c458 | .github/workflows/build.yml | .github/workflows/build.yml | on: push
name: build
jobs:
test:
runs-on: ubuntu-latest
services:
db:
image: postgres:11
ports: ['5432:5432']
options: >-
--health-cmd pg_isready
--health-interval 10s
--health-timeout 5s
--health-retries 5
steps:
- uses: actions/checkout@v2
- uses: actions/setup-elixir@v1.0.0
with:
otp-version: 22.x
elixir-version: 1.9.x
- run: script/cibuild
env:
ACTIONS: true
COVERALLS_REPO_TOKEN: ${{ secrets.COVERALLS_REPO_TOKEN }}
| on: push
name: build
jobs:
test:
runs-on: ubuntu-latest
services:
db:
image: postgres:12.1
ports: ['5432:5432']
options: >-
--health-cmd pg_isready
--health-interval 10s
--health-timeout 5s
--health-retries 5
steps:
- uses: actions/checkout@v2
- uses: actions/setup-elixir@v1.0.0
with:
otp-version: 22.x
elixir-version: 1.9.x
- run: script/cibuild
env:
ACTIONS: true
COVERALLS_REPO_TOKEN: ${{ secrets.COVERALLS_REPO_TOKEN }}
| Use Postgres v12.1 for now until they get their images fixed | Use Postgres v12.1 for now until they get their images fixed
| YAML | mit | lee-dohm/atom-style-tweaks,lee-dohm/atom-style-tweaks,lee-dohm/atom-style-tweaks,lee-dohm/atom-style-tweaks |
76d3d580b5b8f83b5e57183478744a42ea215d01 | .github/workflows/build.yml | .github/workflows/build.yml | name: Build
on:
push:
branches:
- master
pull_request:
branches:
- master
jobs:
build:
runs-on: ubuntu-latest
steps:
- uses: actions/checkout@v2
- uses: actions/cache@v1
with:
path: vendor/bundle
key: ${{ runner.os }}-gems-${{ hashFiles('**/Gemfile.lock') }}
restore-keys: |
${{ runner.os }}-gems-
- name: Set up Ruby 2.6
uses: actions/setup-ruby@v1
with:
ruby-version: 2.6.x
- name: Bundle install
run: |
bundle config path vendor/bundle
bundle install --jobs 4 --retry 3
- name: Build with Jekyll
run: |
bundle exec jekyll build
| name: Build
on:
push:
branches:
- master
pull_request:
branches:
- master
jobs:
build:
runs-on: ubuntu-latest
steps:
- uses: actions/checkout@v2
- uses: actions/cache@v1
with:
path: vendor/bundle
key: ${{ runner.os }}-gems-${{ hashFiles('Gemfile') }}
restore-keys: |
${{ runner.os }}-gems-
- name: Set up Ruby 2.6
uses: actions/setup-ruby@v1
with:
ruby-version: 2.6.x
- name: Bundle install
run: |
bundle config path vendor/bundle
bundle install --jobs 4 --retry 3
- name: Build with Jekyll
run: |
bundle exec jekyll build
| Use cache key of Gemfile | Use cache key of Gemfile
Failure if Gemfile.lock is used:
##[error]The template is not valid.
hashFiles('/home/runner/work/pixyll/pixyll/**/Gemfile.lock') failed.
Search pattern
'/home/runner/work/pixyll/pixyll/**/Gemfile.lock' doesn't match any
file under '/home/runner/work/pixyll/pixyll'
| YAML | mit | muodov/blog,galeone/galeone.github.io,muodov/blog,galeone/galeone.github.io,kongove/kongove.github.com,kongove/kongove.github.com,pepijn/pepijn.github.io,galeone/galeone.github.io,tanmaybinaykiya/tanmaybinaykiya.github.io,deepredsky/deepredsky.github.io,faustovaz/faustovaz.github.io,kongove/kongove.github.com,deepredsky/deepredsky.github.io,faustovaz/faustovaz.github.io,muodov/blog,grimsa/grimsa.github.io,tanmaybinaykiya/tanmaybinaykiya.github.io,pepijn/pepijn.github.io,galeone/galeone.github.io,pepijn/pepijn.github.io,kongove/kongove.github.com,grimsa/grimsa.github.io,galeone/galeone.github.io,tanmaybinaykiya/tanmaybinaykiya.github.io,kongove/kongove.github.com |
9d6df3137a018426b9b125e908abb8a50e7e341b | .github/workflows/build.yml | .github/workflows/build.yml | name: Build
on:
pull_request:
push:
branches:
- main
- develop
jobs:
build:
name: Build
runs-on: ${{ matrix.os }}
strategy:
matrix:
os: [windows-latest, ubuntu-latest, macos-latest]
steps:
- name: Get the sources
uses: actions/checkout@v2
- name: Fetch all history for all tags and branches
run: git fetch --prune --unshallow
- name: Install .NET Core SDK 3.1.301
uses: actions/setup-dotnet@v1
with:
dotnet-version: '3.1.301'
- name: Run Cake script
uses: ecampidoglio/cake-action@v1.1.1
env:
DOTNET_ROLL_FORWARD_ON_NO_CANDIDATE_FX: 2
with:
target: Run-Integration-Tests
cake-bootstrap: true
| name: Build
on:
pull_request:
push:
branches:
- main
- develop
- hotfix/*
jobs:
build:
name: Build
runs-on: ${{ matrix.os }}
strategy:
matrix:
os: [windows-latest, ubuntu-latest, macos-latest]
steps:
- name: Get the sources
uses: actions/checkout@v2
- name: Fetch all history for all tags and branches
run: git fetch --prune --unshallow
- name: Install .NET Core SDK 3.1.301
uses: actions/setup-dotnet@v1
with:
dotnet-version: '3.1.301'
- name: Run Cake script
uses: ecampidoglio/cake-action@v1.1.1
env:
DOTNET_ROLL_FORWARD_ON_NO_CANDIDATE_FX: 2
with:
target: Run-Integration-Tests
cake-bootstrap: true
| Build hotfix branches on GitHub actions | Build hotfix branches on GitHub actions
| YAML | mit | gep13/cake,patriksvensson/cake,devlead/cake,cake-build/cake,patriksvensson/cake,devlead/cake,gep13/cake,cake-build/cake |
3b2a3ad5117480294de822b18fffbc07c22ca75e | .github/workflows/build.yml | .github/workflows/build.yml | name: Build project
on: [push]
jobs:
build_project:
runs-on: ubuntu-20.04
steps:
- name: Set up Git repository
uses: actions/checkout@v1
- name: Build
run: |
mvn test --batch-mode
| name: Build project
on: [push]
jobs:
build_project:
runs-on: ubuntu-20.04
steps:
- name: Set up Git repository
uses: actions/checkout@v1
- name: Build
run: |
mvn test --batch-mode
- name: Install
run: |
mvn install --batch-mode
| Test install step in CI | Test install step in CI
| YAML | apache-2.0 | tudelft-atlarge/graphalytics-platforms-reference,ldbc/ldbc_graphalytics_platforms_reference,ldbc/ldbc_graphalytics_platforms_reference,tudelft-atlarge/graphalytics-platforms-reference |
a961f0ef25663f787096149ad978ca357152eace | .github/workflows/build.yml | .github/workflows/build.yml | name: build
on:
push:
branches: [ main ]
pull_request:
branches: [ main ]
workflow_dispatch:
jobs:
build:
name: ${{ matrix.os }}
runs-on: ${{ matrix.os }}
strategy:
fail-fast: false
matrix:
os: [ macos-latest, ubuntu-latest, windows-latest ]
steps:
- name: Checkout code
uses: actions/checkout@v3
- name: Setup .NET Core SDK
uses: actions/setup-dotnet@v2
- name: Build, Test and Publish
shell: pwsh
run: ./build.ps1
env:
DOTNET_CLI_TELEMETRY_OPTOUT: true
DOTNET_SKIP_FIRST_TIME_EXPERIENCE: true
NUGET_XMLDOC_MODE: skip
| name: build
on:
push:
branches: [ main ]
pull_request:
branches: [ main ]
workflow_dispatch:
jobs:
build:
name: ${{ matrix.os }}
runs-on: ${{ matrix.os }}
strategy:
fail-fast: false
matrix:
os: [ macos-latest, ubuntu-latest, windows-latest ]
steps:
- name: Checkout code
uses: actions/checkout@v3
- name: Setup .NET Core SDK
uses: actions/setup-dotnet@v2
- name: Build, Test and Publish
shell: pwsh
run: ./build.ps1
env:
DOTNET_CLI_TELEMETRY_OPTOUT: true
DOTNET_SKIP_FIRST_TIME_EXPERIENCE: true
DOTNET_SYSTEM_CONSOLE_ALLOW_ANSI_COLOR_REDIRECTION: 1
NUGET_XMLDOC_MODE: skip
TERM: xterm
| Enable colour output from .NET CLI | Enable colour output from .NET CLI
Enable coloured output to the terminal in GitHub Actions for Linux and macOS. | YAML | apache-2.0 | martincostello/project-euler |
dbbe2b9d6bdccb7b869a24532313e47aad49d01d | .github/workflows/maven.yml | .github/workflows/maven.yml | # Licensed to the Apache Software Foundation (ASF) under one or more
# contributor license agreements. See the NOTICE file distributed with
# this work for additional information regarding copyright ownership.
# The ASF licenses this file to You under the Apache License, Version 2.0
# (the "License"); you may not use this file except in compliance with
# the License. You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
name: Java CI
on: [push, pull_request]
jobs:
build:
runs-on: ${{ matrix.os }}
continue-on-error: ${{ matrix.experimental }}
strategy:
matrix:
os: [ubuntu-latest, windows-latest]
java: [ 8, 11, 17 ]
experimental: [false]
# include:
# - java: 18-ea
# os: ubuntu-latest
# experimental: true
steps:
- uses: actions/checkout@v3
- name: Set up JDK ${{ matrix.java }}
uses: actions/setup-java@v3
with:
distribution: adopt
java-version: ${{ matrix.java }}
- name: Build with Maven
run: mvn -V --no-transfer-progress
| # Licensed to the Apache Software Foundation (ASF) under one or more
# contributor license agreements. See the NOTICE file distributed with
# this work for additional information regarding copyright ownership.
# The ASF licenses this file to You under the Apache License, Version 2.0
# (the "License"); you may not use this file except in compliance with
# the License. You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
name: Java CI
on: [push, pull_request]
permissions:
contents: read
jobs:
build:
runs-on: ${{ matrix.os }}
continue-on-error: ${{ matrix.experimental }}
strategy:
matrix:
os: [ubuntu-latest, windows-latest]
java: [ 8, 11, 17 ]
experimental: [false]
# include:
# - java: 18-ea
# os: ubuntu-latest
# experimental: true
steps:
- uses: actions/checkout@v3
- name: Set up JDK ${{ matrix.java }}
uses: actions/setup-java@v3
with:
distribution: adopt
java-version: ${{ matrix.java }}
- name: Build with Maven
run: mvn -V --no-transfer-progress
| Set permissions for GH Actions | Set permissions for GH Actions
| YAML | apache-2.0 | apache/commons-imaging,apache/commons-imaging |
e50a38e9d257067149e62695c6437754fa625562 | .github/workflows/stale.yml | .github/workflows/stale.yml | name: Mark stale issues and pull requests
on:
schedule:
- cron: "30 1 * * *"
jobs:
stale:
runs-on: ubuntu-latest
steps:
- uses: actions/stale@v3
with:
repo-token: ${{ secrets.GITHUB_TOKEN }}
stale-issue-message: 'As this issue has received no new activity, it has been marked as stale'
stale-pr-message: 'As this pull request has received no new activity, it has been marked as stale'
stale-issue-label: 'no-issue-activity'
stale-pr-label: 'no-pr-activity'
exempt-issue-labels: 'Up for Grabs'
| name: Mark stale issues and pull requests
on:
schedule:
- cron: "30 1 * * *"
jobs:
stale:
runs-on: ubuntu-latest
steps:
- uses: actions/stale@v3
with:
repo-token: ${{ secrets.GITHUB_TOKEN }}
stale-issue-message: 'As this issue has received no new activity, it has been marked as stale'
stale-pr-message: 'As this pull request has received no new activity, it has been marked as stale'
stale-issue-label: 'no-issue-activity'
stale-pr-label: 'no-pr-activity'
exempt-issue-labels: 'Up for Grabs,Blocked'
exempt-pr-labels: 'Work in progress'
| Expand list of exempt labels | Expand list of exempt labels | YAML | bsd-3-clause | zarlo/Cosmos,CosmosOS/Cosmos,CosmosOS/Cosmos,zarlo/Cosmos,CosmosOS/Cosmos,zarlo/Cosmos,CosmosOS/Cosmos,zarlo/Cosmos |
1cf1b0248d23cd0789332df868c7690672291710 | .github/workflows/tests.yml | .github/workflows/tests.yml | name: Node CI
on: [push, pull_request]
jobs:
build:
runs-on: ubuntu-latest
strategy:
matrix:
node-version: [12.x, 14.x, 16.x]
steps:
- uses: actions/checkout@v1
- name: Use Node.js ${{ matrix.node-version }}
uses: actions/setup-node@v1
with:
node-version: ${{ matrix.node-version }}
- name: npm install, build, and test
run: |
npm install
npm run ci-test
env:
CI: true | name: Node CI
on:
push:
branches:
- master
pull_request:
branches:
- master
jobs:
build:
runs-on: ubuntu-latest
strategy:
matrix:
node-version: [12.x, 14.x, 16.x]
steps:
- uses: actions/checkout@v1
- name: Use Node.js ${{ matrix.node-version }}
uses: actions/setup-node@v1
with:
node-version: ${{ matrix.node-version }}
- name: npm install, build, and test
run: |
npm install
npm run ci-test
env:
CI: true | Change PR to only run on master | Change PR to only run on master
| YAML | mit | taylorhakes/promise-polyfill |
d75f3a5391a46ea267602dbea79c780436d63320 | ansible/main.yml | ansible/main.yml | ---
- hosts: all
vars:
cf_cli_version: "6.32.0"
maven_version: "3.5.2"
intellij_version: "2017.2.6"
sts_version: "3.9.1"
eclipse_version: "4.7.1a"
docker_compose_version: "1.17.1"
atom_version: "1.20.1"
gradle_version: "4.3.1"
go_version: "1.9.2"
compliance_masonry_version: "1.1.2"
bosh_cli_version: "2.0.45"
yaml_linux_version: "1.13.1"
terraform_version: "0.10.8"
bosh_bootloader_version: "5.4.1"
postman_version: "5.3.2"
gather_facts: yes
become: yes
roles:
- disable_auto_update
- developer_packages
- git
- atom
- docker_service
- docker_compose
- cf_cli
- maven
- gradle
- intellij
- sts
- unity_desktop
- go
- postman
- gitbook
- compliance_masonry
- gcloud
- newrelic
- bosh
| ---
- hosts: all
vars:
cf_cli_version: "6.33.0"
maven_version: "3.5.2"
intellij_version: "2017.2.6"
sts_version: "3.9.1"
eclipse_version: "4.7.1a"
docker_compose_version: "1.17.1"
atom_version: "1.20.1"
gradle_version: "4.3.1"
go_version: "1.9.2"
compliance_masonry_version: "1.1.2"
bosh_cli_version: "2.0.45"
yaml_linux_version: "1.13.1"
terraform_version: "0.11.0"
bosh_bootloader_version: "5.6.0"
postman_version: "5.3.2"
gather_facts: yes
become: yes
roles:
- disable_auto_update
- developer_packages
- git
- atom
- docker_service
- docker_compose
- cf_cli
- maven
- gradle
- intellij
- sts
- unity_desktop
- go
- postman
- gitbook
- compliance_masonry
- gcloud
- newrelic
- bosh
| Update cf-cli, bosh-bootloader & terraform | Update cf-cli, bosh-bootloader & terraform
| YAML | apache-2.0 | cagiti/developer-environment |
e0ad4860f5461a9a1b7bbca45009d12cc6999a42 | .github/workflows/continuous-integration-workflow.yml | .github/workflows/continuous-integration-workflow.yml | name: Continuous Integration
on: [push]
jobs:
build:
runs-on: ubuntu-latest
strategy:
fail-fast: false
matrix:
python-version: ['3.7', '3.8']
steps:
- uses: actions/checkout@v2
- uses: conda-incubator/setup-miniconda@v2
with:
auto-update-conda: true
python-version: ${{ matrix.python }}
- name: Install Tox and any other packages
shell: bash -l {0}
run: conda install -c conda-forge tox-conda
- name: Run Tox
shell: bash -l {0}
run: tox | name: Continuous Integration
on: [push]
jobs:
build:
runs-on: ubuntu-latest
strategy:
fail-fast: false
matrix:
python-version: ['3.7', '3.8']
steps:
- uses: actions/checkout@v2
- uses: conda-incubator/setup-miniconda@v2
with:
auto-update-conda: true
python-version: ${{ matrix.python }}
- name: Install Tox and any other packages
# shell: bash -l {0}
run: conda install -c conda-forge tox-conda
- name: Run Tox
# shell: bash -l {0}
run: tox -e pytest | Remove shell bash command and only run tox -e pytest. | Remove shell bash command and only run tox -e pytest.
| YAML | mit | grmToolbox/grmpy |
7a1ba5e4e1668750b89051d1220383ad7f6e8bbc | spec/acceptance/nodesets/fedora-22.yml | spec/acceptance/nodesets/fedora-22.yml | HOSTS:
fedora-22:
roles:
- master
platform: fedora-22-x86_64
box: boxcutter/fedora22
box_url: https://atlas.hashicorp.com/boxcutter/boxes/fedora22
hypervisor: vagrant
CONFIG:
log_level: verbose
type: foss
| HOSTS:
fedora-22:
roles:
- master
platform: fedora-22-x86_64
box: fedora/22-cloud-base
box_url: https://dl.fedoraproject.org/pub/fedora/linux/releases/22/Cloud/x86_64/Images/Fedora-Cloud-Base-Vagrant-22-20150521.x86_64.vagrant-virtualbox.box
hypervisor: vagrant
CONFIG:
log_level: verbose
type: foss
| Use official Fedora 22 base box for Beaker tests | Use official Fedora 22 base box for Beaker tests
Signed-off-by: Daniel Farrell <65eaad88897e615d9128b5e890b98414aac720cc@redhat.com>
| YAML | bsd-2-clause | srisankethu/puppet-opendaylight,srisankethu/puppet-opendaylight,trozet/puppet-opendaylight,trozet/puppet-opendaylight,umeshksingla/puppet-opendaylight,umeshksingla/puppet-opendaylight |
a9f8fe9eaf2e4cec0189fa6db410b38f1a1ecb28 | .pyup.yml | .pyup.yml | # see https://pyup.io/docs/configuration/ for all available options
schedule: "every week on monday"
# Security updates only - not just any newer version
update: insecure
search: False
requirements:
- requirements.in
- requirements_for_test.txt
| # see https://pyup.io/docs/configuration/ for all available options
schedule: "every week on monday"
# Security updates only - not just any newer version
update: insecure
search: False
requirements:
- requirements.in:
update: insecure
- requirements_for_test.txt:
update: insecure
| Fix PyUp insecure-only version bumps. | Fix PyUp insecure-only version bumps.
PyUp docs:
```
requirements:
- requirements/staging.txt:
# update all dependencies and pin them
update: all
pin: True
- requirements/dev.txt:
# don't update dependencies, use global 'pin' default
update: False
- requirements/prod.txt:
# update insecure only, pin all
update: insecure
pin: True
```
Despite their own documentation implying settings configured at a global
level apply everywhere, it's not true.
https://github.com/pyupio/pyup/issues/238
| YAML | mit | alphagov/notifications-admin,alphagov/notifications-admin,alphagov/notifications-admin,alphagov/notifications-admin |
f57be67dddf857c85e202abdea064a1fa9288ede | tests/all-roles.yml | tests/all-roles.yml | ---
- name: Dummy playbook that includes all roles for syntax checking
hosts: all
sudo: True
roles:
- development
- multimedia
- office
- vagrant
| ---
- name: Dummy playbook that includes all roles for syntax checking
hosts: all
sudo: True
roles:
- development
- multimedia
- office
| Remove Vagrant roles of tests | Remove Vagrant roles of tests
| YAML | mit | JeanFred/APPA |
4ffa7e5c7a0ea1068b4bdead028224b837b573fd | spec/support/broken/site.yml | spec/support/broken/site.yml | :site:
:name: My broken site
:url: http://broke.it
:copyright: 2010
:how-to-break-it:
Just start writing like you don't know what you're doing. | :site:
:name: My broken site
:url: http://broke.it
:copyright: 2010
:how-to-br eak-it:
Just start writing like you don't know what you're doing. | Use really broken yml to test | Use really broken yml to test | YAML | mit | benschwarz/bonsai,benschwarz/bonsai,benschwarz/bonsai |
1a73d55bae86fc746b8e64dafd4ca60007401522 | roles/ceph-mon/tasks/docker/create_configs.yml | roles/ceph-mon/tasks/docker/create_configs.yml | ---
- name: create ceph conf directory
file:
path: /etc/ceph
state: directory
owner: root
group: root
mode: 0644
- name: generate ceph configuration file
config_template:
src: "{{ playbook_dir }}/roles/ceph-common/templates/ceph.conf.j2"
dest: /etc/ceph/ceph.conf
owner: "root"
group: "root"
mode: "0644"
config_overrides: "{{ ceph_conf_overrides }}"
config_type: ini
| ---
- name: create ceph conf directory
file:
path: /etc/ceph
state: directory
owner: root
group: root
mode: 0644
- name: generate ceph configuration file
action: config_template
args:
src: "{{ playbook_dir }}/roles/ceph-common/templates/ceph.conf.j2"
dest: /etc/ceph/ceph.conf
owner: "root"
group: "root"
mode: "0644"
config_overrides: "{{ ceph_conf_overrides }}"
config_type: ini
| Fix mis-use of config_template in docker tasks | Fix mis-use of config_template in docker tasks
As written, generating the config file for ceph-mon in Docker yielded:
ERROR: config_template is not a legal parameter in an Ansible task or
handler
This fixes that error condition.
| YAML | apache-2.0 | guits/ceph-ansible,mcsage/ceph-ansible,ceph/ceph-ansible,rootfs/ceph-ansible,dang/ceph-ansible,jfcoz/ceph-ansible,font/ceph-ansible,travmi/ceph-ansible,bengland2/ceph-ansible,jsaintrocc/ceph-ansible,albertomurillo/ceph-ansible,fgal/ceph-ansible,jfcoz/ceph-ansible,jtaleric/ceph-ansible,fgal/ceph-ansible,verdurin/ceph-ansible-1,guits/ceph-ansible,ceph/ceph-ansible,WingkaiHo/ceph-ansible,jtaleric/ceph-ansible,albertomurillo/ceph-ansible,bengland2/ceph-ansible,jsaintrocc/ceph-ansible,rootfs/ceph-ansible,mcsage/ceph-ansible,travmi/ceph-ansible,WingkaiHo/ceph-ansible,verdurin/ceph-ansible-1,font/ceph-ansible,WingkaiHo/ceph-ansible,albertomurillo/ceph-ansible,dang/ceph-ansible |
600d88a261ec4a04b822d4b3de4eefe07ea0b09d | charts/lagoon-logging/templates/logging.yaml | charts/lagoon-logging/templates/logging.yaml | apiVersion: logging.banzaicloud.io/v1beta1
kind: Logging
metadata:
namespace: {{ .Release.Namespace | quote }}
name: {{ include "lagoon-logging.fullname" . }}
labels:
{{- include "lagoon-logging.labels" . | nindent 4 }}
spec:
fluentd: {}
fluentbit: {}
controlNamespace: {{ .Release.Namespace | quote }}
| apiVersion: logging.banzaicloud.io/v1beta1
kind: Logging
metadata:
namespace: {{ .Release.Namespace | quote }}
name: {{ include "lagoon-logging.fullname" . }}
labels:
{{- include "lagoon-logging.labels" . | nindent 4 }}
spec:
fluentd:
security:
podSecurityContext:
runAsUser: 100
fluentbit: {}
controlNamespace: {{ .Release.Namespace | quote }}
| Define PodSecurityContext for OpenShift compatibility | Define PodSecurityContext for OpenShift compatibility
| YAML | apache-2.0 | amazeeio/lagoon,amazeeio/lagoon,amazeeio/lagoon,amazeeio/lagoon,amazeeio/lagoon,amazeeio/lagoon,amazeeio/lagoon |
3591429347aaa2b8df502f37a6ef54da0ce6cc7b | tasks/promote-candidate-to-rc/task.yml | tasks/promote-candidate-to-rc/task.yml | ---
platform: linux
image_resource:
type: docker-image
source:
repository: cfbuildpacks/ci
inputs:
- name: buildpacks-ci
- name: candidate
- name: version
outputs:
- name: release-candidate
run: buildpacks-ci/tasks/promote-candidate-to-rc/run.sh
| ---
platform: linux
image_resource:
type: docker-image
source:
repository: cfbuildpacks/ci
inputs:
- name: buildpacks-ci
- name: candidate
- name: version
outputs:
- name: release-candidate
run:
path: buildpacks-ci/tasks/promote-candidate-to-rc/run.sh
| Add correct path to promote-candidate-to-rc | Add correct path to promote-candidate-to-rc
Co-authored-by: David Freilich <0106e5ee349bb58d7c51800b6acf5b57cf80aef6@pivotal.io>
| YAML | apache-2.0 | cloudfoundry/buildpacks-ci,cloudfoundry/buildpacks-ci,cloudfoundry/buildpacks-ci,cloudfoundry/buildpacks-ci,cloudfoundry/buildpacks-ci |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.