commit stringlengths 40 40 | old_file stringlengths 4 184 | new_file stringlengths 4 184 | old_contents stringlengths 1 3.6k | new_contents stringlengths 5 3.38k | subject stringlengths 15 778 | message stringlengths 16 6.74k | lang stringclasses 201 values | license stringclasses 13 values | repos stringlengths 6 116k | config stringclasses 201 values | content stringlengths 137 7.24k | diff stringlengths 26 5.55k | diff_length int64 1 123 | relative_diff_length float64 0.01 89 | n_lines_added int64 0 108 | n_lines_deleted int64 0 106 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
2e7b435fd1ba374c78d664b2f93e3a660ef943a7 | History.txt | History.txt | === 0.1.4 / 2009-03-11
* Add clean method to clear out Sitemaps directory
* Make methods chainable
=== 0.1.3 / 2009-03-10
* Initial release
| === 0.2.1 / 2009-03-12
* Normalize path arguments so it no longer matters whether a leading slash is used or not
=== 0.2.0 / 2009-03-11
* Methods are now chainable
=== 0.1.4 / 2009-03-11
* Add clean method to clear out Sitemaps directory
* Make methods chainable
=== 0.1.3 / 2009-03-10
* Initial release
| Update history with 0.2.0 and 0.2.1 version info | Update history with 0.2.0 and 0.2.1 version info
| Text | mit | alexrabarts/big_sitemap | text | ## Code Before:
=== 0.1.4 / 2009-03-11
* Add clean method to clear out Sitemaps directory
* Make methods chainable
=== 0.1.3 / 2009-03-10
* Initial release
## Instruction:
Update history with 0.2.0 and 0.2.1 version info
## Code After:
=== 0.2.1 / 2009-03-12
* Normalize path arguments so it no longer matters whether a leading slash is used or not
=== 0.2.0 / 2009-03-11
* Methods are now chainable
=== 0.1.4 / 2009-03-11
* Add clean method to clear out Sitemaps directory
* Make methods chainable
=== 0.1.3 / 2009-03-10
* Initial release
| + === 0.2.1 / 2009-03-12
+
+ * Normalize path arguments so it no longer matters whether a leading slash is used or not
+
+ === 0.2.0 / 2009-03-11
+
+ * Methods are now chainable
+
=== 0.1.4 / 2009-03-11
* Add clean method to clear out Sitemaps directory
* Make methods chainable
=== 0.1.3 / 2009-03-10
* Initial release
| 8 | 0.888889 | 8 | 0 |
6bd2c1ec5ea16a17163002fafff29b664ba1078a | app/src/main/java/nl/dionsegijn/steppertouchdemo/MainActivity.kt | app/src/main/java/nl/dionsegijn/steppertouchdemo/MainActivity.kt | package nl.dionsegijn.steppertouchdemo
import android.os.Bundle
import androidx.appcompat.app.AppCompatActivity
import android.widget.Toast
import nl.dionsegijn.steppertouch.OnStepCallback
import nl.dionsegijn.steppertouch.StepperTouch
class MainActivity : AppCompatActivity() {
override fun onCreate(savedInstanceState: Bundle?) {
super.onCreate(savedInstanceState)
setContentView(R.layout.activity_main)
val stepperTouch = findViewById(R.id.stepperTouch) as StepperTouch
stepperTouch.stepper.setMin(0)
stepperTouch.stepper.setMax(5)
stepperTouch.stepper.addStepCallback(object : OnStepCallback {
override fun onStep(value: Int, positive: Boolean) {
Toast.makeText(applicationContext, value.toString(), Toast.LENGTH_SHORT).show()
}
})
val bottemStepperTouch = findViewById(R.id.bottomStepperTouch) as StepperTouch
bottemStepperTouch.stepper.setMin(-10)
bottemStepperTouch.stepper.setMax(10)
bottemStepperTouch.enableSideTap(true)
}
}
| package nl.dionsegijn.steppertouchdemo
import android.os.Bundle
import androidx.appcompat.app.AppCompatActivity
import android.widget.Toast
import kotlinx.android.synthetic.main.activity_main.*
import nl.dionsegijn.steppertouch.OnStepCallback
import nl.dionsegijn.steppertouch.StepperTouch
class MainActivity : AppCompatActivity() {
override fun onCreate(savedInstanceState: Bundle?) {
super.onCreate(savedInstanceState)
setContentView(R.layout.activity_main)
stepperTouch.stepper.setMin(0)
stepperTouch.stepper.setMax(5)
stepperTouch.stepper.addStepCallback(object : OnStepCallback {
override fun onStep(value: Int, positive: Boolean) {
Toast.makeText(applicationContext, value.toString(), Toast.LENGTH_SHORT).show()
}
})
bottomStepperTouch.stepper.setMin(-10)
bottomStepperTouch.stepper.setMax(10)
bottomStepperTouch.enableSideTap(true)
}
}
| Use kotlin extensions to retrieve the view | Use kotlin extensions to retrieve the view
| Kotlin | apache-2.0 | DanielMartinus/Stepper-Touch | kotlin | ## Code Before:
package nl.dionsegijn.steppertouchdemo
import android.os.Bundle
import androidx.appcompat.app.AppCompatActivity
import android.widget.Toast
import nl.dionsegijn.steppertouch.OnStepCallback
import nl.dionsegijn.steppertouch.StepperTouch
class MainActivity : AppCompatActivity() {
override fun onCreate(savedInstanceState: Bundle?) {
super.onCreate(savedInstanceState)
setContentView(R.layout.activity_main)
val stepperTouch = findViewById(R.id.stepperTouch) as StepperTouch
stepperTouch.stepper.setMin(0)
stepperTouch.stepper.setMax(5)
stepperTouch.stepper.addStepCallback(object : OnStepCallback {
override fun onStep(value: Int, positive: Boolean) {
Toast.makeText(applicationContext, value.toString(), Toast.LENGTH_SHORT).show()
}
})
val bottemStepperTouch = findViewById(R.id.bottomStepperTouch) as StepperTouch
bottemStepperTouch.stepper.setMin(-10)
bottemStepperTouch.stepper.setMax(10)
bottemStepperTouch.enableSideTap(true)
}
}
## Instruction:
Use kotlin extensions to retrieve the view
## Code After:
package nl.dionsegijn.steppertouchdemo
import android.os.Bundle
import androidx.appcompat.app.AppCompatActivity
import android.widget.Toast
import kotlinx.android.synthetic.main.activity_main.*
import nl.dionsegijn.steppertouch.OnStepCallback
import nl.dionsegijn.steppertouch.StepperTouch
class MainActivity : AppCompatActivity() {
override fun onCreate(savedInstanceState: Bundle?) {
super.onCreate(savedInstanceState)
setContentView(R.layout.activity_main)
stepperTouch.stepper.setMin(0)
stepperTouch.stepper.setMax(5)
stepperTouch.stepper.addStepCallback(object : OnStepCallback {
override fun onStep(value: Int, positive: Boolean) {
Toast.makeText(applicationContext, value.toString(), Toast.LENGTH_SHORT).show()
}
})
bottomStepperTouch.stepper.setMin(-10)
bottomStepperTouch.stepper.setMax(10)
bottomStepperTouch.enableSideTap(true)
}
}
| package nl.dionsegijn.steppertouchdemo
import android.os.Bundle
import androidx.appcompat.app.AppCompatActivity
import android.widget.Toast
+ import kotlinx.android.synthetic.main.activity_main.*
import nl.dionsegijn.steppertouch.OnStepCallback
import nl.dionsegijn.steppertouch.StepperTouch
class MainActivity : AppCompatActivity() {
override fun onCreate(savedInstanceState: Bundle?) {
super.onCreate(savedInstanceState)
setContentView(R.layout.activity_main)
- val stepperTouch = findViewById(R.id.stepperTouch) as StepperTouch
stepperTouch.stepper.setMin(0)
stepperTouch.stepper.setMax(5)
stepperTouch.stepper.addStepCallback(object : OnStepCallback {
override fun onStep(value: Int, positive: Boolean) {
Toast.makeText(applicationContext, value.toString(), Toast.LENGTH_SHORT).show()
}
})
- val bottemStepperTouch = findViewById(R.id.bottomStepperTouch) as StepperTouch
- bottemStepperTouch.stepper.setMin(-10)
? ^
+ bottomStepperTouch.stepper.setMin(-10)
? ^
- bottemStepperTouch.stepper.setMax(10)
? ^
+ bottomStepperTouch.stepper.setMax(10)
? ^
- bottemStepperTouch.enableSideTap(true)
? ^
+ bottomStepperTouch.enableSideTap(true)
? ^
}
} | 9 | 0.290323 | 4 | 5 |
77e3c9dcbb0890d0166d0ee98c58f06839a2de0c | app/assets/stylesheets/outofplace.css | app/assets/stylesheets/outofplace.css | .out-of-place {
display: inline-block;
position: relative;
}
.one {
color: magenta;
top: 500px;
}
.two {
color: blue;
top: 65px;
}
.three {
color: yellow;
top: 205px;
}
.four {
top: 430px;
color: cyan;
}
.five {
color: lime;
top: 130px;
} | .out-of-place {
display: inline-block;
position: relative;
-webkit-transform: rotate(45deg);
}
.one {
color: magenta;
top: 500px;
}
.two {
color: blue;
top: 165px;
}
.three {
color: yellow;
top: 205px;
}
.four {
top: 430px;
color: cyan;
}
.five {
color: lime;
top: 130px;
} | Put random letters at an angle | Put random letters at an angle
| CSS | mit | mhgbrown/website,mhgbrown/website,mhgbrown/website | css | ## Code Before:
.out-of-place {
display: inline-block;
position: relative;
}
.one {
color: magenta;
top: 500px;
}
.two {
color: blue;
top: 65px;
}
.three {
color: yellow;
top: 205px;
}
.four {
top: 430px;
color: cyan;
}
.five {
color: lime;
top: 130px;
}
## Instruction:
Put random letters at an angle
## Code After:
.out-of-place {
display: inline-block;
position: relative;
-webkit-transform: rotate(45deg);
}
.one {
color: magenta;
top: 500px;
}
.two {
color: blue;
top: 165px;
}
.three {
color: yellow;
top: 205px;
}
.four {
top: 430px;
color: cyan;
}
.five {
color: lime;
top: 130px;
} | .out-of-place {
display: inline-block;
position: relative;
+ -webkit-transform: rotate(45deg);
}
.one {
color: magenta;
top: 500px;
}
.two {
color: blue;
- top: 65px;
+ top: 165px;
? +
}
.three {
color: yellow;
top: 205px;
}
.four {
top: 430px;
color: cyan;
}
.five {
color: lime;
top: 130px;
} | 3 | 0.103448 | 2 | 1 |
9bf94604a3d4bb8a35e325e0800635b4a614c55c | cf-plugin/scripts/build-plugins.sh | cf-plugin/scripts/build-plugins.sh |
set -x
go get github.com/Sirupsen/logrus
GOOS=linux GOARCH=386 go build -o ssh-plugin-linux-386 plugin.go
GOOS=linux GOARCH=amd64 go build -o ssh-plugin-linux-amd64 plugin.go
GOOS=windows GOARCH=386 go build -o ssh-plugin-win32.exe plugin.go
GOOS=windows GOARCH=amd64 go build -o ssh-plugin-win64.exe plugin.go
go build -o ssh-plugin-darwin-amd64 plugin.go
shasum -a1 ssh-plugin-*
|
set -x
go get github.com/Sirupsen/logrus
GOOS=linux GOARCH=386 go build -o ssh-plugin-linux-386 .
GOOS=linux GOARCH=amd64 go build -o ssh-plugin-linux-amd64 .
GOOS=windows GOARCH=386 go build -o ssh-plugin-win32.exe .
GOOS=windows GOARCH=amd64 go build -o ssh-plugin-win64.exe .
go build -o ssh-plugin-darwin-amd64 .
shasum -a1 ssh-plugin-*
| Fix the build script to handle the split main | Fix the build script to handle the split main
| Shell | apache-2.0 | sykesm/diego-ssh,cloudfoundry-incubator/diego-ssh,jeaniejung/diego-ssh,cloudfoundry-incubator/diego-ssh-windows,sykesm/diego-ssh,jeaniejung/diego-ssh,cloudfoundry-incubator/diego-ssh-windows | shell | ## Code Before:
set -x
go get github.com/Sirupsen/logrus
GOOS=linux GOARCH=386 go build -o ssh-plugin-linux-386 plugin.go
GOOS=linux GOARCH=amd64 go build -o ssh-plugin-linux-amd64 plugin.go
GOOS=windows GOARCH=386 go build -o ssh-plugin-win32.exe plugin.go
GOOS=windows GOARCH=amd64 go build -o ssh-plugin-win64.exe plugin.go
go build -o ssh-plugin-darwin-amd64 plugin.go
shasum -a1 ssh-plugin-*
## Instruction:
Fix the build script to handle the split main
## Code After:
set -x
go get github.com/Sirupsen/logrus
GOOS=linux GOARCH=386 go build -o ssh-plugin-linux-386 .
GOOS=linux GOARCH=amd64 go build -o ssh-plugin-linux-amd64 .
GOOS=windows GOARCH=386 go build -o ssh-plugin-win32.exe .
GOOS=windows GOARCH=amd64 go build -o ssh-plugin-win64.exe .
go build -o ssh-plugin-darwin-amd64 .
shasum -a1 ssh-plugin-*
|
set -x
go get github.com/Sirupsen/logrus
- GOOS=linux GOARCH=386 go build -o ssh-plugin-linux-386 plugin.go
? ------ --
+ GOOS=linux GOARCH=386 go build -o ssh-plugin-linux-386 .
- GOOS=linux GOARCH=amd64 go build -o ssh-plugin-linux-amd64 plugin.go
? ------ --
+ GOOS=linux GOARCH=amd64 go build -o ssh-plugin-linux-amd64 .
- GOOS=windows GOARCH=386 go build -o ssh-plugin-win32.exe plugin.go
? ------ --
+ GOOS=windows GOARCH=386 go build -o ssh-plugin-win32.exe .
- GOOS=windows GOARCH=amd64 go build -o ssh-plugin-win64.exe plugin.go
? ------ --
+ GOOS=windows GOARCH=amd64 go build -o ssh-plugin-win64.exe .
- go build -o ssh-plugin-darwin-amd64 plugin.go
? ------ --
+ go build -o ssh-plugin-darwin-amd64 .
shasum -a1 ssh-plugin-* | 10 | 0.833333 | 5 | 5 |
377cd1caeb3ae9cdc09f75f5c3305c5dc29de0e4 | README.md | README.md | SoapAsmvar
==========
SoapAsmvar is a software for detecting variants, including Indel & SV, base on long sequence(or de novo Assembly) alignement.
LICENSE
--------
Copyright © 2014-2015
__Author & contributors:__ Shujia Huang, Siyang Liu, Weijian Ye & Junhua Rao <br/>
__Contact :__ huangshujia@genomics.cn & liusiyang.ocean@gmail.com <br/>
__Institute :__ BGI & KU <br/>
__Last Version :__ 2014-05-30 <br/>
1. Overview
2. How to use SoapAsmvar
3. Speed and Memory required
4. Install the software
5. Run the software
6. Software output
7. Please cite the paper ..., if you need .
8. Acknowledgement :
| Asmvar
==========
Asmvar is a software for detecting variants, including Indel & SV, base on long sequence(or de novo Assembly) alignement.
LICENSE
--------
Copyright © 2014-2015
__Author & contributors:__ Shujia Huang, Siyang Liu, Weijian Ye & Junhua Rao <br/>
__Contact :__ huangshujia@genomics.cn & liusiyang.ocean@gmail.com <br/>
__Institute :__ BGI & KU <br/>
__Last Version :__ 2014-05-30 <br/>
1. Overview
2. How to use Asmvar
3. Speed and Memory required
4. Install the software
5. Run the software
6. Software output
7. Please cite the paper ..., if you need .
8. Acknowledgement :
| Change the name of software | Change the name of software
| Markdown | mit | ShujiaHuang/AsmVar,ShujiaHuang/AsmVar,ShujiaHuang/AsmVar,bioinformatics-centre/AsmVar,bioinformatics-centre/AsmVar,bioinformatics-centre/AsmVar,bioinformatics-centre/AsmVar,bioinformatics-centre/AsmVar,ShujiaHuang/AsmVar,ShujiaHuang/AsmVar,bioinformatics-centre/AsmVar,ShujiaHuang/AsmVar | markdown | ## Code Before:
SoapAsmvar
==========
SoapAsmvar is a software for detecting variants, including Indel & SV, base on long sequence(or de novo Assembly) alignement.
LICENSE
--------
Copyright © 2014-2015
__Author & contributors:__ Shujia Huang, Siyang Liu, Weijian Ye & Junhua Rao <br/>
__Contact :__ huangshujia@genomics.cn & liusiyang.ocean@gmail.com <br/>
__Institute :__ BGI & KU <br/>
__Last Version :__ 2014-05-30 <br/>
1. Overview
2. How to use SoapAsmvar
3. Speed and Memory required
4. Install the software
5. Run the software
6. Software output
7. Please cite the paper ..., if you need .
8. Acknowledgement :
## Instruction:
Change the name of software
## Code After:
Asmvar
==========
Asmvar is a software for detecting variants, including Indel & SV, base on long sequence(or de novo Assembly) alignement.
LICENSE
--------
Copyright © 2014-2015
__Author & contributors:__ Shujia Huang, Siyang Liu, Weijian Ye & Junhua Rao <br/>
__Contact :__ huangshujia@genomics.cn & liusiyang.ocean@gmail.com <br/>
__Institute :__ BGI & KU <br/>
__Last Version :__ 2014-05-30 <br/>
1. Overview
2. How to use Asmvar
3. Speed and Memory required
4. Install the software
5. Run the software
6. Software output
7. Please cite the paper ..., if you need .
8. Acknowledgement :
| - SoapAsmvar
? ----
+ Asmvar
==========
- SoapAsmvar is a software for detecting variants, including Indel & SV, base on long sequence(or de novo Assembly) alignement.
? ----
+ Asmvar is a software for detecting variants, including Indel & SV, base on long sequence(or de novo Assembly) alignement.
LICENSE
--------
Copyright © 2014-2015
__Author & contributors:__ Shujia Huang, Siyang Liu, Weijian Ye & Junhua Rao <br/>
__Contact :__ huangshujia@genomics.cn & liusiyang.ocean@gmail.com <br/>
__Institute :__ BGI & KU <br/>
__Last Version :__ 2014-05-30 <br/>
1. Overview
- 2. How to use SoapAsmvar
? ----
+ 2. How to use Asmvar
3. Speed and Memory required
4. Install the software
5. Run the software
6. Software output
7. Please cite the paper ..., if you need .
8. Acknowledgement :
| 6 | 0.26087 | 3 | 3 |
844342dd5e458e3ac677e5491384526059b69577 | models/userModel.js | models/userModel.js | var bcrypt = require('bcrypt');
var mongoose = require('mongoose');
const saltRounds = 10;
var userSchema = new mongoose.Schema({
signupDate: { type: Date, default: Date.now },
username: { type: String, required: true },
password: { type: String, required: true },
email: { type: String, required: true },
name: { type: String }
});
userSchema.pre('save', function(next) {
if (!this.isModified('password')) {
return next();
}
bcrypt.hash(this.password, saltRounds, function(err, hash) {
if (err) {
console.log(err);
return next(err);
}
this.password = hash;
next();
}.bind(this));
});
userSchema.methods.checkPassword = function(password, cb) {
bcrypt.compare(password, this.password, function(err, result) {
if (err) {
console.log(err);
return next(err);
}
cb(result);
});
};
module.exports = mongoose.model('User', userSchema);
| var bcrypt = require('bcrypt');
var mongoose = require('mongoose');
const saltRounds = 10;
var userSchema = new mongoose.Schema({
signupDate: { type: Date, default: Date.now },
username: { type: String, required: true, unique: true },
email: { type: String, required: true, unique: true },
password: { type: String, required: true },
name: { type: String }
});
userSchema.pre('save', function(next) {
if (!this.isModified('password')) {
return next();
}
bcrypt.hash(this.password, saltRounds, function(err, hash) {
if (err) {
console.log(err);
return next(err);
}
this.password = hash;
next();
}.bind(this));
});
userSchema.methods.checkPassword = function(password, cb) {
bcrypt.compare(password, this.password, function(err, result) {
if (err) {
console.log(err);
return next(err);
}
cb(result);
});
};
module.exports = mongoose.model('User', userSchema);
| Add unique index to username and email for Users | Add unique index to username and email for Users
| JavaScript | mit | heymanhn/journey-backend | javascript | ## Code Before:
var bcrypt = require('bcrypt');
var mongoose = require('mongoose');
const saltRounds = 10;
var userSchema = new mongoose.Schema({
signupDate: { type: Date, default: Date.now },
username: { type: String, required: true },
password: { type: String, required: true },
email: { type: String, required: true },
name: { type: String }
});
userSchema.pre('save', function(next) {
if (!this.isModified('password')) {
return next();
}
bcrypt.hash(this.password, saltRounds, function(err, hash) {
if (err) {
console.log(err);
return next(err);
}
this.password = hash;
next();
}.bind(this));
});
userSchema.methods.checkPassword = function(password, cb) {
bcrypt.compare(password, this.password, function(err, result) {
if (err) {
console.log(err);
return next(err);
}
cb(result);
});
};
module.exports = mongoose.model('User', userSchema);
## Instruction:
Add unique index to username and email for Users
## Code After:
var bcrypt = require('bcrypt');
var mongoose = require('mongoose');
const saltRounds = 10;
var userSchema = new mongoose.Schema({
signupDate: { type: Date, default: Date.now },
username: { type: String, required: true, unique: true },
email: { type: String, required: true, unique: true },
password: { type: String, required: true },
name: { type: String }
});
userSchema.pre('save', function(next) {
if (!this.isModified('password')) {
return next();
}
bcrypt.hash(this.password, saltRounds, function(err, hash) {
if (err) {
console.log(err);
return next(err);
}
this.password = hash;
next();
}.bind(this));
});
userSchema.methods.checkPassword = function(password, cb) {
bcrypt.compare(password, this.password, function(err, result) {
if (err) {
console.log(err);
return next(err);
}
cb(result);
});
};
module.exports = mongoose.model('User', userSchema);
| var bcrypt = require('bcrypt');
var mongoose = require('mongoose');
const saltRounds = 10;
var userSchema = new mongoose.Schema({
signupDate: { type: Date, default: Date.now },
- username: { type: String, required: true },
+ username: { type: String, required: true, unique: true },
? ++++++++++++++
+ email: { type: String, required: true, unique: true },
password: { type: String, required: true },
- email: { type: String, required: true },
name: { type: String }
});
userSchema.pre('save', function(next) {
if (!this.isModified('password')) {
return next();
}
bcrypt.hash(this.password, saltRounds, function(err, hash) {
if (err) {
console.log(err);
return next(err);
}
this.password = hash;
next();
}.bind(this));
});
userSchema.methods.checkPassword = function(password, cb) {
bcrypt.compare(password, this.password, function(err, result) {
if (err) {
console.log(err);
return next(err);
}
cb(result);
});
};
module.exports = mongoose.model('User', userSchema); | 4 | 0.1 | 2 | 2 |
da1799861ba902f7137339f11a425ee5c326345c | README.markdown | README.markdown |
[](http://travis-ci.org/doctrine/mongodb)
The Doctrine MongoDB project is a library that provides a wrapper around the native PHP Mongo PECL extension to provide additional functionality.
## More resources:
* [Website](http://www.doctrine-project.org)
* [Documentation](http://docs.doctrine-project.org/projects/doctrine-mongodb-odm/en/latest/index.html)
* [Issue Tracker](http://www.doctrine-project.org/jira/browse/MODM)
* [Downloads](http://github.com/doctrine/mongodb/downloads) |
[](http://travis-ci.org/doctrine/mongodb)
The Doctrine MongoDB project is a library that provides a wrapper around the native PHP Mongo PECL extension to provide additional functionality.
## More resources:
* [Website](http://www.doctrine-project.org)
* [Documentation](http://docs.doctrine-project.org/projects/doctrine-mongodb-odm/en/latest/index.html)
* [Issue Tracker](http://www.doctrine-project.org/jira/browse/MODM)
* [Downloads](http://github.com/doctrine/mongodb/downloads)
| Use master branch for build status image | Use master branch for build status image | Markdown | mit | doctrine/mongodb,solocommand/mongodb,socialcdetest/test,alcaeus/mongodb,taohaolong/mongodb-1,researchgate/mongodb,talkspirit/mongodb,bjori/mongodb,lyft/mongodb | markdown | ## Code Before:
[](http://travis-ci.org/doctrine/mongodb)
The Doctrine MongoDB project is a library that provides a wrapper around the native PHP Mongo PECL extension to provide additional functionality.
## More resources:
* [Website](http://www.doctrine-project.org)
* [Documentation](http://docs.doctrine-project.org/projects/doctrine-mongodb-odm/en/latest/index.html)
* [Issue Tracker](http://www.doctrine-project.org/jira/browse/MODM)
* [Downloads](http://github.com/doctrine/mongodb/downloads)
## Instruction:
Use master branch for build status image
## Code After:
[](http://travis-ci.org/doctrine/mongodb)
The Doctrine MongoDB project is a library that provides a wrapper around the native PHP Mongo PECL extension to provide additional functionality.
## More resources:
* [Website](http://www.doctrine-project.org)
* [Documentation](http://docs.doctrine-project.org/projects/doctrine-mongodb-odm/en/latest/index.html)
* [Issue Tracker](http://www.doctrine-project.org/jira/browse/MODM)
* [Downloads](http://github.com/doctrine/mongodb/downloads)
|
- [](http://travis-ci.org/doctrine/mongodb)
+ [](http://travis-ci.org/doctrine/mongodb)
? ++++++++++++++
The Doctrine MongoDB project is a library that provides a wrapper around the native PHP Mongo PECL extension to provide additional functionality.
## More resources:
* [Website](http://www.doctrine-project.org)
* [Documentation](http://docs.doctrine-project.org/projects/doctrine-mongodb-odm/en/latest/index.html)
* [Issue Tracker](http://www.doctrine-project.org/jira/browse/MODM)
* [Downloads](http://github.com/doctrine/mongodb/downloads) | 2 | 0.181818 | 1 | 1 |
45bf3a24bac37feabeb38144d991b79e5b9d6c7a | examples/simple.rb | examples/simple.rb | require 'sidetiq'
# We're only loading this so we don't actually have to connect to redis.
require 'sidekiq/testing'
class ExampleWorker
include Sidekiq::Worker
include Sidetiq::Schedulable
def self.perform_at(time)
puts "Enqueued to run at #{time}"
end
# Run every 2 seconds
tiq { secondly(2) }
end
puts "Hit C-c to quit."
sleep 1000000
|
require 'sidekiq'
require 'sidetiq'
Sidekiq.options[:poll_interval] = 1
class MyWorker
include Sidekiq::Worker
include Sidetiq::Schedulable
tiq { secondly }
def perform(*args)
Sidekiq.logger.info "#perform"
end
end
| Add a quicker running example. | Add a quicker running example.
| Ruby | bsd-3-clause | endofunky/sidetiq,brennovich/sidetiq,slategroup/sidetiq,sfroehler/sidetiq,tomoyuki-yanagisawa/sidetiq-mod,danilogcarolino/sidetiq,rjocoleman/sidetiq,akhiln/sidetiq,flowerett/sidetiq,tobiassvn/sidetiq,akhiln/sidetiq,brennovich/sidetiq,kickserv/sidetiq,PaulMest/sidetiq,atpay/sidetiq,tomoyuki-yanagisawa/sidetiq-mod,slategroup/sidetiq,Altonymous/sidetiq,atpay/sidetiq,PaulMest/sidetiq,endofunky/sidetiq,stanchan/sidetiq,b1tw1se/sidetiq,stanchan/sidetiq,talentpad/sidetiq,vayu-technology/sidetiq,rjocoleman/sidetiq,b1tw1se/sidetiq,danilogcarolino/sidetiq,kickserv/sidetiq,sfroehler/sidetiq,grzlus/sidetiq,grzlus/sidetiq,flowerett/sidetiq,vayu-technology/sidetiq,talentpad/sidetiq,tobiassvn/sidetiq | ruby | ## Code Before:
require 'sidetiq'
# We're only loading this so we don't actually have to connect to redis.
require 'sidekiq/testing'
class ExampleWorker
include Sidekiq::Worker
include Sidetiq::Schedulable
def self.perform_at(time)
puts "Enqueued to run at #{time}"
end
# Run every 2 seconds
tiq { secondly(2) }
end
puts "Hit C-c to quit."
sleep 1000000
## Instruction:
Add a quicker running example.
## Code After:
require 'sidekiq'
require 'sidetiq'
Sidekiq.options[:poll_interval] = 1
class MyWorker
include Sidekiq::Worker
include Sidetiq::Schedulable
tiq { secondly }
def perform(*args)
Sidekiq.logger.info "#perform"
end
end
| +
+ require 'sidekiq'
require 'sidetiq'
+ Sidekiq.options[:poll_interval] = 1
- # We're only loading this so we don't actually have to connect to redis.
- require 'sidekiq/testing'
- class ExampleWorker
+ class MyWorker
include Sidekiq::Worker
include Sidetiq::Schedulable
- def self.perform_at(time)
- puts "Enqueued to run at #{time}"
+ tiq { secondly }
+
+ def perform(*args)
+ Sidekiq.logger.info "#perform"
end
-
- # Run every 2 seconds
- tiq { secondly(2) }
end
- puts "Hit C-c to quit."
- sleep 1000000 | 18 | 0.947368 | 8 | 10 |
d91c61a12f99f7888756a5b475c8f424bd75c87a | _posts/2015-05-18-openssl-nginx-server-linux.markdown | _posts/2015-05-18-openssl-nginx-server-linux.markdown | ---
layout: post
title: "OpenSSL - Os X - Linux on nginx Server Linux"
subtitle: "Run it over https"
date: 2015-05-18 14:00:44
author: "Lucas Gatsas"
header-img: "img/mirror.jpg"
---
<h2 class="section-heading">How To Create a SSL Certificate on nginx for Ubuntu 12.04</h2>
<h2 class="section-heading">Install Transport Layer Security - Secure Sockets Layer (SSL)</h2>
<strong> Set Up </strong>
<code> $ sudo apt-get install nginx </code>
<strong> Step One—Create a Directory for the Certificate </strong>
<code>sudo mkdir /etc/nginx/ssl</code>
We will perform the next few steps within the directory:
<code> cd /etc/nginx/ssl </code>
<strong>Step Two—Create the Server Key and Certificate Signing Request</strong>
Link: <a href="https://github.com/openssl/openssl">OpenSSL</a> Link: <a href="http://www.openssl.org/source/">OpenSSL Website</a>
<br>
<blockquote>
“Zeit für etwas zu haben ist das Talent Dinge nach ihrer Wichtigkeit ordnen zu können”
</blockquote>
| ---
layout: post
title: "OpenSSL - Os X - Linux on nginx Server Linux"
subtitle: "Run it over https"
date: 2015-05-18 14:00:44
author: "Lucas Gatsas"
header-img: "img/mirror.jpg"
---
<h2 class="section-heading">How To Create a SSL Certificate on nginx for Ubuntu 12.04</h2>
<h2 class="section-heading">Install Transport Layer Security - Secure Sockets Layer (SSL)</h2>
<strong> Set Up </strong>
<code> $ sudo apt-get install nginx </code>
<strong> Step One—Create a Directory for the Certificate </strong>
<code>sudo mkdir /etc/nginx/ssl</code>
We will perform the next few steps within the directory:
<code> cd /etc/nginx/ssl </code>
<strong>Step Two—Create the Server Key and Certificate Signing Request</strong>
<code> sudo openssl genrsa -des3 -out server.key 2048 </code>
Link: <a href="https://github.com/openssl/openssl">OpenSSL</a> Link: <a href="http://www.openssl.org/source/">OpenSSL Website</a>
<br>
<blockquote>
“Zeit für etwas zu haben ist das Talent Dinge nach ihrer Wichtigkeit ordnen zu können”
</blockquote>
| Update /// sudo openssl genrsa -des3 -out server.key 2048 | Update /// sudo openssl genrsa -des3 -out server.key 2048
| Markdown | mit | SpaceG/spaceg.github.io,SpaceG/spaceg.github.io | markdown | ## Code Before:
---
layout: post
title: "OpenSSL - Os X - Linux on nginx Server Linux"
subtitle: "Run it over https"
date: 2015-05-18 14:00:44
author: "Lucas Gatsas"
header-img: "img/mirror.jpg"
---
<h2 class="section-heading">How To Create a SSL Certificate on nginx for Ubuntu 12.04</h2>
<h2 class="section-heading">Install Transport Layer Security - Secure Sockets Layer (SSL)</h2>
<strong> Set Up </strong>
<code> $ sudo apt-get install nginx </code>
<strong> Step One—Create a Directory for the Certificate </strong>
<code>sudo mkdir /etc/nginx/ssl</code>
We will perform the next few steps within the directory:
<code> cd /etc/nginx/ssl </code>
<strong>Step Two—Create the Server Key and Certificate Signing Request</strong>
Link: <a href="https://github.com/openssl/openssl">OpenSSL</a> Link: <a href="http://www.openssl.org/source/">OpenSSL Website</a>
<br>
<blockquote>
“Zeit für etwas zu haben ist das Talent Dinge nach ihrer Wichtigkeit ordnen zu können”
</blockquote>
## Instruction:
Update /// sudo openssl genrsa -des3 -out server.key 2048
## Code After:
---
layout: post
title: "OpenSSL - Os X - Linux on nginx Server Linux"
subtitle: "Run it over https"
date: 2015-05-18 14:00:44
author: "Lucas Gatsas"
header-img: "img/mirror.jpg"
---
<h2 class="section-heading">How To Create a SSL Certificate on nginx for Ubuntu 12.04</h2>
<h2 class="section-heading">Install Transport Layer Security - Secure Sockets Layer (SSL)</h2>
<strong> Set Up </strong>
<code> $ sudo apt-get install nginx </code>
<strong> Step One—Create a Directory for the Certificate </strong>
<code>sudo mkdir /etc/nginx/ssl</code>
We will perform the next few steps within the directory:
<code> cd /etc/nginx/ssl </code>
<strong>Step Two—Create the Server Key and Certificate Signing Request</strong>
<code> sudo openssl genrsa -des3 -out server.key 2048 </code>
Link: <a href="https://github.com/openssl/openssl">OpenSSL</a> Link: <a href="http://www.openssl.org/source/">OpenSSL Website</a>
<br>
<blockquote>
“Zeit für etwas zu haben ist das Talent Dinge nach ihrer Wichtigkeit ordnen zu können”
</blockquote>
| ---
layout: post
title: "OpenSSL - Os X - Linux on nginx Server Linux"
subtitle: "Run it over https"
date: 2015-05-18 14:00:44
author: "Lucas Gatsas"
header-img: "img/mirror.jpg"
---
<h2 class="section-heading">How To Create a SSL Certificate on nginx for Ubuntu 12.04</h2>
<h2 class="section-heading">Install Transport Layer Security - Secure Sockets Layer (SSL)</h2>
<strong> Set Up </strong>
<code> $ sudo apt-get install nginx </code>
<strong> Step One—Create a Directory for the Certificate </strong>
<code>sudo mkdir /etc/nginx/ssl</code>
We will perform the next few steps within the directory:
<code> cd /etc/nginx/ssl </code>
<strong>Step Two—Create the Server Key and Certificate Signing Request</strong>
+ <code> sudo openssl genrsa -des3 -out server.key 2048 </code>
+
+
Link: <a href="https://github.com/openssl/openssl">OpenSSL</a> Link: <a href="http://www.openssl.org/source/">OpenSSL Website</a>
<br>
<blockquote>
“Zeit für etwas zu haben ist das Talent Dinge nach ihrer Wichtigkeit ordnen zu können”
</blockquote>
| 3 | 0.069767 | 3 | 0 |
4b96477d686506be6c00401ed624f686b3856cc0 | run_test.sh | run_test.sh | set -e
SCRIPT_DIR="$(cd "$(dirname "$0")" && pwd)"
if python -m pytest --fixtures | grep 'xdist\.plugin' > /dev/null ; then
PYTEST_OPTS='-n auto'
else
echo "pytest-xdist not detected, tests won't be run in parellel"
fi
for PYTHON in python2 python3 ; do
echo ------------------------------------------------
echo Testing with $PYTHON...
if [[ $PYTEST_OPTS ]] ; then
$PYTHON "$SCRIPT_DIR/setup.py" test --addopts "$PYTEST_OPTS"
else
$PYTHON "$SCRIPT_DIR/setup.py" test
fi
done
| set -e
SCRIPT_DIR="$(cd "$(dirname "$0")" && pwd)"
for PYTHON in python2 python3 ; do
echo ------------------------------------------------
echo Testing with $PYTHON...
if $PYTHON -m pytest --fixtures | grep 'xdist\.plugin' > /dev/null ; then
PYTEST_OPTS='-n auto'
else
echo "pytest-xdist not detected, tests won't be run in parellel"
fi
if [[ $PYTEST_OPTS ]] ; then
$PYTHON "$SCRIPT_DIR/setup.py" test --addopts "$PYTEST_OPTS"
else
$PYTHON "$SCRIPT_DIR/setup.py" test
fi
done
| Check pytest-xdist validity on each Python version | Check pytest-xdist validity on each Python version
| Shell | mit | tiagoshibata/exrsplit,tiagoshibata/exrsplit | shell | ## Code Before:
set -e
SCRIPT_DIR="$(cd "$(dirname "$0")" && pwd)"
if python -m pytest --fixtures | grep 'xdist\.plugin' > /dev/null ; then
PYTEST_OPTS='-n auto'
else
echo "pytest-xdist not detected, tests won't be run in parellel"
fi
for PYTHON in python2 python3 ; do
echo ------------------------------------------------
echo Testing with $PYTHON...
if [[ $PYTEST_OPTS ]] ; then
$PYTHON "$SCRIPT_DIR/setup.py" test --addopts "$PYTEST_OPTS"
else
$PYTHON "$SCRIPT_DIR/setup.py" test
fi
done
## Instruction:
Check pytest-xdist validity on each Python version
## Code After:
set -e
SCRIPT_DIR="$(cd "$(dirname "$0")" && pwd)"
for PYTHON in python2 python3 ; do
echo ------------------------------------------------
echo Testing with $PYTHON...
if $PYTHON -m pytest --fixtures | grep 'xdist\.plugin' > /dev/null ; then
PYTEST_OPTS='-n auto'
else
echo "pytest-xdist not detected, tests won't be run in parellel"
fi
if [[ $PYTEST_OPTS ]] ; then
$PYTHON "$SCRIPT_DIR/setup.py" test --addopts "$PYTEST_OPTS"
else
$PYTHON "$SCRIPT_DIR/setup.py" test
fi
done
| set -e
SCRIPT_DIR="$(cd "$(dirname "$0")" && pwd)"
- if python -m pytest --fixtures | grep 'xdist\.plugin' > /dev/null ; then
- PYTEST_OPTS='-n auto'
- else
- echo "pytest-xdist not detected, tests won't be run in parellel"
- fi
for PYTHON in python2 python3 ; do
echo ------------------------------------------------
echo Testing with $PYTHON...
+
+ if $PYTHON -m pytest --fixtures | grep 'xdist\.plugin' > /dev/null ; then
+ PYTEST_OPTS='-n auto'
+ else
+ echo "pytest-xdist not detected, tests won't be run in parellel"
+ fi
+
if [[ $PYTEST_OPTS ]] ; then
$PYTHON "$SCRIPT_DIR/setup.py" test --addopts "$PYTEST_OPTS"
else
$PYTHON "$SCRIPT_DIR/setup.py" test
fi
done | 12 | 0.666667 | 7 | 5 |
587dc3f78c3b100ed84476849de7fc9d59a2a865 | src/app/constants/index.js | src/app/constants/index.js | // Workflow State
export const WORKFLOW_STATE_RUNNING = 'running';
export const WORKFLOW_STATE_PAUSED = 'paused';
export const WORKFLOW_STATE_IDLE = 'idle';
| // Workflow State
export const WORKFLOW_STATE_RUNNING = 'running';
export const WORKFLOW_STATE_PAUSED = 'paused';
export const WORKFLOW_STATE_IDLE = 'idle';
// Macro action
export const MACRO_ACTION_START = 'start';
export const MACRO_ACTION_STOP = 'stop';
| Define constants for macro actions | Define constants for macro actions
| JavaScript | mit | cheton/cnc.js,cheton/cnc,cncjs/cncjs,cheton/cnc.js,cheton/piduino-grbl,cheton/cnc.js,cncjs/cncjs,cncjs/cncjs,cheton/piduino-grbl,cheton/piduino-grbl,cheton/cnc,cheton/cnc | javascript | ## Code Before:
// Workflow State
export const WORKFLOW_STATE_RUNNING = 'running';
export const WORKFLOW_STATE_PAUSED = 'paused';
export const WORKFLOW_STATE_IDLE = 'idle';
## Instruction:
Define constants for macro actions
## Code After:
// Workflow State
export const WORKFLOW_STATE_RUNNING = 'running';
export const WORKFLOW_STATE_PAUSED = 'paused';
export const WORKFLOW_STATE_IDLE = 'idle';
// Macro action
export const MACRO_ACTION_START = 'start';
export const MACRO_ACTION_STOP = 'stop';
| // Workflow State
export const WORKFLOW_STATE_RUNNING = 'running';
export const WORKFLOW_STATE_PAUSED = 'paused';
export const WORKFLOW_STATE_IDLE = 'idle';
+
+ // Macro action
+ export const MACRO_ACTION_START = 'start';
+ export const MACRO_ACTION_STOP = 'stop'; | 4 | 1 | 4 | 0 |
80e5ef61f3a43a41d18df7540754545fa040d350 | docs/uploading-images.md | docs/uploading-images.md | Uploading Images
================
Images should be uploaded as a Zip file of JPG images. You'll need to upload any new images that you want to add before you add any metadata. Depending on how many images you add they can take a while to process (hours initially, to be fully indexed and searchable it could take many more hours or days).
The images shouldn't be smaller than about 300 pixels on any side. Ideally they would be much larger. Having them be at least 1000 pixels on each side would be good to start with. It's important to note that the names of the images should match the file names provided in the metadata file.
If you ever upload an image that was previously uploaded then the old image will be replaced with the new one.
If you ever upload any new images you'll also need to upload a new metadata file in order for the new images to be displayed on the site. (Metadata records that have no images associated with them are never added to the database.) | Uploading Images
================
Images should be uploaded as a Zip file of JPG images. You'll need to upload any new images that you want to add before you add any metadata. Depending on how many images you add they can take a while to process (hours initially, to be fully indexed and searchable it could take many more hours or days).
The images shouldn't be smaller than about 300 pixels on any side. Ideally they would be much larger. Having them be at least 1000 pixels on each side would be good to start with. It's important to note that the names of the images should match the file names provided in the metadata.
If you upload an image that was previously uploaded then the old image will be replaced with the new one.
If you ever upload any new images you'll also need to upload a new metadata file in order for the new images to be displayed on the site. (Metadata records that have no images associated with them are never added to the database.)
Once you've uploaded some images, and the images have been fully processed, you will be presented with the number of images successfully uploaded, or that failed to upload. If you click the filename of the Zip file that you uploaded you'll be able to see additional information about what went wrong with the image uploads (if anything).
Common errors and warnings include:
* Having an improperly-formatted JPG image that is unable to be read or displayed. This can be fixed by providing a correct JPG file (verify that it's able to open on your computer).
* Having multiple images that have the same file name. Only one image will be used, the others will be ignored. You'll need to ensure that no duplicate files are provided.
* The image is smaller than 150 pixels on a side (this prevents it from being able to be indexed by the similarity search). Providing a larger image will resolve this warning. | Expand the uploading images docs a bit. | Expand the uploading images docs a bit.
| Markdown | mit | jeresig/pharos-images | markdown | ## Code Before:
Uploading Images
================
Images should be uploaded as a Zip file of JPG images. You'll need to upload any new images that you want to add before you add any metadata. Depending on how many images you add they can take a while to process (hours initially, to be fully indexed and searchable it could take many more hours or days).
The images shouldn't be smaller than about 300 pixels on any side. Ideally they would be much larger. Having them be at least 1000 pixels on each side would be good to start with. It's important to note that the names of the images should match the file names provided in the metadata file.
If you ever upload an image that was previously uploaded then the old image will be replaced with the new one.
If you ever upload any new images you'll also need to upload a new metadata file in order for the new images to be displayed on the site. (Metadata records that have no images associated with them are never added to the database.)
## Instruction:
Expand the uploading images docs a bit.
## Code After:
Uploading Images
================
Images should be uploaded as a Zip file of JPG images. You'll need to upload any new images that you want to add before you add any metadata. Depending on how many images you add they can take a while to process (hours initially, to be fully indexed and searchable it could take many more hours or days).
The images shouldn't be smaller than about 300 pixels on any side. Ideally they would be much larger. Having them be at least 1000 pixels on each side would be good to start with. It's important to note that the names of the images should match the file names provided in the metadata.
If you upload an image that was previously uploaded then the old image will be replaced with the new one.
If you ever upload any new images you'll also need to upload a new metadata file in order for the new images to be displayed on the site. (Metadata records that have no images associated with them are never added to the database.)
Once you've uploaded some images, and the images have been fully processed, you will be presented with the number of images successfully uploaded, or that failed to upload. If you click the filename of the Zip file that you uploaded you'll be able to see additional information about what went wrong with the image uploads (if anything).
Common errors and warnings include:
* Having an improperly-formatted JPG image that is unable to be read or displayed. This can be fixed by providing a correct JPG file (verify that it's able to open on your computer).
* Having multiple images that have the same file name. Only one image will be used, the others will be ignored. You'll need to ensure that no duplicate files are provided.
* The image is smaller than 150 pixels on a side (this prevents it from being able to be indexed by the similarity search). Providing a larger image will resolve this warning. | Uploading Images
================
Images should be uploaded as a Zip file of JPG images. You'll need to upload any new images that you want to add before you add any metadata. Depending on how many images you add they can take a while to process (hours initially, to be fully indexed and searchable it could take many more hours or days).
- The images shouldn't be smaller than about 300 pixels on any side. Ideally they would be much larger. Having them be at least 1000 pixels on each side would be good to start with. It's important to note that the names of the images should match the file names provided in the metadata file.
? ^^^^^^
+ The images shouldn't be smaller than about 300 pixels on any side. Ideally they would be much larger. Having them be at least 1000 pixels on each side would be good to start with. It's important to note that the names of the images should match the file names provided in the metadata.
? ^
- If you ever upload an image that was previously uploaded then the old image will be replaced with the new one.
? -----
+ If you upload an image that was previously uploaded then the old image will be replaced with the new one.
If you ever upload any new images you'll also need to upload a new metadata file in order for the new images to be displayed on the site. (Metadata records that have no images associated with them are never added to the database.)
+
+ Once you've uploaded some images, and the images have been fully processed, you will be presented with the number of images successfully uploaded, or that failed to upload. If you click the filename of the Zip file that you uploaded you'll be able to see additional information about what went wrong with the image uploads (if anything).
+
+ Common errors and warnings include:
+
+ * Having an improperly-formatted JPG image that is unable to be read or displayed. This can be fixed by providing a correct JPG file (verify that it's able to open on your computer).
+ * Having multiple images that have the same file name. Only one image will be used, the others will be ignored. You'll need to ensure that no duplicate files are provided.
+ * The image is smaller than 150 pixels on a side (this prevents it from being able to be indexed by the similarity search). Providing a larger image will resolve this warning. | 12 | 1.2 | 10 | 2 |
f3f1617333375267ad84555440a237ee5c5057de | README.md | README.md | This is my [cider](https://github.com/msanders/cider) config. It downloads and put in place every file and folder, so to format should be painless from now :)
Installation
============
```bash
curl -Ls https://raw.githubusercontent.com/patoroco/dotfiles/master/setup.sh | sh
```
Notes
=====
iTerm
-----
- [Jump to beginning / end of line](http://stackoverflow.com/questions/6205157/iterm2-how-to-get-jump-to-beginning-end-of-line-in-bash-shell)
- [Jump
words](https://coderwall.com/p/h6yfda/use-and-to-jump-forwards-backwards-words-in-iterm-2-on-os-x).
Acknowledgements
================
Thanks to [gtramontina](https://github.com/gtramontina/dotfiles) I discovered
**cider**, and I caught some of their config files.
I steal some config files from [igalarza's dotfiles](https://github.com/igalarzab/dotfiles) too.
| This is my [cider](https://github.com/msanders/cider) config. It downloads and put in place every file and folder, so to format should be painless from now :)
Installation
============
```bash
curl -Ls https://raw.githubusercontent.com/patoroco/dotfiles/master/setup.sh | sh
```
Notes
=====
iTerm
-----
### Profile keys
- `Cmd + ->`: Hex Code 0x05
- `Cmd + <-`: Hex Code 0x01
- `Alt + ->`: Esc f
- `Alt + <-`: Esc b
- `Alt + Backspace`: Hex Code 0x17
- `Alt + Supr` (`fn + backspace`): Esc d
####Source
- [Jump to beginning / end of line](http://stackoverflow.com/questions/6205157/iterm2-how-to-get-jump-to-beginning-end-of-line-in-bash-shell)
- [Jump
words](https://coderwall.com/p/h6yfda/use-and-to-jump-forwards-backwards-words-in-iterm-2-on-os-x)
- [stackoverflow](http://stackoverflow.com/questions/12335787/with-iterm2-on-mac-how-to-delete-forward-a-word-from-cursor-on-command-line)
Acknowledgements
================
Thanks to [gtramontina](https://github.com/gtramontina/dotfiles) I discovered
**cider**, and I caught some of their config files.
I steal some config files from [igalarza's dotfiles](https://github.com/igalarzab/dotfiles) too.
| Add more shortcuts for iTerm | Add more shortcuts for iTerm
| Markdown | mit | patoroco/dotfiles,patoroco/dotfiles,patoroco/dotfiles,patoroco/dotfiles | markdown | ## Code Before:
This is my [cider](https://github.com/msanders/cider) config. It downloads and put in place every file and folder, so to format should be painless from now :)
Installation
============
```bash
curl -Ls https://raw.githubusercontent.com/patoroco/dotfiles/master/setup.sh | sh
```
Notes
=====
iTerm
-----
- [Jump to beginning / end of line](http://stackoverflow.com/questions/6205157/iterm2-how-to-get-jump-to-beginning-end-of-line-in-bash-shell)
- [Jump
words](https://coderwall.com/p/h6yfda/use-and-to-jump-forwards-backwards-words-in-iterm-2-on-os-x).
Acknowledgements
================
Thanks to [gtramontina](https://github.com/gtramontina/dotfiles) I discovered
**cider**, and I caught some of their config files.
I steal some config files from [igalarza's dotfiles](https://github.com/igalarzab/dotfiles) too.
## Instruction:
Add more shortcuts for iTerm
## Code After:
This is my [cider](https://github.com/msanders/cider) config. It downloads and put in place every file and folder, so to format should be painless from now :)
Installation
============
```bash
curl -Ls https://raw.githubusercontent.com/patoroco/dotfiles/master/setup.sh | sh
```
Notes
=====
iTerm
-----
### Profile keys
- `Cmd + ->`: Hex Code 0x05
- `Cmd + <-`: Hex Code 0x01
- `Alt + ->`: Esc f
- `Alt + <-`: Esc b
- `Alt + Backspace`: Hex Code 0x17
- `Alt + Supr` (`fn + backspace`): Esc d
####Source
- [Jump to beginning / end of line](http://stackoverflow.com/questions/6205157/iterm2-how-to-get-jump-to-beginning-end-of-line-in-bash-shell)
- [Jump
words](https://coderwall.com/p/h6yfda/use-and-to-jump-forwards-backwards-words-in-iterm-2-on-os-x)
- [stackoverflow](http://stackoverflow.com/questions/12335787/with-iterm2-on-mac-how-to-delete-forward-a-word-from-cursor-on-command-line)
Acknowledgements
================
Thanks to [gtramontina](https://github.com/gtramontina/dotfiles) I discovered
**cider**, and I caught some of their config files.
I steal some config files from [igalarza's dotfiles](https://github.com/igalarzab/dotfiles) too.
| This is my [cider](https://github.com/msanders/cider) config. It downloads and put in place every file and folder, so to format should be painless from now :)
Installation
============
```bash
curl -Ls https://raw.githubusercontent.com/patoroco/dotfiles/master/setup.sh | sh
```
Notes
=====
iTerm
-----
+ ### Profile keys
+ - `Cmd + ->`: Hex Code 0x05
+ - `Cmd + <-`: Hex Code 0x01
+ - `Alt + ->`: Esc f
+ - `Alt + <-`: Esc b
+ - `Alt + Backspace`: Hex Code 0x17
+ - `Alt + Supr` (`fn + backspace`): Esc d
+ ####Source
- [Jump to beginning / end of line](http://stackoverflow.com/questions/6205157/iterm2-how-to-get-jump-to-beginning-end-of-line-in-bash-shell)
- [Jump
- words](https://coderwall.com/p/h6yfda/use-and-to-jump-forwards-backwards-words-in-iterm-2-on-os-x).
? -
+ words](https://coderwall.com/p/h6yfda/use-and-to-jump-forwards-backwards-words-in-iterm-2-on-os-x)
+ - [stackoverflow](http://stackoverflow.com/questions/12335787/with-iterm2-on-mac-how-to-delete-forward-a-word-from-cursor-on-command-line)
Acknowledgements
================
Thanks to [gtramontina](https://github.com/gtramontina/dotfiles) I discovered
**cider**, and I caught some of their config files.
I steal some config files from [igalarza's dotfiles](https://github.com/igalarzab/dotfiles) too. | 11 | 0.37931 | 10 | 1 |
78a0e69ae0fc572156674074a23d46826bd2d923 | app-backend/migrations/src_migrations/main/scala/126.scala | app-backend/migrations/src_migrations/main/scala/126.scala | import slick.jdbc.PostgresProfile.api._
import com.liyaos.forklift.slick.SqlMigration
object M126 {
RFMigrations.migrations = RFMigrations.migrations :+ SqlMigration(126)(List(
sqlu"""
DELETE FROM access_control_rules WHERE subject_type = 'ALL' AND object_type = 'SCENE' AND action_type = 'VIEW';
"""
))
}
| import slick.jdbc.PostgresProfile.api._
import com.liyaos.forklift.slick.SqlMigration
object M126 {
RFMigrations.migrations = RFMigrations.migrations :+ SqlMigration(126)(List(
sqlu"""
DELETE FROM access_control_rules WHERE subject_type = 'ALL' AND object_type = 'SCENE' AND
(action_type = 'VIEW' OR action_type = 'DOWNLOAD');
"""
))
}
| Update access control rule migration | Update access control rule migration
| Scala | apache-2.0 | aaronxsu/raster-foundry,aaronxsu/raster-foundry,aaronxsu/raster-foundry,azavea/raster-foundry,azavea/raster-foundry,azavea/raster-foundry,azavea/raster-foundry,aaronxsu/raster-foundry,raster-foundry/raster-foundry,azavea/raster-foundry,raster-foundry/raster-foundry,raster-foundry/raster-foundry | scala | ## Code Before:
import slick.jdbc.PostgresProfile.api._
import com.liyaos.forklift.slick.SqlMigration
object M126 {
RFMigrations.migrations = RFMigrations.migrations :+ SqlMigration(126)(List(
sqlu"""
DELETE FROM access_control_rules WHERE subject_type = 'ALL' AND object_type = 'SCENE' AND action_type = 'VIEW';
"""
))
}
## Instruction:
Update access control rule migration
## Code After:
import slick.jdbc.PostgresProfile.api._
import com.liyaos.forklift.slick.SqlMigration
object M126 {
RFMigrations.migrations = RFMigrations.migrations :+ SqlMigration(126)(List(
sqlu"""
DELETE FROM access_control_rules WHERE subject_type = 'ALL' AND object_type = 'SCENE' AND
(action_type = 'VIEW' OR action_type = 'DOWNLOAD');
"""
))
}
| import slick.jdbc.PostgresProfile.api._
import com.liyaos.forklift.slick.SqlMigration
object M126 {
RFMigrations.migrations = RFMigrations.migrations :+ SqlMigration(126)(List(
sqlu"""
- DELETE FROM access_control_rules WHERE subject_type = 'ALL' AND object_type = 'SCENE' AND action_type = 'VIEW';
? ----------------------
+ DELETE FROM access_control_rules WHERE subject_type = 'ALL' AND object_type = 'SCENE' AND
+ (action_type = 'VIEW' OR action_type = 'DOWNLOAD');
"""
))
} | 3 | 0.3 | 2 | 1 |
f99ec25488a71b3e0112b6f353bda76e96fc7c60 | tox.ini | tox.ini | [tox]
envlist = py26,py27,py33
[testenv:py26]
deps =
unittest2
discover
coverage
commands =
coverage erase
coverage run --source=icalendar --omit=*tests* {envbindir}/discover icalendar
coverage report --omit=*tests*
coverage html --omit=*tests*
[testenv]
deps =
discover
coverage
commands =
coverage erase
coverage run --source=icalendar --omit=*tests* {envbindir}/discover icalendar
coverage report --omit=*tests*
coverage html --omit=*tests*
| [tox]
envlist = py26,py27,py33
[testenv:py26]
deps =
unittest2
pytest
coverage
[testenv]
deps =
pytest
coverage
commands =
coverage run --source=src/icalendar --module pytest src/icalendar
coverage report
coverage html
| Use pytest instead of discover | Use pytest instead of discover
| INI | bsd-2-clause | untitaker/icalendar,geier/icalendar,nylas/icalendar | ini | ## Code Before:
[tox]
envlist = py26,py27,py33
[testenv:py26]
deps =
unittest2
discover
coverage
commands =
coverage erase
coverage run --source=icalendar --omit=*tests* {envbindir}/discover icalendar
coverage report --omit=*tests*
coverage html --omit=*tests*
[testenv]
deps =
discover
coverage
commands =
coverage erase
coverage run --source=icalendar --omit=*tests* {envbindir}/discover icalendar
coverage report --omit=*tests*
coverage html --omit=*tests*
## Instruction:
Use pytest instead of discover
## Code After:
[tox]
envlist = py26,py27,py33
[testenv:py26]
deps =
unittest2
pytest
coverage
[testenv]
deps =
pytest
coverage
commands =
coverage run --source=src/icalendar --module pytest src/icalendar
coverage report
coverage html
| [tox]
envlist = py26,py27,py33
[testenv:py26]
deps =
unittest2
- discover
+ pytest
coverage
- commands =
- coverage erase
- coverage run --source=icalendar --omit=*tests* {envbindir}/discover icalendar
- coverage report --omit=*tests*
- coverage html --omit=*tests*
[testenv]
deps =
- discover
+ pytest
coverage
commands =
+ coverage run --source=src/icalendar --module pytest src/icalendar
- coverage erase
? ^^^
+ coverage report
? + ++ ^
+ coverage html
- coverage run --source=icalendar --omit=*tests* {envbindir}/discover icalendar
- coverage report --omit=*tests*
- coverage html --omit=*tests* | 16 | 0.695652 | 5 | 11 |
0ea6a328654399c5b8f21508913f362fbcb99e06 | tests/tests.js | tests/tests.js | var path = require('path');
var root = path.dirname(__filename);
require.paths.unshift(path.join(root, '../build/default'));
require.paths.unshift(path.join(root, '../lib'));
var sys = require('sys');
var Buffer = require('buffer').Buffer;
var archive = require('archive');
var ar = new archive.ArchiveReader();
var buf = new Buffer(8000);
ar.addListener('ready', function() {
sys.log('In ready function....');
ar.nextEntry();
});
ar.addListener('entry', function(entry) {
sys.log('path: '+ entry.getPath());
sys.log(' size: '+ entry.getSize());
sys.log(' mtime: '+ entry.getMtime());
function reader(length, err) {
if (!err) {
if (length === 0) {
entry.emit('end');
}
else {
var b = buf.slice(0, length);
entry.emit('data', b);
entry.read(buf, reader);
}
}
else {
sys.log(err);
entry.emit('end');
}
}
entry.read(buf, reader);
entry.addListener('data', function(data) {
sys.log("got data of length: "+ data.length);
});
entry.addListener('end', function() {
process.nextTick(function() {
ar.nextEntry();
});
});
});
ar.openFile("nofile.tar.gz", function(err){
if (err) {
sys.log(err);
}
});
| var path = require('path');
var root = path.dirname(__filename);
require.paths.unshift(path.join(root, '../build/default'));
require.paths.unshift(path.join(root, '../lib'));
var sys = require('sys');
var Buffer = require('buffer').Buffer;
var archive = require('archive');
var ar = new archive.ArchiveReader();
var buf = new Buffer(1600);
ar.addListener('ready', function() {
sys.log('In ready function....');
ar.nextEntry();
});
ar.addListener('entry', function(entry) {
sys.log('path: '+ entry.getPath());
sys.log(' size: '+ entry.getSize());
sys.log(' mtime: '+ entry.getMtime());
function reader(length, err) {
if (!err) {
if (length === 0) {
entry.emit('end');
}
else {
var b = buf.slice(0, length);
entry.emit('data', b);
entry.read(buf, reader);
}
}
else {
sys.log(err);
entry.emit('end');
}
}
entry.read(buf, reader);
entry.addListener('data', function(data) {
sys.log("got data of length: "+ data.length);
});
entry.addListener('end', function() {
process.nextTick(function() {
ar.nextEntry();
});
});
});
ar.openFile(path.join(root, "nofile.tar.gz"), function(err){
if (err) {
sys.log(err);
}
});
| Use relative path for the test tarfile | Use relative path for the test tarfile
| JavaScript | apache-2.0 | pquerna/node-archive,pquerna/node-archive | javascript | ## Code Before:
var path = require('path');
var root = path.dirname(__filename);
require.paths.unshift(path.join(root, '../build/default'));
require.paths.unshift(path.join(root, '../lib'));
var sys = require('sys');
var Buffer = require('buffer').Buffer;
var archive = require('archive');
var ar = new archive.ArchiveReader();
var buf = new Buffer(8000);
ar.addListener('ready', function() {
sys.log('In ready function....');
ar.nextEntry();
});
ar.addListener('entry', function(entry) {
sys.log('path: '+ entry.getPath());
sys.log(' size: '+ entry.getSize());
sys.log(' mtime: '+ entry.getMtime());
function reader(length, err) {
if (!err) {
if (length === 0) {
entry.emit('end');
}
else {
var b = buf.slice(0, length);
entry.emit('data', b);
entry.read(buf, reader);
}
}
else {
sys.log(err);
entry.emit('end');
}
}
entry.read(buf, reader);
entry.addListener('data', function(data) {
sys.log("got data of length: "+ data.length);
});
entry.addListener('end', function() {
process.nextTick(function() {
ar.nextEntry();
});
});
});
ar.openFile("nofile.tar.gz", function(err){
if (err) {
sys.log(err);
}
});
## Instruction:
Use relative path for the test tarfile
## Code After:
var path = require('path');
var root = path.dirname(__filename);
require.paths.unshift(path.join(root, '../build/default'));
require.paths.unshift(path.join(root, '../lib'));
var sys = require('sys');
var Buffer = require('buffer').Buffer;
var archive = require('archive');
var ar = new archive.ArchiveReader();
var buf = new Buffer(1600);
ar.addListener('ready', function() {
sys.log('In ready function....');
ar.nextEntry();
});
ar.addListener('entry', function(entry) {
sys.log('path: '+ entry.getPath());
sys.log(' size: '+ entry.getSize());
sys.log(' mtime: '+ entry.getMtime());
function reader(length, err) {
if (!err) {
if (length === 0) {
entry.emit('end');
}
else {
var b = buf.slice(0, length);
entry.emit('data', b);
entry.read(buf, reader);
}
}
else {
sys.log(err);
entry.emit('end');
}
}
entry.read(buf, reader);
entry.addListener('data', function(data) {
sys.log("got data of length: "+ data.length);
});
entry.addListener('end', function() {
process.nextTick(function() {
ar.nextEntry();
});
});
});
ar.openFile(path.join(root, "nofile.tar.gz"), function(err){
if (err) {
sys.log(err);
}
});
| var path = require('path');
var root = path.dirname(__filename);
require.paths.unshift(path.join(root, '../build/default'));
require.paths.unshift(path.join(root, '../lib'));
var sys = require('sys');
var Buffer = require('buffer').Buffer;
var archive = require('archive');
var ar = new archive.ArchiveReader();
- var buf = new Buffer(8000);
? ^^
+ var buf = new Buffer(1600);
? ^^
ar.addListener('ready', function() {
sys.log('In ready function....');
ar.nextEntry();
});
ar.addListener('entry', function(entry) {
sys.log('path: '+ entry.getPath());
sys.log(' size: '+ entry.getSize());
sys.log(' mtime: '+ entry.getMtime());
function reader(length, err) {
if (!err) {
if (length === 0) {
entry.emit('end');
}
else {
var b = buf.slice(0, length);
entry.emit('data', b);
entry.read(buf, reader);
}
}
else {
sys.log(err);
entry.emit('end');
}
}
entry.read(buf, reader);
entry.addListener('data', function(data) {
sys.log("got data of length: "+ data.length);
});
entry.addListener('end', function() {
process.nextTick(function() {
ar.nextEntry();
});
});
});
- ar.openFile("nofile.tar.gz", function(err){
+ ar.openFile(path.join(root, "nofile.tar.gz"), function(err){
? ++++++++++++++++ +
if (err) {
sys.log(err);
}
});
| 4 | 0.068966 | 2 | 2 |
6493d76919c737e5f6b655f51115cc91523dda92 | Document/Event.php | Document/Event.php | <?php
namespace Itkg\DelayEventBundle\Document;
use Gedmo\Timestampable\Traits\TimestampableDocument;
use Itkg\DelayEventBundle\Model\Event as BaseEvent;
use Doctrine\ODM\MongoDB\Mapping\Annotations as ODM;
/**
* Class Event
*
* @ODM\Document(
* repositoryClass="Itkg\DelayEventBundle\Repository\EventRepository",
* collection="itkg_delay_event"
* )
* @ODM\DiscriminatorField(fieldName="type")
* @ODM\InheritanceType("SINGLE_COLLECTION")
*/
class Event extends BaseEvent
{
use TimestampableDocument;
/**
* @var string
*
* @ODM\id
*/
protected $id;
/**
* @var string
*
* @ODM\String
*/
protected $originalName;
/**
* @var bool
*
* @ODM\Boolean
*/
protected $async = true;
/**
* @var bool
*
* @ODM\Boolean
*/
protected $failed = false;
/**
* @var int
*
* @ODM\Int
*/
protected $tryCount = 0;
/**
* @var string
*
* @ODM\String
*/
protected $groupFieldIdentifier = self::DEFAULT_GROUP_IDENTIFIER;
}
| <?php
namespace Itkg\DelayEventBundle\Document;
use Gedmo\Timestampable\Traits\TimestampableDocument;
use Itkg\DelayEventBundle\Model\Event as BaseEvent;
use Doctrine\ODM\MongoDB\Mapping\Annotations as ODM;
/**
* Class Event
*
* @ODM\Document(
* repositoryClass="Itkg\DelayEventBundle\Repository\EventRepository",
* collection="itkg_delay_event"
* )
* @ODM\DiscriminatorField(fieldName="type")
* @ODM\InheritanceType("SINGLE_COLLECTION")
*
* @ODM\Indexes({
* @ODM\Index(keys={"failed"="asc", "originalName"="asc", "groupFieldIdentifier"="asc", "createdAt":"asc"}, name="failed_name_group_date")
* })
*/
class Event extends BaseEvent
{
use TimestampableDocument;
/**
* @var string
*
* @ODM\id
*/
protected $id;
/**
* @var string
*
* @ODM\String
*/
protected $originalName;
/**
* @var bool
*
* @ODM\Boolean
*/
protected $async = true;
/**
* @var bool
*
* @ODM\Boolean
*/
protected $failed = false;
/**
* @var int
*
* @ODM\Int
*/
protected $tryCount = 0;
/**
* @var string
*
* @ODM\String
*/
protected $groupFieldIdentifier = self::DEFAULT_GROUP_IDENTIFIER;
}
| Add index on delay events collection | Add index on delay events collection
| PHP | mit | itkg/delay-event-bundle,itkg/delay-event-bundle | php | ## Code Before:
<?php
namespace Itkg\DelayEventBundle\Document;
use Gedmo\Timestampable\Traits\TimestampableDocument;
use Itkg\DelayEventBundle\Model\Event as BaseEvent;
use Doctrine\ODM\MongoDB\Mapping\Annotations as ODM;
/**
* Class Event
*
* @ODM\Document(
* repositoryClass="Itkg\DelayEventBundle\Repository\EventRepository",
* collection="itkg_delay_event"
* )
* @ODM\DiscriminatorField(fieldName="type")
* @ODM\InheritanceType("SINGLE_COLLECTION")
*/
class Event extends BaseEvent
{
use TimestampableDocument;
/**
* @var string
*
* @ODM\id
*/
protected $id;
/**
* @var string
*
* @ODM\String
*/
protected $originalName;
/**
* @var bool
*
* @ODM\Boolean
*/
protected $async = true;
/**
* @var bool
*
* @ODM\Boolean
*/
protected $failed = false;
/**
* @var int
*
* @ODM\Int
*/
protected $tryCount = 0;
/**
* @var string
*
* @ODM\String
*/
protected $groupFieldIdentifier = self::DEFAULT_GROUP_IDENTIFIER;
}
## Instruction:
Add index on delay events collection
## Code After:
<?php
namespace Itkg\DelayEventBundle\Document;
use Gedmo\Timestampable\Traits\TimestampableDocument;
use Itkg\DelayEventBundle\Model\Event as BaseEvent;
use Doctrine\ODM\MongoDB\Mapping\Annotations as ODM;
/**
* Class Event
*
* @ODM\Document(
* repositoryClass="Itkg\DelayEventBundle\Repository\EventRepository",
* collection="itkg_delay_event"
* )
* @ODM\DiscriminatorField(fieldName="type")
* @ODM\InheritanceType("SINGLE_COLLECTION")
*
* @ODM\Indexes({
* @ODM\Index(keys={"failed"="asc", "originalName"="asc", "groupFieldIdentifier"="asc", "createdAt":"asc"}, name="failed_name_group_date")
* })
*/
class Event extends BaseEvent
{
use TimestampableDocument;
/**
* @var string
*
* @ODM\id
*/
protected $id;
/**
* @var string
*
* @ODM\String
*/
protected $originalName;
/**
* @var bool
*
* @ODM\Boolean
*/
protected $async = true;
/**
* @var bool
*
* @ODM\Boolean
*/
protected $failed = false;
/**
* @var int
*
* @ODM\Int
*/
protected $tryCount = 0;
/**
* @var string
*
* @ODM\String
*/
protected $groupFieldIdentifier = self::DEFAULT_GROUP_IDENTIFIER;
}
| <?php
namespace Itkg\DelayEventBundle\Document;
use Gedmo\Timestampable\Traits\TimestampableDocument;
use Itkg\DelayEventBundle\Model\Event as BaseEvent;
use Doctrine\ODM\MongoDB\Mapping\Annotations as ODM;
/**
* Class Event
*
* @ODM\Document(
* repositoryClass="Itkg\DelayEventBundle\Repository\EventRepository",
* collection="itkg_delay_event"
* )
* @ODM\DiscriminatorField(fieldName="type")
* @ODM\InheritanceType("SINGLE_COLLECTION")
+ *
+ * @ODM\Indexes({
+ * @ODM\Index(keys={"failed"="asc", "originalName"="asc", "groupFieldIdentifier"="asc", "createdAt":"asc"}, name="failed_name_group_date")
+ * })
*/
class Event extends BaseEvent
{
use TimestampableDocument;
/**
* @var string
*
* @ODM\id
*/
protected $id;
/**
* @var string
*
* @ODM\String
*/
protected $originalName;
/**
* @var bool
*
* @ODM\Boolean
*/
protected $async = true;
/**
* @var bool
*
* @ODM\Boolean
*/
protected $failed = false;
/**
* @var int
*
* @ODM\Int
*/
protected $tryCount = 0;
/**
* @var string
*
* @ODM\String
*/
protected $groupFieldIdentifier = self::DEFAULT_GROUP_IDENTIFIER;
} | 4 | 0.0625 | 4 | 0 |
7c23567183357eefe685cd8adfdfa9e41dc371d2 | .travis.yml | .travis.yml | language: android
android:
components:
# Use the latest revision of Android SDK Tools
- tools
- platform-tools
# The BuildTools version used by your project
- build-tools-23.0.2
# The SDK version used to compile your project
- android-23
# For AppCompat
- extra
before_cache:
- rm -f $HOME/.gradle/caches/modules-2/modules-2.lock
cache:
directories:
- $HOME/.gradle/caches/
- $HOME/.gradle/wrapper/
jdk:
- oraclejdk8
| language: android
android:
components:
# Use the latest revision of Android SDK Tools
- tools
- platform-tools
# The BuildTools version used by your project
- build-tools-23.0.2
# The SDK version used to compile your project
- android-23
# For AppCompat
- extra
before_install:
- wget -O /tmp/protoc.zip https://github.com/google/protobuf/releases/download/v3.0.0-beta-2/protoc-3.0.0-beta-2-linux-x86_64.zip
- mkdir -p $HOME/bin
- unzip /tmp/protoc.zip protoc -d $HOME/bin
- export PATH="$PATH:$HOME/bin"
before_cache:
- rm -f $HOME/.gradle/caches/modules-2/modules-2.lock
cache:
directories:
- $HOME/.gradle/caches/
- $HOME/.gradle/wrapper/
jdk:
- oraclejdk8
| Install protoc (protobuf compiler) in Travis | Install protoc (protobuf compiler) in Travis | YAML | mit | dasfoo/rover-android-client | yaml | ## Code Before:
language: android
android:
components:
# Use the latest revision of Android SDK Tools
- tools
- platform-tools
# The BuildTools version used by your project
- build-tools-23.0.2
# The SDK version used to compile your project
- android-23
# For AppCompat
- extra
before_cache:
- rm -f $HOME/.gradle/caches/modules-2/modules-2.lock
cache:
directories:
- $HOME/.gradle/caches/
- $HOME/.gradle/wrapper/
jdk:
- oraclejdk8
## Instruction:
Install protoc (protobuf compiler) in Travis
## Code After:
language: android
android:
components:
# Use the latest revision of Android SDK Tools
- tools
- platform-tools
# The BuildTools version used by your project
- build-tools-23.0.2
# The SDK version used to compile your project
- android-23
# For AppCompat
- extra
before_install:
- wget -O /tmp/protoc.zip https://github.com/google/protobuf/releases/download/v3.0.0-beta-2/protoc-3.0.0-beta-2-linux-x86_64.zip
- mkdir -p $HOME/bin
- unzip /tmp/protoc.zip protoc -d $HOME/bin
- export PATH="$PATH:$HOME/bin"
before_cache:
- rm -f $HOME/.gradle/caches/modules-2/modules-2.lock
cache:
directories:
- $HOME/.gradle/caches/
- $HOME/.gradle/wrapper/
jdk:
- oraclejdk8
| language: android
android:
components:
# Use the latest revision of Android SDK Tools
- tools
- platform-tools
# The BuildTools version used by your project
- build-tools-23.0.2
# The SDK version used to compile your project
- android-23
# For AppCompat
- extra
+ before_install:
+ - wget -O /tmp/protoc.zip https://github.com/google/protobuf/releases/download/v3.0.0-beta-2/protoc-3.0.0-beta-2-linux-x86_64.zip
+ - mkdir -p $HOME/bin
+ - unzip /tmp/protoc.zip protoc -d $HOME/bin
+ - export PATH="$PATH:$HOME/bin"
+
before_cache:
- rm -f $HOME/.gradle/caches/modules-2/modules-2.lock
cache:
directories:
- $HOME/.gradle/caches/
- $HOME/.gradle/wrapper/
jdk:
- oraclejdk8 | 6 | 0.230769 | 6 | 0 |
fee30e2365fe2af2500e1d51207c3ab4a4eeefe0 | yard.gemspec | yard.gemspec | require File.expand_path('../lib/yard/version', __FILE__)
Gem::Specification.new do |s|
s.name = "yard"
s.summary = "Documentation tool for consistent and usable documentation in Ruby."
s.description = <<-eof
YARD is a documentation generation tool for the Ruby programming language.
It enables the user to generate consistent, usable documentation that can be
exported to a number of formats very easily, and also supports extending for
custom Ruby constructs such as custom class level definitions.
eof
s.version = YARD::VERSION
s.date = Time.now.strftime('%Y-%m-%d')
s.author = "Loren Segal"
s.email = "lsegal@soen.ca"
s.homepage = "http://yardoc.org"
s.platform = Gem::Platform::RUBY
s.files = Dir.glob("{docs,bin,lib,spec,templates,benchmarks}/**/*") +
['CHANGELOG.md', 'LICENSE', 'LEGAL', 'README.md', 'Rakefile', '.yardopts', __FILE__]
s.require_paths = ['lib']
s.executables = ['yard', 'yardoc', 'yri']
s.has_rdoc = 'yard'
s.rubyforge_project = 'yard'
s.license = 'MIT' if s.respond_to?(:license=)
end | require File.expand_path('../lib/yard/version', __FILE__)
Gem::Specification.new do |s|
s.name = "yard"
s.summary = "Documentation tool for consistent and usable documentation in Ruby."
s.description = <<-eof
YARD is a documentation generation tool for the Ruby programming language.
It enables the user to generate consistent, usable documentation that can be
exported to a number of formats very easily, and also supports extending for
custom Ruby constructs such as custom class level definitions.
eof
s.version = YARD::VERSION
s.date = Time.now.strftime('%Y-%m-%d')
s.author = "Loren Segal"
s.email = "lsegal@soen.ca"
s.homepage = "http://yardoc.org"
s.platform = Gem::Platform::RUBY
s.files = Dir.glob("{docs,bin,lib,spec,templates,benchmarks}/**/*") +
['CHANGELOG.md', 'LICENSE', 'LEGAL', 'README.md', 'Rakefile', '.yardopts', __FILE__]
s.require_paths = ['lib']
s.executables = ['yard', 'yardoc', 'yri']
s.has_rdoc = 'yard'
s.license = 'MIT' if s.respond_to?(:license=)
end
| Remove an obsolete reference to RubyForge. | Remove an obsolete reference to RubyForge.
| Ruby | mit | herosky/yard,herosky/yard,iankronquist/yard,jreinert/yard,iankronquist/yard,iankronquist/yard,ohai/yard,ohai/yard,alexdowad/yard,jreinert/yard,lsegal/yard,alexdowad/yard,lsegal/yard,amclain/yard,herosky/yard,lsegal/yard,thomthom/yard,jreinert/yard,ohai/yard,amclain/yard,thomthom/yard,thomthom/yard,alexdowad/yard,amclain/yard | ruby | ## Code Before:
require File.expand_path('../lib/yard/version', __FILE__)
Gem::Specification.new do |s|
s.name = "yard"
s.summary = "Documentation tool for consistent and usable documentation in Ruby."
s.description = <<-eof
YARD is a documentation generation tool for the Ruby programming language.
It enables the user to generate consistent, usable documentation that can be
exported to a number of formats very easily, and also supports extending for
custom Ruby constructs such as custom class level definitions.
eof
s.version = YARD::VERSION
s.date = Time.now.strftime('%Y-%m-%d')
s.author = "Loren Segal"
s.email = "lsegal@soen.ca"
s.homepage = "http://yardoc.org"
s.platform = Gem::Platform::RUBY
s.files = Dir.glob("{docs,bin,lib,spec,templates,benchmarks}/**/*") +
['CHANGELOG.md', 'LICENSE', 'LEGAL', 'README.md', 'Rakefile', '.yardopts', __FILE__]
s.require_paths = ['lib']
s.executables = ['yard', 'yardoc', 'yri']
s.has_rdoc = 'yard'
s.rubyforge_project = 'yard'
s.license = 'MIT' if s.respond_to?(:license=)
end
## Instruction:
Remove an obsolete reference to RubyForge.
## Code After:
require File.expand_path('../lib/yard/version', __FILE__)
Gem::Specification.new do |s|
s.name = "yard"
s.summary = "Documentation tool for consistent and usable documentation in Ruby."
s.description = <<-eof
YARD is a documentation generation tool for the Ruby programming language.
It enables the user to generate consistent, usable documentation that can be
exported to a number of formats very easily, and also supports extending for
custom Ruby constructs such as custom class level definitions.
eof
s.version = YARD::VERSION
s.date = Time.now.strftime('%Y-%m-%d')
s.author = "Loren Segal"
s.email = "lsegal@soen.ca"
s.homepage = "http://yardoc.org"
s.platform = Gem::Platform::RUBY
s.files = Dir.glob("{docs,bin,lib,spec,templates,benchmarks}/**/*") +
['CHANGELOG.md', 'LICENSE', 'LEGAL', 'README.md', 'Rakefile', '.yardopts', __FILE__]
s.require_paths = ['lib']
s.executables = ['yard', 'yardoc', 'yri']
s.has_rdoc = 'yard'
s.license = 'MIT' if s.respond_to?(:license=)
end
| require File.expand_path('../lib/yard/version', __FILE__)
Gem::Specification.new do |s|
s.name = "yard"
s.summary = "Documentation tool for consistent and usable documentation in Ruby."
s.description = <<-eof
YARD is a documentation generation tool for the Ruby programming language.
It enables the user to generate consistent, usable documentation that can be
exported to a number of formats very easily, and also supports extending for
custom Ruby constructs such as custom class level definitions.
eof
s.version = YARD::VERSION
s.date = Time.now.strftime('%Y-%m-%d')
s.author = "Loren Segal"
s.email = "lsegal@soen.ca"
s.homepage = "http://yardoc.org"
s.platform = Gem::Platform::RUBY
s.files = Dir.glob("{docs,bin,lib,spec,templates,benchmarks}/**/*") +
['CHANGELOG.md', 'LICENSE', 'LEGAL', 'README.md', 'Rakefile', '.yardopts', __FILE__]
s.require_paths = ['lib']
s.executables = ['yard', 'yardoc', 'yri']
s.has_rdoc = 'yard'
- s.rubyforge_project = 'yard'
s.license = 'MIT' if s.respond_to?(:license=)
end | 1 | 0.04 | 0 | 1 |
8fee0f89cceca5a9417eea40ed66371be6284857 | 3.9-minimal/test/from-dockerfile/mod_wsgi.Dockerfile.tpl | 3.9-minimal/test/from-dockerfile/mod_wsgi.Dockerfile.tpl | from #FULL_IMAGE_NAME# as builder
# Add application sources to a directory that the assemble script expects them
# and set permissions so that the container runs without root access
USER 0
ADD app-src /tmp/src
RUN /usr/bin/fix-permissions /tmp/src
USER 1001
# Install the application's dependencies from PyPI
RUN /usr/libexec/s2i/assemble
from #IMAGE_NAME#
# Copy app sources together with the whole virtual environment from the builder image
COPY --from=builder $APP_ROOT $APP_ROOT
# Install httpd package - runtime dependency of our application
USER 0
RUN microdnf install -y httpd
USER 1001
# Set the default command for the resulting image
CMD /usr/libexec/s2i/run
| from #FULL_IMAGE_NAME# as builder
# Add application sources to a directory that the assemble script expects them
# and set permissions so that the container runs without root access
USER 0
ADD app-src /tmp/src
RUN /usr/bin/fix-permissions /tmp/src
USER 1001
# Install the application's dependencies from PyPI
RUN /usr/libexec/s2i/assemble
from #IMAGE_NAME#
# Copy app sources together with the whole virtual environment from the builder image
COPY --from=builder $APP_ROOT $APP_ROOT
# Install httpd and echant packages - runtime dependencies of our application
USER 0
RUN microdnf install -y httpd enchant
USER 1001
# Set the default command for the resulting image
CMD /usr/libexec/s2i/run
| Fix test of the minimal image | Fix test of the minimal image
| Smarty | apache-2.0 | sclorg/s2i-python-container,sclorg/s2i-python-container | smarty | ## Code Before:
from #FULL_IMAGE_NAME# as builder
# Add application sources to a directory that the assemble script expects them
# and set permissions so that the container runs without root access
USER 0
ADD app-src /tmp/src
RUN /usr/bin/fix-permissions /tmp/src
USER 1001
# Install the application's dependencies from PyPI
RUN /usr/libexec/s2i/assemble
from #IMAGE_NAME#
# Copy app sources together with the whole virtual environment from the builder image
COPY --from=builder $APP_ROOT $APP_ROOT
# Install httpd package - runtime dependency of our application
USER 0
RUN microdnf install -y httpd
USER 1001
# Set the default command for the resulting image
CMD /usr/libexec/s2i/run
## Instruction:
Fix test of the minimal image
## Code After:
from #FULL_IMAGE_NAME# as builder
# Add application sources to a directory that the assemble script expects them
# and set permissions so that the container runs without root access
USER 0
ADD app-src /tmp/src
RUN /usr/bin/fix-permissions /tmp/src
USER 1001
# Install the application's dependencies from PyPI
RUN /usr/libexec/s2i/assemble
from #IMAGE_NAME#
# Copy app sources together with the whole virtual environment from the builder image
COPY --from=builder $APP_ROOT $APP_ROOT
# Install httpd and echant packages - runtime dependencies of our application
USER 0
RUN microdnf install -y httpd enchant
USER 1001
# Set the default command for the resulting image
CMD /usr/libexec/s2i/run
| from #FULL_IMAGE_NAME# as builder
# Add application sources to a directory that the assemble script expects them
# and set permissions so that the container runs without root access
USER 0
ADD app-src /tmp/src
RUN /usr/bin/fix-permissions /tmp/src
USER 1001
# Install the application's dependencies from PyPI
RUN /usr/libexec/s2i/assemble
from #IMAGE_NAME#
# Copy app sources together with the whole virtual environment from the builder image
COPY --from=builder $APP_ROOT $APP_ROOT
- # Install httpd package - runtime dependency of our application
? ^
+ # Install httpd and echant packages - runtime dependencies of our application
? +++++++++++ + ^^^
USER 0
- RUN microdnf install -y httpd
+ RUN microdnf install -y httpd enchant
? ++++++++
USER 1001
# Set the default command for the resulting image
CMD /usr/libexec/s2i/run | 4 | 0.166667 | 2 | 2 |
32c8438b48deeab57b7ccfb40fecffefe6b474d8 | website/organizationWhitelist.txt | website/organizationWhitelist.txt | Two Sigma - twosigma.com
Cornell - cornell.edu
NYU - nyu.edu
Stanford - stanford.edu
Harvard - harvard.edu
Harvard - g.harvard.edu
Harvard - fas.harvard.edu
Princeton - princeton.edu
Yale - yale.edu
Columbia - columbia.edu
UPenn - upenn.edu
Brown - brown.edu
Dartmouth - dartmouth.edu
MIT - mit.edu
MIT - alum.mit.edu
CMU - andrew.cmu.edu
RIT - rit.edu
NJU - lamda.nju.edu.cn
Chalmers - student.chalmers.se
McGill - mail.mcgill.ca
Purdue - purdue.edu
| Two Sigma - twosigma.com
Cornell - cornell.edu
NYU - nyu.edu
Stanford - stanford.edu
Harvard - harvard.edu
Harvard - g.harvard.edu
Harvard - fas.harvard.edu
Princeton - princeton.edu
Yale - yale.edu
Columbia - columbia.edu
UPenn - upenn.edu
Brown - brown.edu
Dartmouth - dartmouth.edu
MIT - mit.edu
MIT - alum.mit.edu
CMU - andrew.cmu.edu
RIT - rit.edu
NJU - lamda.nju.edu.cn
Chalmers - student.chalmers.se
McGill - mail.mcgill.ca
Purdue - purdue.edu
Imperial - ic.ac.uk
Imperial - imperial.ac.uk
| Add Imperial to organisation whitelist | Add Imperial to organisation whitelist | Text | mit | yangle/HaliteIO,HaliteChallenge/Halite,lanyudhy/Halite-II,lanyudhy/Halite-II,yangle/HaliteIO,HaliteChallenge/Halite,HaliteChallenge/Halite,lanyudhy/Halite-II,lanyudhy/Halite-II,yangle/HaliteIO,HaliteChallenge/Halite-II,HaliteChallenge/Halite,yangle/HaliteIO,yangle/HaliteIO,HaliteChallenge/Halite-II,yangle/HaliteIO,HaliteChallenge/Halite,HaliteChallenge/Halite-II,yangle/HaliteIO,HaliteChallenge/Halite-II,HaliteChallenge/Halite,HaliteChallenge/Halite-II,lanyudhy/Halite-II,HaliteChallenge/Halite-II,yangle/HaliteIO,HaliteChallenge/Halite-II,HaliteChallenge/Halite-II,lanyudhy/Halite-II,lanyudhy/Halite-II,yangle/HaliteIO,HaliteChallenge/Halite-II,HaliteChallenge/Halite-II,HaliteChallenge/Halite,HaliteChallenge/Halite,yangle/HaliteIO,HaliteChallenge/Halite-II,lanyudhy/Halite-II,lanyudhy/Halite-II,lanyudhy/Halite-II,HaliteChallenge/Halite-II,HaliteChallenge/Halite-II,HaliteChallenge/Halite-II,HaliteChallenge/Halite,lanyudhy/Halite-II,HaliteChallenge/Halite,HaliteChallenge/Halite,HaliteChallenge/Halite-II,yangle/HaliteIO,HaliteChallenge/Halite-II,HaliteChallenge/Halite,yangle/HaliteIO | text | ## Code Before:
Two Sigma - twosigma.com
Cornell - cornell.edu
NYU - nyu.edu
Stanford - stanford.edu
Harvard - harvard.edu
Harvard - g.harvard.edu
Harvard - fas.harvard.edu
Princeton - princeton.edu
Yale - yale.edu
Columbia - columbia.edu
UPenn - upenn.edu
Brown - brown.edu
Dartmouth - dartmouth.edu
MIT - mit.edu
MIT - alum.mit.edu
CMU - andrew.cmu.edu
RIT - rit.edu
NJU - lamda.nju.edu.cn
Chalmers - student.chalmers.se
McGill - mail.mcgill.ca
Purdue - purdue.edu
## Instruction:
Add Imperial to organisation whitelist
## Code After:
Two Sigma - twosigma.com
Cornell - cornell.edu
NYU - nyu.edu
Stanford - stanford.edu
Harvard - harvard.edu
Harvard - g.harvard.edu
Harvard - fas.harvard.edu
Princeton - princeton.edu
Yale - yale.edu
Columbia - columbia.edu
UPenn - upenn.edu
Brown - brown.edu
Dartmouth - dartmouth.edu
MIT - mit.edu
MIT - alum.mit.edu
CMU - andrew.cmu.edu
RIT - rit.edu
NJU - lamda.nju.edu.cn
Chalmers - student.chalmers.se
McGill - mail.mcgill.ca
Purdue - purdue.edu
Imperial - ic.ac.uk
Imperial - imperial.ac.uk
| Two Sigma - twosigma.com
Cornell - cornell.edu
NYU - nyu.edu
Stanford - stanford.edu
Harvard - harvard.edu
Harvard - g.harvard.edu
Harvard - fas.harvard.edu
Princeton - princeton.edu
Yale - yale.edu
Columbia - columbia.edu
UPenn - upenn.edu
Brown - brown.edu
Dartmouth - dartmouth.edu
MIT - mit.edu
MIT - alum.mit.edu
CMU - andrew.cmu.edu
RIT - rit.edu
NJU - lamda.nju.edu.cn
Chalmers - student.chalmers.se
McGill - mail.mcgill.ca
Purdue - purdue.edu
+ Imperial - ic.ac.uk
+ Imperial - imperial.ac.uk | 2 | 0.095238 | 2 | 0 |
f30e0457d45bc88d81c697288ec099e1155ff3ce | usr.sbin/mtest/Makefile | usr.sbin/mtest/Makefile |
PROG= mtest
MAN= mtest.8
WARNS?= 6
.include <bsd.prog.mk>
|
PROG= mtest
MAN= mtest.8
.include <bsd.prog.mk>
| Reduce WARNS count due to alignment warning on ia64. | Reduce WARNS count due to alignment warning on ia64.
| unknown | bsd-3-clause | jrobhoward/SCADAbase,jrobhoward/SCADAbase,jrobhoward/SCADAbase,jrobhoward/SCADAbase,jrobhoward/SCADAbase,jrobhoward/SCADAbase,jrobhoward/SCADAbase,jrobhoward/SCADAbase,jrobhoward/SCADAbase,jrobhoward/SCADAbase,jrobhoward/SCADAbase | unknown | ## Code Before:
PROG= mtest
MAN= mtest.8
WARNS?= 6
.include <bsd.prog.mk>
## Instruction:
Reduce WARNS count due to alignment warning on ia64.
## Code After:
PROG= mtest
MAN= mtest.8
.include <bsd.prog.mk>
|
PROG= mtest
MAN= mtest.8
- WARNS?= 6
.include <bsd.prog.mk> | 1 | 0.166667 | 0 | 1 |
7c3cad2e0db4273e3791770ac99441f3d1a86fcc | _assets/javascripts/fonts.js.coffee.erb | _assets/javascripts/fonts.js.coffee.erb | 'use strict'
WebFont.load
google:
families: [
]
custom:
families: [
'foundation-icons'
]
urls: [
'https://cdnjs.cloudflare.com/ajax/libs/foundicons/3.0.0/foundation-icons.min.css'
]
testStrings:
'foundation-icons': '\uf100\uf101\uf102\uf103'
| 'use strict'
WebFont.load
google:
families: [
'Roboto:400,100,100italic,300,300italic,400italic,500,500italic,700,700italic,900,900italic:latin'
]
custom:
families: [
'foundation-icons'
]
urls: [
'https://cdnjs.cloudflare.com/ajax/libs/foundicons/3.0.0/foundation-icons.min.css'
]
testStrings:
'foundation-icons': '\uf100\uf101\uf102\uf103'
| Add missing font loading entry | Add missing font loading entry
Suggested by Evan Sosenko in
https://github.com/razor-x/jekyll-and-zurb/issues/6
Suggested-by: Evan Sosenko <a67ed6a1087d25201d50318213760ef4657fda27@evansosenko.com>
| HTML+ERB | lgpl-2.1 | matthiasbeyer/wiki.template,matthiasbeyer/wiki.template,matthiasbeyer/wiki.template | html+erb | ## Code Before:
'use strict'
WebFont.load
google:
families: [
]
custom:
families: [
'foundation-icons'
]
urls: [
'https://cdnjs.cloudflare.com/ajax/libs/foundicons/3.0.0/foundation-icons.min.css'
]
testStrings:
'foundation-icons': '\uf100\uf101\uf102\uf103'
## Instruction:
Add missing font loading entry
Suggested by Evan Sosenko in
https://github.com/razor-x/jekyll-and-zurb/issues/6
Suggested-by: Evan Sosenko <a67ed6a1087d25201d50318213760ef4657fda27@evansosenko.com>
## Code After:
'use strict'
WebFont.load
google:
families: [
'Roboto:400,100,100italic,300,300italic,400italic,500,500italic,700,700italic,900,900italic:latin'
]
custom:
families: [
'foundation-icons'
]
urls: [
'https://cdnjs.cloudflare.com/ajax/libs/foundicons/3.0.0/foundation-icons.min.css'
]
testStrings:
'foundation-icons': '\uf100\uf101\uf102\uf103'
| 'use strict'
WebFont.load
google:
families: [
+ 'Roboto:400,100,100italic,300,300italic,400italic,500,500italic,700,700italic,900,900italic:latin'
]
custom:
families: [
'foundation-icons'
]
urls: [
'https://cdnjs.cloudflare.com/ajax/libs/foundicons/3.0.0/foundation-icons.min.css'
]
testStrings:
'foundation-icons': '\uf100\uf101\uf102\uf103' | 1 | 0.066667 | 1 | 0 |
87639f6b3e24fe627963c2ba471b7864afcaed6f | HelloCoreClrApp.Ui/wwwroot/app/greeting/helloworld.html | HelloCoreClrApp.Ui/wwwroot/app/greeting/helloworld.html | <form class="form-horizontal" ng-submit="vm.executeHelloWorld();">
<div class=" form-group">
<div class="col-lg-2">
<input type="text" class="form-control input-lg" placeholder="Give me some text" ng-model="vm.inputText" required />
</div>
<button type="submit" class="col-lg-2 btn btn-primary btn-lg">
Say Hello! <span class="glyphicon glyphicon-play-circle"></span>
</button>
<div class="col-lg-3">
<input class="form-control input-lg" type="text" placeholder="Something will happen over here" ng-model="vm.labelText" disabled>
</div>
</div>
</form>
| <!DOCTYPE html>
<form class="form-horizontal" ng-submit="vm.executeHelloWorld();">
<div class=" form-group">
<div class="col-lg-2">
<input type="text" class="form-control input-lg" placeholder="Give me some text" ng-model="vm.inputText" required />
</div>
<button type="submit" class="col-lg-2 btn btn-primary btn-lg">
Say Hello! <span class="glyphicon glyphicon-play-circle"></span>
</button>
<div class="col-lg-3">
<input class="form-control input-lg" type="text" placeholder="Something will happen over here" ng-model="vm.labelText" disabled>
</div>
</div>
</form>
| Declare doctype to satisfy htmlhint. The doctype wont appear twice in the browser page. | Declare doctype to satisfy htmlhint. The doctype wont appear twice in the browser page.
| HTML | mit | jp7677/hellocoreclr,jp7677/hellocoreclr,jp7677/hellocoreclr,jp7677/hellocoreclr | html | ## Code Before:
<form class="form-horizontal" ng-submit="vm.executeHelloWorld();">
<div class=" form-group">
<div class="col-lg-2">
<input type="text" class="form-control input-lg" placeholder="Give me some text" ng-model="vm.inputText" required />
</div>
<button type="submit" class="col-lg-2 btn btn-primary btn-lg">
Say Hello! <span class="glyphicon glyphicon-play-circle"></span>
</button>
<div class="col-lg-3">
<input class="form-control input-lg" type="text" placeholder="Something will happen over here" ng-model="vm.labelText" disabled>
</div>
</div>
</form>
## Instruction:
Declare doctype to satisfy htmlhint. The doctype wont appear twice in the browser page.
## Code After:
<!DOCTYPE html>
<form class="form-horizontal" ng-submit="vm.executeHelloWorld();">
<div class=" form-group">
<div class="col-lg-2">
<input type="text" class="form-control input-lg" placeholder="Give me some text" ng-model="vm.inputText" required />
</div>
<button type="submit" class="col-lg-2 btn btn-primary btn-lg">
Say Hello! <span class="glyphicon glyphicon-play-circle"></span>
</button>
<div class="col-lg-3">
<input class="form-control input-lg" type="text" placeholder="Something will happen over here" ng-model="vm.labelText" disabled>
</div>
</div>
</form>
| + <!DOCTYPE html>
- <form class="form-horizontal" ng-submit="vm.executeHelloWorld();">
? -
+ <form class="form-horizontal" ng-submit="vm.executeHelloWorld();">
<div class=" form-group">
<div class="col-lg-2">
<input type="text" class="form-control input-lg" placeholder="Give me some text" ng-model="vm.inputText" required />
</div>
<button type="submit" class="col-lg-2 btn btn-primary btn-lg">
Say Hello! <span class="glyphicon glyphicon-play-circle"></span>
</button>
<div class="col-lg-3">
<input class="form-control input-lg" type="text" placeholder="Something will happen over here" ng-model="vm.labelText" disabled>
</div>
</div>
</form> | 3 | 0.176471 | 2 | 1 |
897825a36bb422c7a14424c6cef904cd0f4edbab | app/Policies/APIUserPolicy.php | app/Policies/APIUserPolicy.php | <?php
namespace App\Policies;
use App\User;
use App\APIUser;
use Illuminate\Auth\Access\HandlesAuthorization;
use Illuminate\Http\Request;
class APIUserPolicy
{
use HandlesAuthorization;
public function view_in_admin(User $user) {
if ($user->site_admin) {
return true;
}
}
public function get_all(User $user)
{
if ($user->site_admin) {
return true;
}
}
public function get(User $user, APIUser $api_user)
{
if ($user->site_admin) {
return true;
}
}
public function create(User $user)
{
if ($user->site_admin) {
return true;
}
}
public function update(User $user, APIUser $api_user)
{
if ($user->site_admin) {
return true;
}
}
public function delete(User $user, APIUser $api_user)
{
if ($user->site_admin) {
return true;
}
}
}
| <?php
namespace App\Policies;
use App\User;
use App\APIUser;
use Illuminate\Auth\Access\HandlesAuthorization;
use Illuminate\Http\Request;
class APIUserPolicy
{
use HandlesAuthorization;
public function view_in_admin(User $user) {
if ($user->site_admin) {
return true;
}
}
// Policy used for search -- available to all users.
public function get_all(User $user)
{
return true;
}
public function get(User $user, APIUser $api_user)
{
if ($user->site_admin) {
return true;
}
}
public function create(User $user)
{
if ($user->site_admin) {
return true;
}
}
public function update(User $user, APIUser $api_user)
{
if ($user->site_admin) {
return true;
}
}
public function delete(User $user, APIUser $api_user)
{
if ($user->site_admin) {
return true;
}
}
}
| Change permissions for user search | Change permissions for user search
| PHP | mit | EscherLabs/Graphene,EscherLabs/Graphene,EscherLabs/Graphene | php | ## Code Before:
<?php
namespace App\Policies;
use App\User;
use App\APIUser;
use Illuminate\Auth\Access\HandlesAuthorization;
use Illuminate\Http\Request;
class APIUserPolicy
{
use HandlesAuthorization;
public function view_in_admin(User $user) {
if ($user->site_admin) {
return true;
}
}
public function get_all(User $user)
{
if ($user->site_admin) {
return true;
}
}
public function get(User $user, APIUser $api_user)
{
if ($user->site_admin) {
return true;
}
}
public function create(User $user)
{
if ($user->site_admin) {
return true;
}
}
public function update(User $user, APIUser $api_user)
{
if ($user->site_admin) {
return true;
}
}
public function delete(User $user, APIUser $api_user)
{
if ($user->site_admin) {
return true;
}
}
}
## Instruction:
Change permissions for user search
## Code After:
<?php
namespace App\Policies;
use App\User;
use App\APIUser;
use Illuminate\Auth\Access\HandlesAuthorization;
use Illuminate\Http\Request;
class APIUserPolicy
{
use HandlesAuthorization;
public function view_in_admin(User $user) {
if ($user->site_admin) {
return true;
}
}
// Policy used for search -- available to all users.
public function get_all(User $user)
{
return true;
}
public function get(User $user, APIUser $api_user)
{
if ($user->site_admin) {
return true;
}
}
public function create(User $user)
{
if ($user->site_admin) {
return true;
}
}
public function update(User $user, APIUser $api_user)
{
if ($user->site_admin) {
return true;
}
}
public function delete(User $user, APIUser $api_user)
{
if ($user->site_admin) {
return true;
}
}
}
| <?php
namespace App\Policies;
use App\User;
use App\APIUser;
use Illuminate\Auth\Access\HandlesAuthorization;
use Illuminate\Http\Request;
class APIUserPolicy
{
use HandlesAuthorization;
public function view_in_admin(User $user) {
if ($user->site_admin) {
return true;
}
}
+ // Policy used for search -- available to all users.
public function get_all(User $user)
{
- if ($user->site_admin) {
- return true;
? ----
+ return true;
- }
}
public function get(User $user, APIUser $api_user)
{
if ($user->site_admin) {
return true;
}
}
public function create(User $user)
{
if ($user->site_admin) {
return true;
}
}
public function update(User $user, APIUser $api_user)
{
if ($user->site_admin) {
return true;
}
}
public function delete(User $user, APIUser $api_user)
{
if ($user->site_admin) {
return true;
}
}
} | 5 | 0.092593 | 2 | 3 |
c5a83a4fefb2724e60a269eaf37858b6ab5123ca | package.json | package.json | {
"name": "api-wrapper",
"version": "1.0.0",
"description": "An global api wrapper",
"main": "index.js",
"author": "Kevin BACAS",
"license": "MIT",
"scripts": {
"test": "jest",
"start": "node index.js"
},
"dependencies": {
"axios": "^0.15.3",
"normalizr": "^3.2.2",
"rxjs": "^5.2.0"
},
"devDependencies": {
"eslint": "^3.17.1",
"eslint-config-standard": "^7.0.1",
"eslint-plugin-import": "^2.2.0",
"eslint-plugin-promise": "^3.5.0",
"eslint-plugin-standard": "^2.1.1",
"jest": "^19.0.2"
}
}
| {
"name": "api-wrapper",
"version": "1.0.0",
"description": "An global api wrapper",
"main": "index.js",
"author": "Kevin BACAS",
"license": "MIT",
"scripts": {
"test": "jest --coverage",
"start": "node index.js"
},
"dependencies": {
"axios": "^0.15.3",
"normalizr": "^3.2.2",
"rxjs": "^5.2.0"
},
"devDependencies": {
"eslint": "^3.17.1",
"eslint-config-standard": "^7.0.1",
"eslint-plugin-import": "^2.2.0",
"eslint-plugin-promise": "^3.5.0",
"eslint-plugin-standard": "^2.1.1",
"jest": "^19.0.2"
}
}
| Add coverage report to jest | Add coverage report to jest
| JSON | mit | LaBetePolitique/api-wrapper,LaBetePolitique/api-wrapper | json | ## Code Before:
{
"name": "api-wrapper",
"version": "1.0.0",
"description": "An global api wrapper",
"main": "index.js",
"author": "Kevin BACAS",
"license": "MIT",
"scripts": {
"test": "jest",
"start": "node index.js"
},
"dependencies": {
"axios": "^0.15.3",
"normalizr": "^3.2.2",
"rxjs": "^5.2.0"
},
"devDependencies": {
"eslint": "^3.17.1",
"eslint-config-standard": "^7.0.1",
"eslint-plugin-import": "^2.2.0",
"eslint-plugin-promise": "^3.5.0",
"eslint-plugin-standard": "^2.1.1",
"jest": "^19.0.2"
}
}
## Instruction:
Add coverage report to jest
## Code After:
{
"name": "api-wrapper",
"version": "1.0.0",
"description": "An global api wrapper",
"main": "index.js",
"author": "Kevin BACAS",
"license": "MIT",
"scripts": {
"test": "jest --coverage",
"start": "node index.js"
},
"dependencies": {
"axios": "^0.15.3",
"normalizr": "^3.2.2",
"rxjs": "^5.2.0"
},
"devDependencies": {
"eslint": "^3.17.1",
"eslint-config-standard": "^7.0.1",
"eslint-plugin-import": "^2.2.0",
"eslint-plugin-promise": "^3.5.0",
"eslint-plugin-standard": "^2.1.1",
"jest": "^19.0.2"
}
}
| {
"name": "api-wrapper",
"version": "1.0.0",
"description": "An global api wrapper",
"main": "index.js",
"author": "Kevin BACAS",
"license": "MIT",
"scripts": {
- "test": "jest",
+ "test": "jest --coverage",
? +++++++++++
"start": "node index.js"
},
"dependencies": {
"axios": "^0.15.3",
"normalizr": "^3.2.2",
"rxjs": "^5.2.0"
},
"devDependencies": {
"eslint": "^3.17.1",
"eslint-config-standard": "^7.0.1",
"eslint-plugin-import": "^2.2.0",
"eslint-plugin-promise": "^3.5.0",
"eslint-plugin-standard": "^2.1.1",
"jest": "^19.0.2"
}
} | 2 | 0.08 | 1 | 1 |
f6c2f222db0f529d3f5906d5de7a7541e835ea77 | litecord/api/guilds.py | litecord/api/guilds.py | '''
guilds.py - All handlers under /guilds/*
'''
import json
import logging
from ..utils import _err, _json, strip_user_data
log = logging.getLogger(__name__)
class GuildsEndpoint:
def __init__(self, server):
self.server = server
async def h_post_guilds(self, request):
pass
| '''
guilds.py - All handlers under /guilds/*
'''
import json
import logging
from ..utils import _err, _json, strip_user_data
log = logging.getLogger(__name__)
class GuildsEndpoint:
def __init__(self, server):
self.server = server
async def h_post_guilds(self, request):
pass
async def h_guilds(self, request):
'''
GuildsEndpoint.h_guilds
Handle `GET /guilds/{guild_id}`
'''
_error = await self.server.check_request(request)
_error_json = json.loads(_error.text)
if _error_json['code'] == 0:
return _error
guild_id = request.match_info['guild_id']
guild = self.server.guild_man.get_guild(guild_id)
if guild is None:
return _err('404: Not Found')
return _json(guild.as_json)
async def h_get_guild_channels(self, request):
'''
GuildsEndpoint.h_get_guild_channels
`GET /guilds/{guild_id}/channels`
'''
_error = await self.server.check_request(request)
_error_json = json.loads(_error.text)
if _error_json['code'] == 0:
return _error
guild_id = request.match_info['guild_id']
return _json('Not Implemented')
| Add some dummy routes in GuildsEndpoint | Add some dummy routes in GuildsEndpoint
| Python | mit | nullpixel/litecord,nullpixel/litecord | python | ## Code Before:
'''
guilds.py - All handlers under /guilds/*
'''
import json
import logging
from ..utils import _err, _json, strip_user_data
log = logging.getLogger(__name__)
class GuildsEndpoint:
def __init__(self, server):
self.server = server
async def h_post_guilds(self, request):
pass
## Instruction:
Add some dummy routes in GuildsEndpoint
## Code After:
'''
guilds.py - All handlers under /guilds/*
'''
import json
import logging
from ..utils import _err, _json, strip_user_data
log = logging.getLogger(__name__)
class GuildsEndpoint:
def __init__(self, server):
self.server = server
async def h_post_guilds(self, request):
pass
async def h_guilds(self, request):
'''
GuildsEndpoint.h_guilds
Handle `GET /guilds/{guild_id}`
'''
_error = await self.server.check_request(request)
_error_json = json.loads(_error.text)
if _error_json['code'] == 0:
return _error
guild_id = request.match_info['guild_id']
guild = self.server.guild_man.get_guild(guild_id)
if guild is None:
return _err('404: Not Found')
return _json(guild.as_json)
async def h_get_guild_channels(self, request):
'''
GuildsEndpoint.h_get_guild_channels
`GET /guilds/{guild_id}/channels`
'''
_error = await self.server.check_request(request)
_error_json = json.loads(_error.text)
if _error_json['code'] == 0:
return _error
guild_id = request.match_info['guild_id']
return _json('Not Implemented')
| '''
guilds.py - All handlers under /guilds/*
'''
import json
import logging
from ..utils import _err, _json, strip_user_data
log = logging.getLogger(__name__)
class GuildsEndpoint:
def __init__(self, server):
self.server = server
async def h_post_guilds(self, request):
pass
+
+ async def h_guilds(self, request):
+ '''
+ GuildsEndpoint.h_guilds
+
+ Handle `GET /guilds/{guild_id}`
+ '''
+ _error = await self.server.check_request(request)
+ _error_json = json.loads(_error.text)
+ if _error_json['code'] == 0:
+ return _error
+
+ guild_id = request.match_info['guild_id']
+
+ guild = self.server.guild_man.get_guild(guild_id)
+ if guild is None:
+ return _err('404: Not Found')
+
+ return _json(guild.as_json)
+
+ async def h_get_guild_channels(self, request):
+ '''
+ GuildsEndpoint.h_get_guild_channels
+
+ `GET /guilds/{guild_id}/channels`
+ '''
+ _error = await self.server.check_request(request)
+ _error_json = json.loads(_error.text)
+ if _error_json['code'] == 0:
+ return _error
+
+ guild_id = request.match_info['guild_id']
+
+ return _json('Not Implemented') | 34 | 2.125 | 34 | 0 |
e81ab40e2203b85c8492b53a20da7a75cf128354 | locales/ach/client.properties | locales/ach/client.properties |
searchResultsTitle=Adwogi me yeny
##################
## Project List ##
##################
############
## Editor ##
############
renameProjectSavingIndicator=Gwoko…
############
## Shared ##
############
|
searchResultsTitle=Adwogi me yeny
##################
## Project List ##
##################
############
## Editor ##
############
fileSavingFailedIndicator=Pe ogwoko, tye ka temo ne...
renameProjectSavingIndicator=Gwoko…
############
## Shared ##
############
newProjectInProgressIndicator=Cako projek manyen...
| Update Acholi (ach) localization of Thimble | Pontoon: Update Acholi (ach) localization of Thimble
Localization authors:
- akilama61 <akilama61@gmail.com>
| INI | mpl-2.0 | mozilla/thimble.mozilla.org,mozilla/thimble.mozilla.org,mozilla/thimble.mozilla.org,mozilla/thimble.mozilla.org | ini | ## Code Before:
searchResultsTitle=Adwogi me yeny
##################
## Project List ##
##################
############
## Editor ##
############
renameProjectSavingIndicator=Gwoko…
############
## Shared ##
############
## Instruction:
Pontoon: Update Acholi (ach) localization of Thimble
Localization authors:
- akilama61 <akilama61@gmail.com>
## Code After:
searchResultsTitle=Adwogi me yeny
##################
## Project List ##
##################
############
## Editor ##
############
fileSavingFailedIndicator=Pe ogwoko, tye ka temo ne...
renameProjectSavingIndicator=Gwoko…
############
## Shared ##
############
newProjectInProgressIndicator=Cako projek manyen...
|
searchResultsTitle=Adwogi me yeny
##################
## Project List ##
##################
############
## Editor ##
############
+ fileSavingFailedIndicator=Pe ogwoko, tye ka temo ne...
renameProjectSavingIndicator=Gwoko…
############
## Shared ##
############
+ newProjectInProgressIndicator=Cako projek manyen... | 2 | 0.111111 | 2 | 0 |
62b006993c3ef2fd106cfd4240724d81511b2bd7 | .travis.yml | .travis.yml | language: c
compiler:
- clang
- gcc
before_install:
- "git submodule update --init --recursive"
- "sudo apt-get -qq update"
- "sudo apt-get -qq install doxygen"
- "if [ ! $CC = 'clang' ]; then sudo apt-get -qq install gcc-msp430 msp430-libc msp430mcu; fi"
- "if [ ! $CC = 'clang' ]; then sudo apt-get -qq install gcc-avr avr-libc; fi"
script:
- "make all"
- "src/bin/libxively_unit_test_suite"
- "if [ ! $CC = 'clang' ]; then make ci_msp430; fi"
- "if [ ! $CC = 'clang' ]; then make ci_avr; fi"
| language: c
compiler:
- clang
- gcc
before_install:
- "git submodule update --init --recursive"
- "sudo apt-get -qq update"
- "sudo apt-get -qq install doxygen"
- "if [ ! $CC = 'clang' ]; then sudo apt-get -qq install gcc-msp430 msp430-libc msp430mcu; fi"
- "if [ ! $CC = 'clang' ]; then sudo apt-get -qq install gcc-avr avr-libc; fi"
script:
- "make -C src libxively"
- "make -C src tests XI_UNIT_TEST_TARGET=native TINYTEST_SRCDIR=`readlink -e src/ext/tinytest`"
- "if [ ! $CC = 'clang' ]; then make ci_msp430; fi"
- "if [ ! $CC = 'clang' ]; then make ci_avr; fi"
| Update Travis script to use new system | [ci] Update Travis script to use new system
| YAML | bsd-3-clause | foss-for-synopsys-dwc-arc-processors/libxively,foss-for-synopsys-dwc-arc-processors/libxively | yaml | ## Code Before:
language: c
compiler:
- clang
- gcc
before_install:
- "git submodule update --init --recursive"
- "sudo apt-get -qq update"
- "sudo apt-get -qq install doxygen"
- "if [ ! $CC = 'clang' ]; then sudo apt-get -qq install gcc-msp430 msp430-libc msp430mcu; fi"
- "if [ ! $CC = 'clang' ]; then sudo apt-get -qq install gcc-avr avr-libc; fi"
script:
- "make all"
- "src/bin/libxively_unit_test_suite"
- "if [ ! $CC = 'clang' ]; then make ci_msp430; fi"
- "if [ ! $CC = 'clang' ]; then make ci_avr; fi"
## Instruction:
[ci] Update Travis script to use new system
## Code After:
language: c
compiler:
- clang
- gcc
before_install:
- "git submodule update --init --recursive"
- "sudo apt-get -qq update"
- "sudo apt-get -qq install doxygen"
- "if [ ! $CC = 'clang' ]; then sudo apt-get -qq install gcc-msp430 msp430-libc msp430mcu; fi"
- "if [ ! $CC = 'clang' ]; then sudo apt-get -qq install gcc-avr avr-libc; fi"
script:
- "make -C src libxively"
- "make -C src tests XI_UNIT_TEST_TARGET=native TINYTEST_SRCDIR=`readlink -e src/ext/tinytest`"
- "if [ ! $CC = 'clang' ]; then make ci_msp430; fi"
- "if [ ! $CC = 'clang' ]; then make ci_avr; fi"
| language: c
compiler:
- clang
- gcc
before_install:
- "git submodule update --init --recursive"
- "sudo apt-get -qq update"
- "sudo apt-get -qq install doxygen"
- "if [ ! $CC = 'clang' ]; then sudo apt-get -qq install gcc-msp430 msp430-libc msp430mcu; fi"
- "if [ ! $CC = 'clang' ]; then sudo apt-get -qq install gcc-avr avr-libc; fi"
script:
- - "make all"
- - "src/bin/libxively_unit_test_suite"
+ - "make -C src libxively"
+ - "make -C src tests XI_UNIT_TEST_TARGET=native TINYTEST_SRCDIR=`readlink -e src/ext/tinytest`"
- "if [ ! $CC = 'clang' ]; then make ci_msp430; fi"
- "if [ ! $CC = 'clang' ]; then make ci_avr; fi" | 4 | 0.266667 | 2 | 2 |
4874a150f0cbd14d69bbb4915ecf78991ba46d77 | templates/default/tags/setup.rb | templates/default/tags/setup.rb | def init
sections :index, [:example, :overload, [T('docstring')], :param, :option, :yields,
:yieldparam, :yieldreturn, :return, :raises, :see, :author, :since, :version]
end
def param
tag :param
end
def yields
tag :yield, :no_names => true, :no_types => true
end
def yieldparam
tag :yieldparam
end
def yieldreturn
tag :yieldreturn, :no_names => true
end
def return
tag :return, :no_names => true
end
def raises
tag :raise, :no_names => true
end
def author
tag :author, :no_types => true, :no_names => true
end
def since
tag :since, :no_types => true, :no_names => true
end
def version
tag :version, :no_types => true, :no_names => true
end
def tag(name, opts = {})
return unless object.has_tag?(name)
@no_names = true if opts[:no_names]
@no_types = true if opts[:no_types]
@name = name
erb('tag')
end
| def init
sections :index, [:example, :overload, [T('docstring')], :param, :option, :yields,
:yieldparam, :yieldreturn, :return, :raises, :see, :author, :since, :version]
end
def param
tag :param
end
def yields
tag :yield, :no_names => true, :no_types => true
end
def yieldparam
tag :yieldparam
end
def yieldreturn
tag :yieldreturn, :no_names => true
end
def return
tag :return, :no_names => true
end
def raises
tag :raise, :no_names => true
end
def author
tag :author, :no_types => true, :no_names => true
end
def since
tag :since, :no_types => true, :no_names => true
end
def version
tag :version, :no_types => true, :no_names => true
end
def tag(name, opts = {})
return unless object.has_tag?(name)
@no_names = true if opts[:no_names]
@no_types = true if opts[:no_types]
@name = name
out = erb('tag')
@no_names, @no_types = nil, nil
out
end
| Fix showing of yieldparam type/name info | Fix showing of yieldparam type/name info
| Ruby | mit | ohai/yard,iankronquist/yard,lsegal/yard,ohai/yard,herosky/yard,iankronquist/yard,amclain/yard,thomthom/yard,herosky/yard,alexdowad/yard,thomthom/yard,herosky/yard,lsegal/yard,iankronquist/yard,alexdowad/yard,travis-repos/yard,travis-repos/yard,alexdowad/yard,amclain/yard,jreinert/yard,jreinert/yard,ohai/yard,jreinert/yard,lsegal/yard,amclain/yard,thomthom/yard | ruby | ## Code Before:
def init
sections :index, [:example, :overload, [T('docstring')], :param, :option, :yields,
:yieldparam, :yieldreturn, :return, :raises, :see, :author, :since, :version]
end
def param
tag :param
end
def yields
tag :yield, :no_names => true, :no_types => true
end
def yieldparam
tag :yieldparam
end
def yieldreturn
tag :yieldreturn, :no_names => true
end
def return
tag :return, :no_names => true
end
def raises
tag :raise, :no_names => true
end
def author
tag :author, :no_types => true, :no_names => true
end
def since
tag :since, :no_types => true, :no_names => true
end
def version
tag :version, :no_types => true, :no_names => true
end
def tag(name, opts = {})
return unless object.has_tag?(name)
@no_names = true if opts[:no_names]
@no_types = true if opts[:no_types]
@name = name
erb('tag')
end
## Instruction:
Fix showing of yieldparam type/name info
## Code After:
def init
sections :index, [:example, :overload, [T('docstring')], :param, :option, :yields,
:yieldparam, :yieldreturn, :return, :raises, :see, :author, :since, :version]
end
def param
tag :param
end
def yields
tag :yield, :no_names => true, :no_types => true
end
def yieldparam
tag :yieldparam
end
def yieldreturn
tag :yieldreturn, :no_names => true
end
def return
tag :return, :no_names => true
end
def raises
tag :raise, :no_names => true
end
def author
tag :author, :no_types => true, :no_names => true
end
def since
tag :since, :no_types => true, :no_names => true
end
def version
tag :version, :no_types => true, :no_names => true
end
def tag(name, opts = {})
return unless object.has_tag?(name)
@no_names = true if opts[:no_names]
@no_types = true if opts[:no_types]
@name = name
out = erb('tag')
@no_names, @no_types = nil, nil
out
end
| def init
sections :index, [:example, :overload, [T('docstring')], :param, :option, :yields,
:yieldparam, :yieldreturn, :return, :raises, :see, :author, :since, :version]
end
def param
tag :param
end
def yields
tag :yield, :no_names => true, :no_types => true
end
def yieldparam
tag :yieldparam
end
def yieldreturn
tag :yieldreturn, :no_names => true
end
def return
tag :return, :no_names => true
end
def raises
tag :raise, :no_names => true
end
def author
tag :author, :no_types => true, :no_names => true
end
def since
tag :since, :no_types => true, :no_names => true
end
def version
tag :version, :no_types => true, :no_names => true
end
def tag(name, opts = {})
return unless object.has_tag?(name)
@no_names = true if opts[:no_names]
@no_types = true if opts[:no_types]
@name = name
- erb('tag')
+ out = erb('tag')
? ++++++
+ @no_names, @no_types = nil, nil
+ out
end | 4 | 0.083333 | 3 | 1 |
09b2c80a30e24e3842c32c52a474feecee3c86a0 | app/views/servers/_nav.html.haml | app/views/servers/_nav.html.haml | - if development?
%nav.jg-navbar.navbar.navbar-default{"role"=>"navigation"}
.navbar-header
.navbar-collapse.collapse
%ul.nav.navbar-nav
%li= link_to server.name, server
%li= link_to 'Backups', server_backups_path(server)
%li= link_to 'Console', console_server_path(server) | %nav.jg-navbar.navbar.navbar-default{"role"=>"navigation"}
.navbar-header
.navbar-collapse.collapse
%ul.nav.navbar-nav
%li= link_to server.name, server
%li= link_to 'Backups', server_backups_path(server)
%li= link_to 'Console', console_server_path(server) | Allow nav menu for servers on all env | Allow nav menu for servers on all env | Haml | apache-2.0 | OnApp/cloudnet,OnApp/cloudnet,OnApp/cloudnet,OnApp/cloudnet | haml | ## Code Before:
- if development?
%nav.jg-navbar.navbar.navbar-default{"role"=>"navigation"}
.navbar-header
.navbar-collapse.collapse
%ul.nav.navbar-nav
%li= link_to server.name, server
%li= link_to 'Backups', server_backups_path(server)
%li= link_to 'Console', console_server_path(server)
## Instruction:
Allow nav menu for servers on all env
## Code After:
%nav.jg-navbar.navbar.navbar-default{"role"=>"navigation"}
.navbar-header
.navbar-collapse.collapse
%ul.nav.navbar-nav
%li= link_to server.name, server
%li= link_to 'Backups', server_backups_path(server)
%li= link_to 'Console', console_server_path(server) | - - if development?
- %nav.jg-navbar.navbar.navbar-default{"role"=>"navigation"}
? --
+ %nav.jg-navbar.navbar.navbar-default{"role"=>"navigation"}
- .navbar-header
? --
+ .navbar-header
- .navbar-collapse.collapse
? --
+ .navbar-collapse.collapse
- %ul.nav.navbar-nav
? --
+ %ul.nav.navbar-nav
- %li= link_to server.name, server
? --
+ %li= link_to server.name, server
- %li= link_to 'Backups', server_backups_path(server)
? --
+ %li= link_to 'Backups', server_backups_path(server)
- %li= link_to 'Console', console_server_path(server)
? -- --
+ %li= link_to 'Console', console_server_path(server) | 15 | 1.875 | 7 | 8 |
ebef9dc7e761411a1f2712fe37a4b9697dc35d9a | lib/cloudstrap/amazon.rb | lib/cloudstrap/amazon.rb | require_relative 'amazon/ec2'
require_relative 'amazon/iam'
require_relative 'amazon/route53'
| require_relative 'amazon/ec2'
require_relative 'amazon/elb'
require_relative 'amazon/iam'
require_relative 'amazon/route53'
| Add ELB to require wrapper | Add ELB to require wrapper
| Ruby | mit | colstrom/cloudstrap | ruby | ## Code Before:
require_relative 'amazon/ec2'
require_relative 'amazon/iam'
require_relative 'amazon/route53'
## Instruction:
Add ELB to require wrapper
## Code After:
require_relative 'amazon/ec2'
require_relative 'amazon/elb'
require_relative 'amazon/iam'
require_relative 'amazon/route53'
| require_relative 'amazon/ec2'
+ require_relative 'amazon/elb'
require_relative 'amazon/iam'
require_relative 'amazon/route53' | 1 | 0.333333 | 1 | 0 |
8ec4047abc9b70b81bcfef5e9e4a0284bf394028 | dinner.md | dinner.md | Dylan - Basic Nachos
| <h3>Dylan</h3>
Basic Nachos.
<h3>Chris</h3>
Chicken Burritto.
| Add mine; better formatting & hierarchy. | Add mine; better formatting & hierarchy. | Markdown | mit | zachlatta/codeday-boulder-food,zachlatta/codeday-boulder-food,zachlatta/codeday-boulder-food,zachlatta/codeday-boulder-food,zachlatta/codeday-boulder-food,zachlatta/codeday-boulder-food | markdown | ## Code Before:
Dylan - Basic Nachos
## Instruction:
Add mine; better formatting & hierarchy.
## Code After:
<h3>Dylan</h3>
Basic Nachos.
<h3>Chris</h3>
Chicken Burritto.
| - Dylan - Basic Nachos
+ <h3>Dylan</h3>
+ Basic Nachos.
+ <h3>Chris</h3>
+ Chicken Burritto. | 5 | 5 | 4 | 1 |
876a970e7d134c0c84bf24c91cf5f78898bcf2a1 | versions/versions.ini | versions/versions.ini | [1.0a93]
nextVersion = "none"
updateScript = "v1_0a92_to_v1_0a93.sh"
[1.0a92]
nextVersion = "1.0a93"
updateScript = "v1_0a91_to_v1_0a92.sh"
[1.0a91]
nextVersion = "1.0a92"
| [1.0a94]
nextVersion = "none"
updateScript = "v1_0a93_to_v1_0a94.sh"
[1.0a93]
nextVersion = "1.0a94"
updateScript = "v1_0a92_to_v1_0a93.sh"
[1.0a92]
nextVersion = "1.0a93"
updateScript = "v1_0a91_to_v1_0a92.sh"
[1.0a91]
nextVersion = "1.0a92"
| Add the new structWSF version 1.094 step. | Add the new structWSF version 1.094 step.
| INI | apache-2.0 | structureddynamics/structWSF-Upgrader | ini | ## Code Before:
[1.0a93]
nextVersion = "none"
updateScript = "v1_0a92_to_v1_0a93.sh"
[1.0a92]
nextVersion = "1.0a93"
updateScript = "v1_0a91_to_v1_0a92.sh"
[1.0a91]
nextVersion = "1.0a92"
## Instruction:
Add the new structWSF version 1.094 step.
## Code After:
[1.0a94]
nextVersion = "none"
updateScript = "v1_0a93_to_v1_0a94.sh"
[1.0a93]
nextVersion = "1.0a94"
updateScript = "v1_0a92_to_v1_0a93.sh"
[1.0a92]
nextVersion = "1.0a93"
updateScript = "v1_0a91_to_v1_0a92.sh"
[1.0a91]
nextVersion = "1.0a92"
| + [1.0a94]
+
+ nextVersion = "none"
+ updateScript = "v1_0a93_to_v1_0a94.sh"
+
[1.0a93]
- nextVersion = "none"
? ^^^^
+ nextVersion = "1.0a94"
? ^^^^^^
updateScript = "v1_0a92_to_v1_0a93.sh"
[1.0a92]
nextVersion = "1.0a93"
updateScript = "v1_0a91_to_v1_0a92.sh"
[1.0a91]
nextVersion = "1.0a92" | 7 | 0.538462 | 6 | 1 |
817d9c78f939de2b01ff518356ed0414178aaa6d | avalonstar/apps/api/serializers.py | avalonstar/apps/api/serializers.py | from rest_framework import serializers
from apps.broadcasts.models import Broadcast, Series
from apps.games.models import Game
class BroadcastSerializer(serializers.ModelSerializer):
class Meta:
depth = 1
model = Broadcast
class SeriesSerializer(serializers.ModelSerializer):
class Meta:
model = Series
class GameSerializer(serializers.ModelSerializer):
class Meta:
model = Game
| from rest_framework import serializers
from apps.broadcasts.models import Broadcast, Raid, Series
from apps.games.models import Game
class BroadcastSerializer(serializers.ModelSerializer):
class Meta:
depth = 1
model = Broadcast
class RaidSerializer(serializers.ModelSerializer):
class Meta:
model = Raid
class SeriesSerializer(serializers.ModelSerializer):
class Meta:
model = Series
class GameSerializer(serializers.ModelSerializer):
class Meta:
model = Game
| Add Raid to the API. | Add Raid to the API.
| Python | apache-2.0 | bryanveloso/avalonstar-tv,bryanveloso/avalonstar-tv,bryanveloso/avalonstar-tv | python | ## Code Before:
from rest_framework import serializers
from apps.broadcasts.models import Broadcast, Series
from apps.games.models import Game
class BroadcastSerializer(serializers.ModelSerializer):
class Meta:
depth = 1
model = Broadcast
class SeriesSerializer(serializers.ModelSerializer):
class Meta:
model = Series
class GameSerializer(serializers.ModelSerializer):
class Meta:
model = Game
## Instruction:
Add Raid to the API.
## Code After:
from rest_framework import serializers
from apps.broadcasts.models import Broadcast, Raid, Series
from apps.games.models import Game
class BroadcastSerializer(serializers.ModelSerializer):
class Meta:
depth = 1
model = Broadcast
class RaidSerializer(serializers.ModelSerializer):
class Meta:
model = Raid
class SeriesSerializer(serializers.ModelSerializer):
class Meta:
model = Series
class GameSerializer(serializers.ModelSerializer):
class Meta:
model = Game
| from rest_framework import serializers
- from apps.broadcasts.models import Broadcast, Series
+ from apps.broadcasts.models import Broadcast, Raid, Series
? ++++++
from apps.games.models import Game
class BroadcastSerializer(serializers.ModelSerializer):
class Meta:
depth = 1
model = Broadcast
+
+
+ class RaidSerializer(serializers.ModelSerializer):
+ class Meta:
+ model = Raid
class SeriesSerializer(serializers.ModelSerializer):
class Meta:
model = Series
class GameSerializer(serializers.ModelSerializer):
class Meta:
model = Game | 7 | 0.35 | 6 | 1 |
b3adec73ca58e00ccd18c10270fc866c1afaab30 | ruby_event_store-rom/spec/rom/sql_spec.rb | ruby_event_store-rom/spec/rom/sql_spec.rb | require 'spec_helper'
module RubyEventStore::ROM
RSpec.describe SQL::SpecHelper do
include SchemaHelper
around(:each) do |example|
begin
establish_database_connection
load_database_schema
example.run
ensure
drop_database
end
end
specify '#env gives access to ROM container' do
expect(subject.env.container).to be_a(::ROM::Container)
end
end
end
| require 'spec_helper'
require 'ruby_event_store/rom/sql'
module RubyEventStore::ROM::SQL
RSpec.describe SpecHelper do
include SchemaHelper
around(:each) do |example|
begin
establish_database_connection
load_database_schema
example.run
ensure
drop_database
end
end
specify '#env gives access to ROM container' do
expect(subject.env.container).to be_a(::ROM::Container)
end
end
end
| FIx reference to SQL adapter | FIx reference to SQL adapter
| Ruby | mit | RailsEventStore/rails_event_store,RailsEventStore/rails_event_store,mpraglowski/rails_event_store,mpraglowski/rails_event_store,mpraglowski/rails_event_store,arkency/rails_event_store,arkency/rails_event_store,mpraglowski/rails_event_store,arkency/rails_event_store,arkency/rails_event_store,RailsEventStore/rails_event_store,RailsEventStore/rails_event_store | ruby | ## Code Before:
require 'spec_helper'
module RubyEventStore::ROM
RSpec.describe SQL::SpecHelper do
include SchemaHelper
around(:each) do |example|
begin
establish_database_connection
load_database_schema
example.run
ensure
drop_database
end
end
specify '#env gives access to ROM container' do
expect(subject.env.container).to be_a(::ROM::Container)
end
end
end
## Instruction:
FIx reference to SQL adapter
## Code After:
require 'spec_helper'
require 'ruby_event_store/rom/sql'
module RubyEventStore::ROM::SQL
RSpec.describe SpecHelper do
include SchemaHelper
around(:each) do |example|
begin
establish_database_connection
load_database_schema
example.run
ensure
drop_database
end
end
specify '#env gives access to ROM container' do
expect(subject.env.container).to be_a(::ROM::Container)
end
end
end
| require 'spec_helper'
+ require 'ruby_event_store/rom/sql'
- module RubyEventStore::ROM
+ module RubyEventStore::ROM::SQL
? +++++
- RSpec.describe SQL::SpecHelper do
? -----
+ RSpec.describe SpecHelper do
include SchemaHelper
around(:each) do |example|
begin
establish_database_connection
load_database_schema
example.run
ensure
drop_database
end
end
specify '#env gives access to ROM container' do
expect(subject.env.container).to be_a(::ROM::Container)
end
end
end | 5 | 0.238095 | 3 | 2 |
c743dbfc389cd12a3615d31aa90f2cca14422895 | resources/views/admin/formElements/date.blade.php | resources/views/admin/formElements/date.blade.php | <div class="form-group datepicker {{ $errors->has($fieldName) ? 'has-error' : '' }}">
{!! Form::label($elementName, $options['label']) !!}
{!! Form::text($elementName, $record->$fieldName, [
'id' => $elementName,
'class' => "form-control",
'placeholder' => $options['label'],
] ) !!}
<div class="text-red">
{{ join($errors->get($fieldName), '<br />') }}
</div>
</div>
| <div class="form-group {{ $errors->has($fieldName) ? 'has-error' : '' }}" style="position: 'relative'">
{!! Form::label($elementName, $options['label']) !!}
<div class="datepicker input-group">
{!! Form::text($elementName, $record->$fieldName, [
'id' => $elementName,
'class' => "form-control",
'placeholder' => $options['label'],
] ) !!}
</div>
<div class="text-red">
{{ join($errors->get($fieldName), '<br />') }}
</div>
</div>
| Fix wrong position of date widget | Fix wrong position of date widget | PHP | mit | despark/ignicms,despark/ignicms,despark/ignicms,despark/ignicms | php | ## Code Before:
<div class="form-group datepicker {{ $errors->has($fieldName) ? 'has-error' : '' }}">
{!! Form::label($elementName, $options['label']) !!}
{!! Form::text($elementName, $record->$fieldName, [
'id' => $elementName,
'class' => "form-control",
'placeholder' => $options['label'],
] ) !!}
<div class="text-red">
{{ join($errors->get($fieldName), '<br />') }}
</div>
</div>
## Instruction:
Fix wrong position of date widget
## Code After:
<div class="form-group {{ $errors->has($fieldName) ? 'has-error' : '' }}" style="position: 'relative'">
{!! Form::label($elementName, $options['label']) !!}
<div class="datepicker input-group">
{!! Form::text($elementName, $record->$fieldName, [
'id' => $elementName,
'class' => "form-control",
'placeholder' => $options['label'],
] ) !!}
</div>
<div class="text-red">
{{ join($errors->get($fieldName), '<br />') }}
</div>
</div>
| - <div class="form-group datepicker {{ $errors->has($fieldName) ? 'has-error' : '' }}">
? -----------
+ <div class="form-group {{ $errors->has($fieldName) ? 'has-error' : '' }}" style="position: 'relative'">
? +++++++++++++++++++++++++++++
{!! Form::label($elementName, $options['label']) !!}
+ <div class="datepicker input-group">
- {!! Form::text($elementName, $record->$fieldName, [
+ {!! Form::text($elementName, $record->$fieldName, [
? ++++
- 'id' => $elementName,
? ^
+ 'id' => $elementName,
? ^^^^^^^^
- 'class' => "form-control",
? ^
+ 'class' => "form-control",
? ^^^^^^^^
- 'placeholder' => $options['label'],
? ^
+ 'placeholder' => $options['label'],
? ^^^^^^^^
- ] ) !!}
+ ] ) !!}
+ </div>
<div class="text-red">
{{ join($errors->get($fieldName), '<br />') }}
</div>
</div> | 14 | 1.272727 | 8 | 6 |
d9782943b48217f1d25f99f9bc1454cee3f42ab8 | parseresults.sh | parseresults.sh | for result in c*result.txt
do
cat $result | grep -e '^HTTP/1.1,4' > $result._errors.txt
cat $result | grep -e '^HTTP/1.1,5' >> $result._errors.txt
cat $result | grep -e '^HTTP/1.1,3' > $result._redirects.txt
cat $result | grep -e '^HTTP/1.1,200' | cut -d, -f 3,5,7 > $result._normal.txt
done
| SQLITECMD=`which sqlite3`
if [ ! -x $SQLITECMD ]
then
echo "Cannot use sqlite3 command. Please check if SQLite3 is installed on this system and you have permission to use it."
exit 1
fi
if [ -f results.db ]
then
rm results.db
fi
for result in c*result.txt
do
cat $result | grep -e '^HTTP/1.1,4' > $result._errors.txt
cat $result | grep -e '^HTTP/1.1,5' >> $result._errors.txt
cat $result | grep -e '^HTTP/1.1,3' > $result._redirects.txt
cat $result | grep -e '^HTTP/1.1,200' | cut -d, -f 3,5,7 > $result._normal.txt
done
| Check for usable SQLite command and cleanup working dir. | [FEATURE] Check for usable SQLite command and cleanup working dir.
| Shell | bsd-3-clause | bvdwiel/loadtest,bvdwiel/loadtest | shell | ## Code Before:
for result in c*result.txt
do
cat $result | grep -e '^HTTP/1.1,4' > $result._errors.txt
cat $result | grep -e '^HTTP/1.1,5' >> $result._errors.txt
cat $result | grep -e '^HTTP/1.1,3' > $result._redirects.txt
cat $result | grep -e '^HTTP/1.1,200' | cut -d, -f 3,5,7 > $result._normal.txt
done
## Instruction:
[FEATURE] Check for usable SQLite command and cleanup working dir.
## Code After:
SQLITECMD=`which sqlite3`
if [ ! -x $SQLITECMD ]
then
echo "Cannot use sqlite3 command. Please check if SQLite3 is installed on this system and you have permission to use it."
exit 1
fi
if [ -f results.db ]
then
rm results.db
fi
for result in c*result.txt
do
cat $result | grep -e '^HTTP/1.1,4' > $result._errors.txt
cat $result | grep -e '^HTTP/1.1,5' >> $result._errors.txt
cat $result | grep -e '^HTTP/1.1,3' > $result._redirects.txt
cat $result | grep -e '^HTTP/1.1,200' | cut -d, -f 3,5,7 > $result._normal.txt
done
| + SQLITECMD=`which sqlite3`
+ if [ ! -x $SQLITECMD ]
+ then
+ echo "Cannot use sqlite3 command. Please check if SQLite3 is installed on this system and you have permission to use it."
+ exit 1
+ fi
+ if [ -f results.db ]
+ then
+ rm results.db
+ fi
+
for result in c*result.txt
do
cat $result | grep -e '^HTTP/1.1,4' > $result._errors.txt
cat $result | grep -e '^HTTP/1.1,5' >> $result._errors.txt
cat $result | grep -e '^HTTP/1.1,3' > $result._redirects.txt
cat $result | grep -e '^HTTP/1.1,200' | cut -d, -f 3,5,7 > $result._normal.txt
done | 11 | 1.571429 | 11 | 0 |
e7a51900c7247fcbd1df1155e34ccbf1290bf463 | .travis.yml | .travis.yml | language: python
python:
- "2.6"
- "2.7"
- "3.2"
- "3.3"
- "3.4"
- "pypy"
sudo: false
install: pip install -r requirements.txt
script:
- py.test -v --cov flake8diff --cov-report term-missing
after_success:
- coveralls
| language: python
python:
- "2.7"
- "3.2"
- "3.3"
- "3.4"
- "pypy"
sudo: false
install: pip install -r requirements.txt
script:
- py.test -v --cov flake8diff --cov-report term-missing
after_success:
- coveralls
| Drop Python 2.6.x for now | Drop Python 2.6.x for now
| YAML | mit | dealertrack/flake8-diff,miki725/flake8-diff | yaml | ## Code Before:
language: python
python:
- "2.6"
- "2.7"
- "3.2"
- "3.3"
- "3.4"
- "pypy"
sudo: false
install: pip install -r requirements.txt
script:
- py.test -v --cov flake8diff --cov-report term-missing
after_success:
- coveralls
## Instruction:
Drop Python 2.6.x for now
## Code After:
language: python
python:
- "2.7"
- "3.2"
- "3.3"
- "3.4"
- "pypy"
sudo: false
install: pip install -r requirements.txt
script:
- py.test -v --cov flake8diff --cov-report term-missing
after_success:
- coveralls
| language: python
python:
- - "2.6"
- "2.7"
- "3.2"
- "3.3"
- "3.4"
- "pypy"
sudo: false
install: pip install -r requirements.txt
script:
- py.test -v --cov flake8diff --cov-report term-missing
after_success:
- coveralls | 1 | 0.071429 | 0 | 1 |
46231cf5938f090521f7d65c881f0df8b6e34511 | app/src/main/java/mozilla/org/webmaker/activity/Tinker.java | app/src/main/java/mozilla/org/webmaker/activity/Tinker.java | package mozilla.org.webmaker.activity;
import android.app.ActionBar;
import android.graphics.Color;
import android.graphics.drawable.ColorDrawable;
import android.os.Bundle;
import android.view.Menu;
import android.view.Window;
import android.view.WindowManager;
import android.widget.RelativeLayout;
import mozilla.org.webmaker.R;
import mozilla.org.webmaker.WebmakerActivity;
import mozilla.org.webmaker.view.WebmakerWebView;
public class Tinker extends WebmakerActivity {
public Tinker() {
super("tinker", R.id.tinker_layout, R.layout.tinker_layout, R.menu.menu_tinker);
}
@Override
protected void onCreate(Bundle savedInstanceState) {
super.onCreate(savedInstanceState);
// Custom styles
ActionBar actionBar = getActionBar();
ColorDrawable colorOne = new ColorDrawable(Color.parseColor("#ff303250"));
ColorDrawable colorTwo = new ColorDrawable(Color.parseColor("#ff303250"));
actionBar.setStackedBackgroundDrawable(colorOne);
actionBar.setBackgroundDrawable(colorTwo);
Window window = getWindow();
window.addFlags(WindowManager.LayoutParams.FLAG_DRAWS_SYSTEM_BAR_BACKGROUNDS);
window.clearFlags(WindowManager.LayoutParams.FLAG_TRANSLUCENT_STATUS);
window.setStatusBarColor(0xff282733);
}
@Override
public boolean onCreateOptionsMenu(Menu menu) {
getMenuInflater().inflate(R.menu.menu_main, menu);
return true;
}
}
| package mozilla.org.webmaker.activity;
import android.app.ActionBar;
import android.content.res.Resources;
import android.graphics.Color;
import android.graphics.drawable.ColorDrawable;
import android.os.Bundle;
import android.view.Menu;
import android.view.Window;
import android.view.WindowManager;
import android.widget.RelativeLayout;
import mozilla.org.webmaker.R;
import mozilla.org.webmaker.WebmakerActivity;
import mozilla.org.webmaker.view.WebmakerWebView;
public class Tinker extends WebmakerActivity {
public Tinker() {
super("tinker", R.id.tinker_layout, R.layout.tinker_layout, R.menu.menu_tinker);
}
@Override
protected void onCreate(Bundle savedInstanceState) {
super.onCreate(savedInstanceState);
// Custom styles
Resources res = getResources();
int shadowPlum = res.getColor(R.color.shadow_plum);
int plum = res.getColor(R.color.plum);
ActionBar actionBar = getActionBar();
actionBar.setStackedBackgroundDrawable(plum);
actionBar.setBackgroundDrawable(plum);
Window window = getWindow();
window.addFlags(WindowManager.LayoutParams.FLAG_DRAWS_SYSTEM_BAR_BACKGROUNDS);
window.clearFlags(WindowManager.LayoutParams.FLAG_TRANSLUCENT_STATUS);
window.setStatusBarColor(shadowPlum);
}
@Override
public boolean onCreateOptionsMenu(Menu menu) {
getMenuInflater().inflate(R.menu.menu_main, menu);
return true;
}
}
| Use plum and shadow_plum from colors.xml | Use plum and shadow_plum from colors.xml
| Java | mpl-2.0 | adamlofting/webmaker-android,k88hudson/webmaker-android,alicoding/webmaker-android,alanmoo/webmaker-android,alicoding/webmaker-android,bolaram/webmaker-android,alanmoo/webmaker-android,j796160836/webmaker-android,gvn/webmaker-android,rodmoreno/webmaker-android,j796160836/webmaker-android,rodmoreno/webmaker-android,mozilla/webmaker-android,gvn/webmaker-android,bolaram/webmaker-android,k88hudson/webmaker-android,mozilla/webmaker-android | java | ## Code Before:
package mozilla.org.webmaker.activity;
import android.app.ActionBar;
import android.graphics.Color;
import android.graphics.drawable.ColorDrawable;
import android.os.Bundle;
import android.view.Menu;
import android.view.Window;
import android.view.WindowManager;
import android.widget.RelativeLayout;
import mozilla.org.webmaker.R;
import mozilla.org.webmaker.WebmakerActivity;
import mozilla.org.webmaker.view.WebmakerWebView;
public class Tinker extends WebmakerActivity {
public Tinker() {
super("tinker", R.id.tinker_layout, R.layout.tinker_layout, R.menu.menu_tinker);
}
@Override
protected void onCreate(Bundle savedInstanceState) {
super.onCreate(savedInstanceState);
// Custom styles
ActionBar actionBar = getActionBar();
ColorDrawable colorOne = new ColorDrawable(Color.parseColor("#ff303250"));
ColorDrawable colorTwo = new ColorDrawable(Color.parseColor("#ff303250"));
actionBar.setStackedBackgroundDrawable(colorOne);
actionBar.setBackgroundDrawable(colorTwo);
Window window = getWindow();
window.addFlags(WindowManager.LayoutParams.FLAG_DRAWS_SYSTEM_BAR_BACKGROUNDS);
window.clearFlags(WindowManager.LayoutParams.FLAG_TRANSLUCENT_STATUS);
window.setStatusBarColor(0xff282733);
}
@Override
public boolean onCreateOptionsMenu(Menu menu) {
getMenuInflater().inflate(R.menu.menu_main, menu);
return true;
}
}
## Instruction:
Use plum and shadow_plum from colors.xml
## Code After:
package mozilla.org.webmaker.activity;
import android.app.ActionBar;
import android.content.res.Resources;
import android.graphics.Color;
import android.graphics.drawable.ColorDrawable;
import android.os.Bundle;
import android.view.Menu;
import android.view.Window;
import android.view.WindowManager;
import android.widget.RelativeLayout;
import mozilla.org.webmaker.R;
import mozilla.org.webmaker.WebmakerActivity;
import mozilla.org.webmaker.view.WebmakerWebView;
public class Tinker extends WebmakerActivity {
public Tinker() {
super("tinker", R.id.tinker_layout, R.layout.tinker_layout, R.menu.menu_tinker);
}
@Override
protected void onCreate(Bundle savedInstanceState) {
super.onCreate(savedInstanceState);
// Custom styles
Resources res = getResources();
int shadowPlum = res.getColor(R.color.shadow_plum);
int plum = res.getColor(R.color.plum);
ActionBar actionBar = getActionBar();
actionBar.setStackedBackgroundDrawable(plum);
actionBar.setBackgroundDrawable(plum);
Window window = getWindow();
window.addFlags(WindowManager.LayoutParams.FLAG_DRAWS_SYSTEM_BAR_BACKGROUNDS);
window.clearFlags(WindowManager.LayoutParams.FLAG_TRANSLUCENT_STATUS);
window.setStatusBarColor(shadowPlum);
}
@Override
public boolean onCreateOptionsMenu(Menu menu) {
getMenuInflater().inflate(R.menu.menu_main, menu);
return true;
}
}
| package mozilla.org.webmaker.activity;
import android.app.ActionBar;
+ import android.content.res.Resources;
import android.graphics.Color;
import android.graphics.drawable.ColorDrawable;
import android.os.Bundle;
import android.view.Menu;
import android.view.Window;
import android.view.WindowManager;
import android.widget.RelativeLayout;
import mozilla.org.webmaker.R;
import mozilla.org.webmaker.WebmakerActivity;
import mozilla.org.webmaker.view.WebmakerWebView;
public class Tinker extends WebmakerActivity {
public Tinker() {
super("tinker", R.id.tinker_layout, R.layout.tinker_layout, R.menu.menu_tinker);
}
@Override
protected void onCreate(Bundle savedInstanceState) {
super.onCreate(savedInstanceState);
// Custom styles
+ Resources res = getResources();
+ int shadowPlum = res.getColor(R.color.shadow_plum);
+ int plum = res.getColor(R.color.plum);
+
ActionBar actionBar = getActionBar();
- ColorDrawable colorOne = new ColorDrawable(Color.parseColor("#ff303250"));
- ColorDrawable colorTwo = new ColorDrawable(Color.parseColor("#ff303250"));
- actionBar.setStackedBackgroundDrawable(colorOne);
? ^^ ^^^^^
+ actionBar.setStackedBackgroundDrawable(plum);
? ^ ^^
- actionBar.setBackgroundDrawable(colorTwo);
? ^^ ^^^^^
+ actionBar.setBackgroundDrawable(plum);
? ^ ^^
Window window = getWindow();
window.addFlags(WindowManager.LayoutParams.FLAG_DRAWS_SYSTEM_BAR_BACKGROUNDS);
window.clearFlags(WindowManager.LayoutParams.FLAG_TRANSLUCENT_STATUS);
- window.setStatusBarColor(0xff282733);
? ^^^^^^^^^^
+ window.setStatusBarColor(shadowPlum);
? ^^^^^^^^^^
}
@Override
public boolean onCreateOptionsMenu(Menu menu) {
getMenuInflater().inflate(R.menu.menu_main, menu);
return true;
}
} | 13 | 0.309524 | 8 | 5 |
48f537e0a7ea31dc905ddefe8ac74ce8e6058512 | after/syntax/vimwiki.vim | after/syntax/vimwiki.vim | if exists("b:vtd_extra_highlights")
finish
endif
"syntax match doneItem '.*DONE.*'
"highlight link doneItem Ignore
"syntax region doneItem start=/\^\z(\s*\)\S.*DONE.*/ end=/write/
"syntax region doneItem start=/^\z(\s\+\)\S.*DONE/ end=/(^\z1\S)\@=/
"syntax region doneItem start='^\z(\s\+\)\S.*DONE' skip='^\z1\s.*$' end='^'
"syntax region doneItem start='^\z(\s*\).*DONE' skip='^\z1' end='^' skipnl
"highlight link doneItem Error
syntax region doneItem start='^\z(\s*\)\S.*DONE.*$' skip='^\z1\s' end='^'me=e-1
syntax region doneItem start='^\z(\s*\)\S.*WONTDO.*$' skip='^\z1\s' end='^'me=e-1
highlight link doneItem Ignore
let b:vtd_extra_highlights = 1
| if exists("b:vtd_extra_highlights")
finish
endif
"syntax match doneItem '.*DONE.*'
"highlight link doneItem Ignore
syntax region doneItem start='^\z(\s*\)\S.*DONE.*$' skip='^\z1\s' end='^'me=e-1
syntax region doneItem start='^\z(\s*\)\S.*WONTDO.*$' skip='^\z1\s' end='^'me=e-1
highlight link doneItem Ignore
let b:vtd_extra_highlights = 1
| Clean up DONE/WONTDO syntax highlighting code | Clean up DONE/WONTDO syntax highlighting code
| VimL | apache-2.0 | chiphogg/vim-vtd | viml | ## Code Before:
if exists("b:vtd_extra_highlights")
finish
endif
"syntax match doneItem '.*DONE.*'
"highlight link doneItem Ignore
"syntax region doneItem start=/\^\z(\s*\)\S.*DONE.*/ end=/write/
"syntax region doneItem start=/^\z(\s\+\)\S.*DONE/ end=/(^\z1\S)\@=/
"syntax region doneItem start='^\z(\s\+\)\S.*DONE' skip='^\z1\s.*$' end='^'
"syntax region doneItem start='^\z(\s*\).*DONE' skip='^\z1' end='^' skipnl
"highlight link doneItem Error
syntax region doneItem start='^\z(\s*\)\S.*DONE.*$' skip='^\z1\s' end='^'me=e-1
syntax region doneItem start='^\z(\s*\)\S.*WONTDO.*$' skip='^\z1\s' end='^'me=e-1
highlight link doneItem Ignore
let b:vtd_extra_highlights = 1
## Instruction:
Clean up DONE/WONTDO syntax highlighting code
## Code After:
if exists("b:vtd_extra_highlights")
finish
endif
"syntax match doneItem '.*DONE.*'
"highlight link doneItem Ignore
syntax region doneItem start='^\z(\s*\)\S.*DONE.*$' skip='^\z1\s' end='^'me=e-1
syntax region doneItem start='^\z(\s*\)\S.*WONTDO.*$' skip='^\z1\s' end='^'me=e-1
highlight link doneItem Ignore
let b:vtd_extra_highlights = 1
| if exists("b:vtd_extra_highlights")
finish
endif
"syntax match doneItem '.*DONE.*'
"highlight link doneItem Ignore
- "syntax region doneItem start=/\^\z(\s*\)\S.*DONE.*/ end=/write/
- "syntax region doneItem start=/^\z(\s\+\)\S.*DONE/ end=/(^\z1\S)\@=/
- "syntax region doneItem start='^\z(\s\+\)\S.*DONE' skip='^\z1\s.*$' end='^'
- "syntax region doneItem start='^\z(\s*\).*DONE' skip='^\z1' end='^' skipnl
- "highlight link doneItem Error
-
syntax region doneItem start='^\z(\s*\)\S.*DONE.*$' skip='^\z1\s' end='^'me=e-1
syntax region doneItem start='^\z(\s*\)\S.*WONTDO.*$' skip='^\z1\s' end='^'me=e-1
highlight link doneItem Ignore
let b:vtd_extra_highlights = 1 | 6 | 0.333333 | 0 | 6 |
252d0e05ff9813de1b1c9fb59b6b82c4655d1b95 | app/controllers/procurement_areas_controller.rb | app/controllers/procurement_areas_controller.rb | class ProcurementAreasController < ApplicationController
def index
@procurement_areas = ProcurementArea.ordered_by_name
end
def new
@procurement_area = ProcurementArea.new
end
def create
@procurement_area = ProcurementArea.new(procurement_area_params)
if @procurement_area.save
redirect_to procurement_areas_path
else
render :new
end
end
def edit
@procurement_area = procurement_area
end
def update
@procurement_area = procurement_area
if @procurement_area.update_attributes(procurement_area_params)
redirect_to procurement_areas_path
else
render :edit
end
end
def destroy
if procurement_area.destroy
redirect_to procurement_areas_path
end
end
private
def procurement_area_params
params.require(:procurement_area).permit(:name)
end
def procurement_area
@_procurement_area ||= ProcurementArea.find(params[:id])
end
end
| class ProcurementAreasController < ApplicationController
def index
@procurement_areas = ProcurementArea.ordered_by_name
end
def new
@procurement_area = ProcurementArea.new
end
def create
@procurement_area = ProcurementArea.new(procurement_area_params)
if @procurement_area.save
redirect_to procurement_areas_path
else
render :new
end
end
def edit
@procurement_area = procurement_area
end
def update
@procurement_area = procurement_area
if @procurement_area.update_attributes(procurement_area_params)
redirect_to procurement_areas_path
else
render :edit
end
end
def destroy
procurement_area.destroy
redirect_to procurement_areas_path
end
private
def procurement_area_params
params.require(:procurement_area).permit(:name)
end
def procurement_area
@_procurement_area ||= ProcurementArea.find(params[:id])
end
end
| Remove conditional around destroy action | Remove conditional around destroy action
| Ruby | mit | ministryofjustice/defence-request-service-rota,ministryofjustice/defence-request-service-rota,ministryofjustice/defence-request-service-rota,ministryofjustice/defence-request-service-rota | ruby | ## Code Before:
class ProcurementAreasController < ApplicationController
def index
@procurement_areas = ProcurementArea.ordered_by_name
end
def new
@procurement_area = ProcurementArea.new
end
def create
@procurement_area = ProcurementArea.new(procurement_area_params)
if @procurement_area.save
redirect_to procurement_areas_path
else
render :new
end
end
def edit
@procurement_area = procurement_area
end
def update
@procurement_area = procurement_area
if @procurement_area.update_attributes(procurement_area_params)
redirect_to procurement_areas_path
else
render :edit
end
end
def destroy
if procurement_area.destroy
redirect_to procurement_areas_path
end
end
private
def procurement_area_params
params.require(:procurement_area).permit(:name)
end
def procurement_area
@_procurement_area ||= ProcurementArea.find(params[:id])
end
end
## Instruction:
Remove conditional around destroy action
## Code After:
class ProcurementAreasController < ApplicationController
def index
@procurement_areas = ProcurementArea.ordered_by_name
end
def new
@procurement_area = ProcurementArea.new
end
def create
@procurement_area = ProcurementArea.new(procurement_area_params)
if @procurement_area.save
redirect_to procurement_areas_path
else
render :new
end
end
def edit
@procurement_area = procurement_area
end
def update
@procurement_area = procurement_area
if @procurement_area.update_attributes(procurement_area_params)
redirect_to procurement_areas_path
else
render :edit
end
end
def destroy
procurement_area.destroy
redirect_to procurement_areas_path
end
private
def procurement_area_params
params.require(:procurement_area).permit(:name)
end
def procurement_area
@_procurement_area ||= ProcurementArea.find(params[:id])
end
end
| class ProcurementAreasController < ApplicationController
def index
@procurement_areas = ProcurementArea.ordered_by_name
end
def new
@procurement_area = ProcurementArea.new
end
def create
@procurement_area = ProcurementArea.new(procurement_area_params)
if @procurement_area.save
redirect_to procurement_areas_path
else
render :new
end
end
def edit
@procurement_area = procurement_area
end
def update
@procurement_area = procurement_area
if @procurement_area.update_attributes(procurement_area_params)
redirect_to procurement_areas_path
else
render :edit
end
end
def destroy
- if procurement_area.destroy
? ---
+ procurement_area.destroy
+
- redirect_to procurement_areas_path
? --
+ redirect_to procurement_areas_path
- end
end
private
def procurement_area_params
params.require(:procurement_area).permit(:name)
end
def procurement_area
@_procurement_area ||= ProcurementArea.find(params[:id])
end
end | 6 | 0.122449 | 3 | 3 |
a25ce6a3fad5f617d83e96da84e92d9f80b76836 | .travis.yml | .travis.yml | sudo: false
os:
- linux
- osx
language: python
python:
- "2.7"
- "3.3"
- "3.4"
- "3.5"
- "pypy"
- "pypy3"
env:
- DJANGO="django1.6"
- DJANGO="django1.7"
- DJANGO="django1.8"
- DJANGO="django1.9"
install:
- pip install tox --use-mirrors
script:
- tox -e ${DJANGO}
| sudo: false
os:
- linux
- osx
language: python
python:
- "2.7"
- "3.3"
- "3.4"
- "3.5"
- "pypy"
- "pypy3"
env:
- DJANGO="django1.6"
- DJANGO="django1.7"
- DJANGO="django1.8"
- DJANGO="django1.9"
install:
- pip install tox
script:
- tox -e ${DJANGO}
| Fix issue where `pip install --use-mirrors ...` doesn't work in Python 3.5 | Fix issue where `pip install --use-mirrors ...` doesn't work in Python 3.5
| YAML | bsd-2-clause | pombredanne/django-classy-settings,funkybob/django-classy-settings | yaml | ## Code Before:
sudo: false
os:
- linux
- osx
language: python
python:
- "2.7"
- "3.3"
- "3.4"
- "3.5"
- "pypy"
- "pypy3"
env:
- DJANGO="django1.6"
- DJANGO="django1.7"
- DJANGO="django1.8"
- DJANGO="django1.9"
install:
- pip install tox --use-mirrors
script:
- tox -e ${DJANGO}
## Instruction:
Fix issue where `pip install --use-mirrors ...` doesn't work in Python 3.5
## Code After:
sudo: false
os:
- linux
- osx
language: python
python:
- "2.7"
- "3.3"
- "3.4"
- "3.5"
- "pypy"
- "pypy3"
env:
- DJANGO="django1.6"
- DJANGO="django1.7"
- DJANGO="django1.8"
- DJANGO="django1.9"
install:
- pip install tox
script:
- tox -e ${DJANGO}
| sudo: false
os:
- linux
- osx
language: python
python:
- "2.7"
- "3.3"
- "3.4"
- "3.5"
- "pypy"
- "pypy3"
env:
- DJANGO="django1.6"
- DJANGO="django1.7"
- DJANGO="django1.8"
- DJANGO="django1.9"
install:
- - pip install tox --use-mirrors
+ - pip install tox
script:
- tox -e ${DJANGO} | 2 | 0.074074 | 1 | 1 |
be6d4d2ffb461a131cdeba6014ca9ac02bddc9d5 | lib/facter/package_provider.rb | lib/facter/package_provider.rb | require 'puppet/type'
require 'puppet/type/package'
Facter.add(:package_provider) do
setcode do
if Gem::Version.new(Facter.value(:puppetversion).split(' ')[0]) >= Gem::Version.new('3.6')
Puppet::Type.type(:package).newpackage(:name => 'dummy', :allow_virtual => 'true')[:provider].to_s
else
Puppet::Type.type(:package).newpackage(:name => 'dummy')[:provider].to_s
end
end
end
| require 'puppet/type'
require 'puppet/type/package'
Facter.add(:package_provider) do
setcode do
if defined? Gem and Gem::Version.new(Facter.value(:puppetversion).split(' ')[0]) >= Gem::Version.new('3.6')
Puppet::Type.type(:package).newpackage(:name => 'dummy', :allow_virtual => 'true')[:provider].to_s
else
Puppet::Type.type(:package).newpackage(:name => 'dummy')[:provider].to_s
end
end
end
| Add check if Gem is defined | Add check if Gem is defined
On e.g. Ubuntu 12.04 LTS Gem is not there by default so i added a check to not fail in that fact if this is the case. | Ruby | apache-2.0 | puppetlabs/puppetlabs-stdlib,fuel-infra/puppetlabs-stdlib,DavidS/puppetlabs-stdlib,jbondpdx/puppetlabs-stdlib,HelenCampbell/puppetlabs-stdlib | ruby | ## Code Before:
require 'puppet/type'
require 'puppet/type/package'
Facter.add(:package_provider) do
setcode do
if Gem::Version.new(Facter.value(:puppetversion).split(' ')[0]) >= Gem::Version.new('3.6')
Puppet::Type.type(:package).newpackage(:name => 'dummy', :allow_virtual => 'true')[:provider].to_s
else
Puppet::Type.type(:package).newpackage(:name => 'dummy')[:provider].to_s
end
end
end
## Instruction:
Add check if Gem is defined
On e.g. Ubuntu 12.04 LTS Gem is not there by default so i added a check to not fail in that fact if this is the case.
## Code After:
require 'puppet/type'
require 'puppet/type/package'
Facter.add(:package_provider) do
setcode do
if defined? Gem and Gem::Version.new(Facter.value(:puppetversion).split(' ')[0]) >= Gem::Version.new('3.6')
Puppet::Type.type(:package).newpackage(:name => 'dummy', :allow_virtual => 'true')[:provider].to_s
else
Puppet::Type.type(:package).newpackage(:name => 'dummy')[:provider].to_s
end
end
end
| require 'puppet/type'
require 'puppet/type/package'
Facter.add(:package_provider) do
setcode do
- if Gem::Version.new(Facter.value(:puppetversion).split(' ')[0]) >= Gem::Version.new('3.6')
+ if defined? Gem and Gem::Version.new(Facter.value(:puppetversion).split(' ')[0]) >= Gem::Version.new('3.6')
? +++++++++++++++++
Puppet::Type.type(:package).newpackage(:name => 'dummy', :allow_virtual => 'true')[:provider].to_s
else
Puppet::Type.type(:package).newpackage(:name => 'dummy')[:provider].to_s
end
end
end | 2 | 0.166667 | 1 | 1 |
799504a3e27f665fc6a5acdf877f88b416191460 | doc/schema.rst | doc/schema.rst | .. _schema:
Schema
======
.. toctree::
:maxdepth: 1
schema-job
schema-defconfig
schema-boot
| .. _schema:
Schema
======
.. toctree::
:maxdepth: 1
schema-job
schema-defconfig
schema-boot
schema-token
| Add missing doc in index. | doc: Add missing doc in index.
Change-Id: I31cb99e3cf9c5ced408edc8bbec4e15bd2e60f19
| reStructuredText | agpl-3.0 | joyxu/kernelci-backend,joyxu/kernelci-backend,kernelci/kernelci-backend,joyxu/kernelci-backend,kernelci/kernelci-backend | restructuredtext | ## Code Before:
.. _schema:
Schema
======
.. toctree::
:maxdepth: 1
schema-job
schema-defconfig
schema-boot
## Instruction:
doc: Add missing doc in index.
Change-Id: I31cb99e3cf9c5ced408edc8bbec4e15bd2e60f19
## Code After:
.. _schema:
Schema
======
.. toctree::
:maxdepth: 1
schema-job
schema-defconfig
schema-boot
schema-token
| .. _schema:
Schema
======
.. toctree::
:maxdepth: 1
schema-job
schema-defconfig
schema-boot
+ schema-token | 1 | 0.090909 | 1 | 0 |
e92b9eabecbb7fb90fc601f356d962fefe4f22c6 | _topics/topic-FAQ.md | _topics/topic-FAQ.md | ---
layout: topic
title: Frequently asked questions
author: Willy McAllister
comments: true
---
I get a questions about electrical engineering that are interesting but don't have a natural place to live. So they camp out here.
----
 [What's the difference between impedance and resistance?]({{ site.baseurl }}{% link _articles/difference-between-Z-and-R.md %})
 [Inductor - how it works]({{ site.baseurl }}{% link _articles/inductor-how-it-works.md %})
 [6 ways to generate voltage]({{ site.baseurl }}{% link _articles/six-ways-to-generate-voltage.md %})
| ---
layout: topic
title: Frequently asked questions
author: Willy McAllister
comments: true
---
I get a questions about electrical engineering that are interesting but don't have a natural place to live. So they camp out here for now.
----
 [What's the difference between impedance and resistance?]({{ site.baseurl }}{% link _articles/difference-between-Z-and-R.md %})
 [Inductor - how it works]({{ site.baseurl }}{% link _articles/inductor-how-it-works.md %})
 [6 ways to generate electricity]({{ site.baseurl }}{% link _articles/six-ways-to-generate-electricity.md %})
| Update link to 6 ways article. | Update link to 6 ways article.
| Markdown | mit | willymcallister/willymcallister.github.io,willymcallister/spinningnumbers,willymcallister/spinningnumbers,willymcallister/willymcallister.github.io,willymcallister/willymcallister.github.io,willymcallister/spinningnumbers | markdown | ## Code Before:
---
layout: topic
title: Frequently asked questions
author: Willy McAllister
comments: true
---
I get a questions about electrical engineering that are interesting but don't have a natural place to live. So they camp out here.
----
 [What's the difference between impedance and resistance?]({{ site.baseurl }}{% link _articles/difference-between-Z-and-R.md %})
 [Inductor - how it works]({{ site.baseurl }}{% link _articles/inductor-how-it-works.md %})
 [6 ways to generate voltage]({{ site.baseurl }}{% link _articles/six-ways-to-generate-voltage.md %})
## Instruction:
Update link to 6 ways article.
## Code After:
---
layout: topic
title: Frequently asked questions
author: Willy McAllister
comments: true
---
I get a questions about electrical engineering that are interesting but don't have a natural place to live. So they camp out here for now.
----
 [What's the difference between impedance and resistance?]({{ site.baseurl }}{% link _articles/difference-between-Z-and-R.md %})
 [Inductor - how it works]({{ site.baseurl }}{% link _articles/inductor-how-it-works.md %})
 [6 ways to generate electricity]({{ site.baseurl }}{% link _articles/six-ways-to-generate-electricity.md %})
| ---
layout: topic
title: Frequently asked questions
author: Willy McAllister
comments: true
---
- I get a questions about electrical engineering that are interesting but don't have a natural place to live. So they camp out here.
+ I get a questions about electrical engineering that are interesting but don't have a natural place to live. So they camp out here for now.
? ++++++++
----
 [What's the difference between impedance and resistance?]({{ site.baseurl }}{% link _articles/difference-between-Z-and-R.md %})
 [Inductor - how it works]({{ site.baseurl }}{% link _articles/inductor-how-it-works.md %})
-  [6 ways to generate voltage]({{ site.baseurl }}{% link _articles/six-ways-to-generate-voltage.md %})
? ^^ ^^^ ^^ ^^^
+  [6 ways to generate electricity]({{ site.baseurl }}{% link _articles/six-ways-to-generate-electricity.md %})
? ^ ++ ^^^^^^ ^ ++ ^^^^^^
| 4 | 0.222222 | 2 | 2 |
4372e243030670d7a4c23dfbd10a514e2328508a | top.sls | top.sls | base:
'*':
- iptables
- logcheck
- logrotate
- monit
- motd
- ntp
- packages
- pam
- postfix
- salt.common
- salt.minion
- ssh
- sudo
- timezone
- users
'os_family:Debian':
- match: grain
- apt
'os:Debian':
- match: grain
- debsecan
- apticron
'helium.djl.io':
- apache
- fstab
- madsonic
- sabnzbd
- sickbeard
- smartmontools
- unbound
'iron.djl.io':
- apache
- php
- salt.master
- znc
'neon.djl.io':
- fstab
- xbmc
| base:
'*':
- iptables
- logcheck
- logrotate
- monit
- motd
- ntp
- packages
- pam
- postfix
- salt.common
- salt.minion
- ssh
- sudo
- timezone
- users
'os_family:Debian':
- match: grain
- apt
'os:Debian':
- match: grain
- debsecan
- apticron
'helium.djl.io':
- apache
- fstab
- madsonic
- sabnzbd
- sickbeard
- smartmontools
- unbound
'iron.djl.io':
- apache
- mysql.server
- mysql.client
- mysql.database
- mysql.user
- php
- salt.master
- znc
'neon.djl.io':
- fstab
- xbmc
| Bring back MySQL on iron. | Bring back MySQL on iron.
| SaltStack | bsd-3-clause | h4ck3rm1k3/states-2,h4ck3rm1k3/states-2 | saltstack | ## Code Before:
base:
'*':
- iptables
- logcheck
- logrotate
- monit
- motd
- ntp
- packages
- pam
- postfix
- salt.common
- salt.minion
- ssh
- sudo
- timezone
- users
'os_family:Debian':
- match: grain
- apt
'os:Debian':
- match: grain
- debsecan
- apticron
'helium.djl.io':
- apache
- fstab
- madsonic
- sabnzbd
- sickbeard
- smartmontools
- unbound
'iron.djl.io':
- apache
- php
- salt.master
- znc
'neon.djl.io':
- fstab
- xbmc
## Instruction:
Bring back MySQL on iron.
## Code After:
base:
'*':
- iptables
- logcheck
- logrotate
- monit
- motd
- ntp
- packages
- pam
- postfix
- salt.common
- salt.minion
- ssh
- sudo
- timezone
- users
'os_family:Debian':
- match: grain
- apt
'os:Debian':
- match: grain
- debsecan
- apticron
'helium.djl.io':
- apache
- fstab
- madsonic
- sabnzbd
- sickbeard
- smartmontools
- unbound
'iron.djl.io':
- apache
- mysql.server
- mysql.client
- mysql.database
- mysql.user
- php
- salt.master
- znc
'neon.djl.io':
- fstab
- xbmc
| base:
'*':
- iptables
- logcheck
- logrotate
- monit
- motd
- ntp
- packages
- pam
- postfix
- salt.common
- salt.minion
- ssh
- sudo
- timezone
- users
'os_family:Debian':
- match: grain
- apt
'os:Debian':
- match: grain
- debsecan
- apticron
'helium.djl.io':
- apache
- fstab
- madsonic
- sabnzbd
- sickbeard
- smartmontools
- unbound
'iron.djl.io':
- apache
+ - mysql.server
+ - mysql.client
+ - mysql.database
+ - mysql.user
- php
- salt.master
- znc
'neon.djl.io':
- fstab
- xbmc | 4 | 0.1 | 4 | 0 |
eb9689e09b6880c8b68b330aca61e306aa119196 | .travis.yml | .travis.yml | language: c
os:
- linux
- osx
install:
- if [ "${TRAVIS_OS_NAME}" = "linux" ]; then ./travis.linux.install.deps.sh; fi
- if [ "${TRAVIS_OS_NAME}" = "osx" ]; then ./travis.osx.install.deps.sh; fi
before_script:
- mkdir -p build
script:
- if [ "${TRAVIS_OS_NAME}" = "linux" ]; then ./travis.linux.script.sh; fi
- if [ "${TRAVIS_OS_NAME}" = "osx" ]; then ./travis.osx.script.sh; fi
git:
submodules: true
notifications:
irc:
channels: "irc.mozilla.org#servo"
on_success: change
on_failure: change
use_notice: true
email: false
branches:
only:
- master
| language: c
os:
- linux
- osx
install:
- if [ "${TRAVIS_OS_NAME}" = "linux" ]; then ./travis.linux.install.deps.sh; fi
- if [ "${TRAVIS_OS_NAME}" = "osx" ]; then ./travis.osx.install.deps.sh; fi
before_script:
- mkdir -p build
script:
- if [ "${TRAVIS_OS_NAME}" = "linux" ]; then ./travis.linux.script.sh; fi
- if [ "${TRAVIS_OS_NAME}" = "osx" ]; then ./travis.osx.script.sh; fi
git:
submodules: true
notifications:
irc:
channels: "irc.mozilla.org#servo"
on_success: change
on_failure: change
use_notice: true
branches:
only:
- master
| Enable email notifications from Travis CI | Enable email notifications from Travis CI
http://docs.travis-ci.com/user/notifications/
> By default, email notifications will be sent to the committer and the commit author, if they are members of the repository (that is, they have push or admin permissions for public repositories, or if they have pull, push or admin permissions for private repositories).
>
> And it will by default send emails when, on the given branch:
>
> * a build was just broken or still is broken
> * a previously broken build was just fixed
| YAML | mpl-2.0 | KiChjang/servo,GyrosOfWar/servo,hyowon/servo,pyfisch/servo,emilio/servo,runarberg/servo,emilio/servo,aweinstock314/servo,meh/servo,SimonSapin/servo,upsuper/servo,srbhklkrn/SERVOENGINE,boghison/servo,rnestler/servo,mt2d2/servo,jgraham/servo,Shraddha512/servo,nnethercote/servo,chotchki/servo,caldwell/servo,paulrouget/servo,rnestler/servo,eddyb/servo,nick-thompson/servo,chotchki/servo,GyrosOfWar/servo,eddyb/servo,peterjoel/servo,aidanhs/servo,cbrewster/servo,GreenRecycleBin/servo,nerith/servo,jlegendary/servo,saneyuki/servo,canaltinova/servo,bfrohs/servo,ConnorGBrewster/servo,jimberlage/servo,GreenRecycleBin/servo,canaltinova/servo,mdibaiee/servo,cbrewster/servo,jimberlage/servo,fiji-flo/servo,dsandeephegde/servo,youprofit/servo,Adenilson/prototype-viewing-distance,evilpie/servo,KiChjang/servo,upsuper/servo,avadacatavra/servo,akosel/servo,mdibaiee/servo,thiagopnts/servo,dvberkel/servo,notriddle/servo,peterjoel/servo,michaelwu/servo,fiji-flo/servo,Adenilson/prototype-viewing-distance,mattnenterprise/servo,thiagopnts/servo,szeged/servo,indykish/servo,bjwbell/servo,srbhklkrn/SERVOENGINE,samfoo/servo,saratang/servo,srbhklkrn/SERVOENGINE,evilpie/servo,anthgur/servo,splav/servo,CJ8664/servo,Shraddha512/servo,jlegendary/servo,deokjinkim/servo,saneyuki/servo,wartman4404/servo,akosel/servo,codemac/servo,nick-thompson/servo,vks/servo,karlito40/servo,pyfisch/servo,rnestler/servo,michaelwu/servo,caldwell/servo,sadmansk/servo,dagnir/servo,dagnir/servo,juzer10/servo,dvberkel/servo,dsandeephegde/servo,samfoo/servo,mdibaiee/servo,mukilan/servo,g-k/servo,sadmansk/servo,mbrubeck/servo,peterjoel/servo,thiagopnts/servo,paulrouget/servo,jlegendary/servo,SimonSapin/servo,evilpie/servo,steveklabnik/servo,youprofit/servo,chotchki/servo,codemac/servo,AnthonyBroadCrawford/servo,aidanhs/servo,Adenilson/prototype-viewing-distance,boghison/servo,juzer10/servo,tafia/servo,caldwell/servo,cbrewster/servo,jlegendary/servo,dsandeephegde/servo,paulrouget/servo,indykish/servo,Adenilson/prototype-viewing-distance,larsbergstrom/servo,nnethercote/servo,tafia/servo,ConnorGBrewster/servo,szeged/servo,karlito40/servo,rixrix/servo,mt2d2/servo,kindersung/servo,ruud-v-a/servo,s142857/servo,notriddle/servo,rentongzhang/servo,aidanhs/servo,jimberlage/servo,samfoo/servo,WriterOfAlicrow/servo,eddyb/servo,shrenikgala/servo,thiagopnts/servo,KiChjang/servo,nrc/servo,canaltinova/servo,s142857/servo,rnestler/servo,ConnorGBrewster/servo,A-deLuna/servo,GyrosOfWar/servo,dvberkel/servo,kindersung/servo,snf/servo,brendandahl/servo,tschneidereit/servo,KiChjang/servo,nrc/servo,walac/servo,codemac/servo,rixrix/servo,michaelwu/servo,tschneidereit/servo,A-deLuna/servo,sadmansk/servo,fiji-flo/servo,huonw/servo,indykish/servo,larsbergstrom/servo,aweinstock314/servo,snf/servo,srbhklkrn/SERVOENGINE,pyfisch/servo,dsandeephegde/servo,peterjoel/servo,zhangjunlei26/servo,mdibaiee/servo,splav/servo,A-deLuna/servo,KiChjang/servo,aidanhs/servo,jimberlage/servo,tempbottle/servo,brendandahl/servo,kindersung/servo,rnestler/servo,akosel/servo,mukilan/servo,nrc/servo,dsandeephegde/servo,nrc/servo,upsuper/servo,karlito40/servo,Shraddha512/servo,aweinstock314/servo,saneyuki/servo,pgonda/servo,avadacatavra/servo,rentongzhang/servo,DominoTree/servo,nerith/servo,saratang/servo,zentner-kyle/servo,wpgallih/servo,tschneidereit/servo,s142857/servo,mbrubeck/servo,emilio/servo,mbrubeck/servo,szeged/servo,mt2d2/servo,GyrosOfWar/servo,jlegendary/servo,shrenikgala/servo,saratang/servo,indykish/servo,splav/servo,saneyuki/servo,Shraddha512/servo,tempbottle/servo,snf/servo,evilpie/servo,emilio/servo,nerith/servo,notriddle/servo,avadacatavra/servo,dati91/servo,karlito40/servo,notriddle/servo,huonw/servo,luniv/servo,nerith/servo,cbrewster/servo,karlito40/servo,juzer10/servo,zentner-kyle/servo,juzer10/servo,aweinstock314/servo,jdramani/servo,rixrix/servo,rixrix/servo,j3parker/servo,froydnj/servo,avadacatavra/servo,indykish/servo,RenaudParis/servo,meh/servo,Adenilson/prototype-viewing-distance,dati91/servo,kindersung/servo,anthgur/servo,deokjinkim/servo,pyecs/servo,bfrohs/servo,fiji-flo/servo,ruud-v-a/servo,hyowon/servo,szeged/servo,larsbergstrom/servo,anthgur/servo,eddyb/servo,jdramani/servo,indykish/servo,dhananjay92/servo,CJ8664/servo,g-k/servo,pyfisch/servo,codemac/servo,tafia/servo,tafia/servo,rixrix/servo,steveklabnik/servo,snf/servo,emilio/servo,wpgallih/servo,bfrohs/servo,runarberg/servo,nrc/servo,rixrix/servo,ConnorGBrewster/servo,thiagopnts/servo,ryancanhelpyou/servo,steveklabnik/servo,steveklabnik/servo,DominoTree/servo,indykish/servo,mattnenterprise/servo,nrc/servo,splav/servo,hyowon/servo,zentner-kyle/servo,sadmansk/servo,mukilan/servo,snf/servo,s142857/servo,avadacatavra/servo,vks/servo,emilio/servo,saneyuki/servo,paulrouget/servo,fiji-flo/servo,SimonSapin/servo,ruud-v-a/servo,nnethercote/servo,A-deLuna/servo,nnethercote/servo,wartman4404/servo,boghison/servo,mattnenterprise/servo,ruud-v-a/servo,jdramani/servo,jimberlage/servo,pgonda/servo,srbhklkrn/SERVOENGINE,mt2d2/servo,KiChjang/servo,anthgur/servo,aweinstock314/servo,dvberkel/servo,szeged/servo,tempbottle/servo,WriterOfAlicrow/servo,pyecs/servo,boghison/servo,mukilan/servo,caldwell/servo,GreenRecycleBin/servo,wpgallih/servo,szeged/servo,RenaudParis/servo,g-k/servo,A-deLuna/servo,nnethercote/servo,vks/servo,tschneidereit/servo,walac/servo,luniv/servo,szeged/servo,boghison/servo,larsbergstrom/servo,rentongzhang/servo,DominoTree/servo,Shraddha512/servo,mbrubeck/servo,splav/servo,indykish/servo,saratang/servo,larsbergstrom/servo,echochamber/servo,ryancanhelpyou/servo,ruud-v-a/servo,zentner-kyle/servo,froydnj/servo,zhangjunlei26/servo,splav/servo,KiChjang/servo,fiji-flo/servo,SimonSapin/servo,tschneidereit/servo,RenaudParis/servo,szeged/servo,mattnenterprise/servo,Shraddha512/servo,dmarcos/servo,DominoTree/servo,sadmansk/servo,mattnenterprise/servo,ruud-v-a/servo,brendandahl/servo,RenaudParis/servo,dati91/servo,rentongzhang/servo,akosel/servo,nnethercote/servo,GyrosOfWar/servo,caldwell/servo,dhananjay92/servo,dsandeephegde/servo,saneyuki/servo,caldwell/servo,ryancanhelpyou/servo,snf/servo,avadacatavra/servo,mbrubeck/servo,luniv/servo,s142857/servo,zhangjunlei26/servo,ryancanhelpyou/servo,AnthonyBroadCrawford/servo,paulrouget/servo,vks/servo,RenaudParis/servo,evilpie/servo,jgraham/servo,deokjinkim/servo,j3parker/servo,WriterOfAlicrow/servo,notriddle/servo,szeged/servo,peterjoel/servo,Adenilson/prototype-viewing-distance,WriterOfAlicrow/servo,wartman4404/servo,peterjoel/servo,tempbottle/servo,luniv/servo,eddyb/servo,tempbottle/servo,evilpie/servo,g-k/servo,DominoTree/servo,dvberkel/servo,saneyuki/servo,juzer10/servo,brendandahl/servo,upsuper/servo,cbrewster/servo,wpgallih/servo,saneyuki/servo,zentner-kyle/servo,kindersung/servo,CJ8664/servo,pyecs/servo,hyowon/servo,SimonSapin/servo,karlito40/servo,rentongzhang/servo,huonw/servo,tafia/servo,A-deLuna/servo,dagnir/servo,zhangjunlei26/servo,DominoTree/servo,splav/servo,akosel/servo,youprofit/servo,wpgallih/servo,akosel/servo,huonw/servo,canaltinova/servo,samfoo/servo,AnthonyBroadCrawford/servo,bfrohs/servo,peterjoel/servo,GreenRecycleBin/servo,jdramani/servo,dagnir/servo,saratang/servo,samfoo/servo,bjwbell/servo,nnethercote/servo,jgraham/servo,GreenRecycleBin/servo,cbrewster/servo,aidanhs/servo,avadacatavra/servo,indykish/servo,eddyb/servo,michaelwu/servo,echochamber/servo,Adenilson/prototype-viewing-distance,mbrubeck/servo,evilpie/servo,dhananjay92/servo,brendandahl/servo,juzer10/servo,bfrohs/servo,wpgallih/servo,mdibaiee/servo,saneyuki/servo,rixrix/servo,froydnj/servo,jgraham/servo,akosel/servo,pgonda/servo,avadacatavra/servo,KiChjang/servo,SimonSapin/servo,AnthonyBroadCrawford/servo,notriddle/servo,jgraham/servo,froydnj/servo,steveklabnik/servo,chotchki/servo,youprofit/servo,tafia/servo,kindersung/servo,dhananjay92/servo,j3parker/servo,jimberlage/servo,wartman4404/servo,splav/servo,walac/servo,nick-thompson/servo,tafia/servo,AnthonyBroadCrawford/servo,cbrewster/servo,thiagopnts/servo,echochamber/servo,thiagopnts/servo,echochamber/servo,larsbergstrom/servo,nerith/servo,ConnorGBrewster/servo,kindersung/servo,brendandahl/servo,paulrouget/servo,shrenikgala/servo,canaltinova/servo,tempbottle/servo,deokjinkim/servo,j3parker/servo,emilio/servo,dagnir/servo,sadmansk/servo,pyfisch/servo,dmarcos/servo,huonw/servo,dmarcos/servo,shrenikgala/servo,dati91/servo,mt2d2/servo,anthgur/servo,ryancanhelpyou/servo,sadmansk/servo,youprofit/servo,runarberg/servo,canaltinova/servo,Shraddha512/servo,jlegendary/servo,canaltinova/servo,pyecs/servo,vks/servo,canaltinova/servo,nnethercote/servo,saratang/servo,dagnir/servo,s142857/servo,vks/servo,pgonda/servo,bjwbell/servo,rixrix/servo,upsuper/servo,peterjoel/servo,paulrouget/servo,wartman4404/servo,pyfisch/servo,deokjinkim/servo,deokjinkim/servo,GreenRecycleBin/servo,walac/servo,GreenRecycleBin/servo,notriddle/servo,anthgur/servo,DominoTree/servo,ConnorGBrewster/servo,WriterOfAlicrow/servo,mattnenterprise/servo,nrc/servo,nick-thompson/servo,eddyb/servo,luniv/servo,wpgallih/servo,paulrouget/servo,vks/servo,codemac/servo,splav/servo,dhananjay92/servo,echochamber/servo,notriddle/servo,dsandeephegde/servo,boghison/servo,runarberg/servo,ConnorGBrewster/servo,dati91/servo,huonw/servo,mattnenterprise/servo,mbrubeck/servo,g-k/servo,chotchki/servo,luniv/servo,wpgallih/servo,zentner-kyle/servo,RenaudParis/servo,juzer10/servo,SimonSapin/servo,larsbergstrom/servo,A-deLuna/servo,pgonda/servo,dmarcos/servo,ConnorGBrewster/servo,srbhklkrn/SERVOENGINE,jdramani/servo,j3parker/servo,zhangjunlei26/servo,mattnenterprise/servo,rentongzhang/servo,paulrouget/servo,eddyb/servo,meh/servo,dhananjay92/servo,WriterOfAlicrow/servo,ryancanhelpyou/servo,bjwbell/servo,luniv/servo,DominoTree/servo,boghison/servo,wartman4404/servo,szeged/servo,tempbottle/servo,anthgur/servo,dagnir/servo,upsuper/servo,meh/servo,mt2d2/servo,j3parker/servo,pyecs/servo,meh/servo,mdibaiee/servo,rnestler/servo,thiagopnts/servo,CJ8664/servo,cbrewster/servo,AnthonyBroadCrawford/servo,froydnj/servo,dmarcos/servo,rnestler/servo,pyfisch/servo,pyfisch/servo,nerith/servo,rnestler/servo,mukilan/servo,rixrix/servo,walac/servo,zhangjunlei26/servo,KiChjang/servo,peterjoel/servo,shrenikgala/servo,mbrubeck/servo,sadmansk/servo,nick-thompson/servo,jimberlage/servo,jgraham/servo,nick-thompson/servo,dmarcos/servo,michaelwu/servo,wpgallih/servo,g-k/servo,dati91/servo,CJ8664/servo,jdramani/servo,KiChjang/servo,notriddle/servo,echochamber/servo,runarberg/servo,jimberlage/servo,AnthonyBroadCrawford/servo,emilio/servo,nick-thompson/servo,emilio/servo,aweinstock314/servo,larsbergstrom/servo,zhangjunlei26/servo,anthgur/servo,hyowon/servo,caldwell/servo,evilpie/servo,dvberkel/servo,wartman4404/servo,CJ8664/servo,dhananjay92/servo,WriterOfAlicrow/servo,DominoTree/servo,saneyuki/servo,dsandeephegde/servo,GyrosOfWar/servo,nnethercote/servo,pyecs/servo,upsuper/servo,tschneidereit/servo,aidanhs/servo,fiji-flo/servo,walac/servo,nnethercote/servo,jimberlage/servo,upsuper/servo,deokjinkim/servo,pgonda/servo,chotchki/servo,bjwbell/servo,samfoo/servo,codemac/servo,snf/servo,huonw/servo,zentner-kyle/servo,nerith/servo,jlegendary/servo,hyowon/servo,froydnj/servo,emilio/servo,saratang/servo,mdibaiee/servo,fiji-flo/servo,michaelwu/servo,SimonSapin/servo,meh/servo,karlito40/servo,GreenRecycleBin/servo,jdramani/servo,splav/servo,zhangjunlei26/servo,dmarcos/servo,aidanhs/servo,tschneidereit/servo,steveklabnik/servo,larsbergstrom/servo,pyecs/servo,dvberkel/servo,srbhklkrn/SERVOENGINE,brendandahl/servo,bjwbell/servo,pyfisch/servo,larsbergstrom/servo,dati91/servo,shrenikgala/servo,bfrohs/servo,DominoTree/servo,jgraham/servo,meh/servo,GyrosOfWar/servo,pgonda/servo,samfoo/servo,shrenikgala/servo,runarberg/servo,rentongzhang/servo,zhangjunlei26/servo,aweinstock314/servo,paulrouget/servo,mt2d2/servo,echochamber/servo,bjwbell/servo,CJ8664/servo,walac/servo,g-k/servo,hyowon/servo,youprofit/servo,GreenRecycleBin/servo,RenaudParis/servo,youprofit/servo,codemac/servo,CJ8664/servo,s142857/servo,dati91/servo,runarberg/servo,notriddle/servo,steveklabnik/servo,mukilan/servo,ryancanhelpyou/servo,pyfisch/servo,j3parker/servo,evilpie/servo,peterjoel/servo | yaml | ## Code Before:
language: c
os:
- linux
- osx
install:
- if [ "${TRAVIS_OS_NAME}" = "linux" ]; then ./travis.linux.install.deps.sh; fi
- if [ "${TRAVIS_OS_NAME}" = "osx" ]; then ./travis.osx.install.deps.sh; fi
before_script:
- mkdir -p build
script:
- if [ "${TRAVIS_OS_NAME}" = "linux" ]; then ./travis.linux.script.sh; fi
- if [ "${TRAVIS_OS_NAME}" = "osx" ]; then ./travis.osx.script.sh; fi
git:
submodules: true
notifications:
irc:
channels: "irc.mozilla.org#servo"
on_success: change
on_failure: change
use_notice: true
email: false
branches:
only:
- master
## Instruction:
Enable email notifications from Travis CI
http://docs.travis-ci.com/user/notifications/
> By default, email notifications will be sent to the committer and the commit author, if they are members of the repository (that is, they have push or admin permissions for public repositories, or if they have pull, push or admin permissions for private repositories).
>
> And it will by default send emails when, on the given branch:
>
> * a build was just broken or still is broken
> * a previously broken build was just fixed
## Code After:
language: c
os:
- linux
- osx
install:
- if [ "${TRAVIS_OS_NAME}" = "linux" ]; then ./travis.linux.install.deps.sh; fi
- if [ "${TRAVIS_OS_NAME}" = "osx" ]; then ./travis.osx.install.deps.sh; fi
before_script:
- mkdir -p build
script:
- if [ "${TRAVIS_OS_NAME}" = "linux" ]; then ./travis.linux.script.sh; fi
- if [ "${TRAVIS_OS_NAME}" = "osx" ]; then ./travis.osx.script.sh; fi
git:
submodules: true
notifications:
irc:
channels: "irc.mozilla.org#servo"
on_success: change
on_failure: change
use_notice: true
branches:
only:
- master
| language: c
os:
- linux
- osx
install:
- if [ "${TRAVIS_OS_NAME}" = "linux" ]; then ./travis.linux.install.deps.sh; fi
- if [ "${TRAVIS_OS_NAME}" = "osx" ]; then ./travis.osx.install.deps.sh; fi
before_script:
- mkdir -p build
script:
- if [ "${TRAVIS_OS_NAME}" = "linux" ]; then ./travis.linux.script.sh; fi
- if [ "${TRAVIS_OS_NAME}" = "osx" ]; then ./travis.osx.script.sh; fi
git:
submodules: true
notifications:
irc:
channels: "irc.mozilla.org#servo"
on_success: change
on_failure: change
use_notice: true
- email: false
branches:
only:
- master | 1 | 0.032258 | 0 | 1 |
2838bda117bd51688a4f3224af6f7a23cb767c04 | lib/openers/sumatra-opener.js | lib/openers/sumatra-opener.js | /** @babel */
import fs from 'fs-plus'
import Opener from '../opener'
import { isPdfFile } from '../werkzeug'
export default class SumatraOpener extends Opener {
async open (filePath, texPath, lineNumber) {
const sumatraPath = `"${atom.config.get('latex.sumatraPath')}"`
const atomPath = `"${process.argv[0]}"`
const args = [
'-reuse-instance',
'-forward-search',
`"${texPath}"`,
lineNumber,
'-inverse-search',
`"\\"${atomPath}\\" \\"%f:%l\\""`,
`"${filePath}"`
]
const command = `${sumatraPath} ${args.join(' ')}`
console.log(command)
await latex.process.executeChildProcess(command)
}
canOpen (filePath) {
return process.platform === 'win32' && isPdfFile(filePath) &&
fs.existsSync(atom.config.get('latex.sumatraPath'))
}
hasSynctex () {
return true
}
}
| /** @babel */
import fs from 'fs-plus'
import Opener from '../opener'
import { isPdfFile } from '../werkzeug'
export default class SumatraOpener extends Opener {
async open (filePath, texPath, lineNumber) {
const sumatraPath = `"${atom.config.get('latex.sumatraPath')}"`
const atomPath = process.argv[0]
const args = [
'-reuse-instance',
'-forward-search',
`"${texPath}"`,
lineNumber,
'-inverse-search',
`"\\"${atomPath}\\" \\"%f:%l\\""`,
`"${filePath}"`
]
const command = `${sumatraPath} ${args.join(' ')}`
await latex.process.executeChildProcess(command)
}
canOpen (filePath) {
return process.platform === 'win32' && isPdfFile(filePath) &&
fs.existsSync(atom.config.get('latex.sumatraPath'))
}
hasSynctex () {
return true
}
}
| Remove extra quoting of Atom path | Remove extra quoting of Atom path
| JavaScript | mit | thomasjo/atom-latex,thomasjo/atom-latex,thomasjo/atom-latex | javascript | ## Code Before:
/** @babel */
import fs from 'fs-plus'
import Opener from '../opener'
import { isPdfFile } from '../werkzeug'
export default class SumatraOpener extends Opener {
async open (filePath, texPath, lineNumber) {
const sumatraPath = `"${atom.config.get('latex.sumatraPath')}"`
const atomPath = `"${process.argv[0]}"`
const args = [
'-reuse-instance',
'-forward-search',
`"${texPath}"`,
lineNumber,
'-inverse-search',
`"\\"${atomPath}\\" \\"%f:%l\\""`,
`"${filePath}"`
]
const command = `${sumatraPath} ${args.join(' ')}`
console.log(command)
await latex.process.executeChildProcess(command)
}
canOpen (filePath) {
return process.platform === 'win32' && isPdfFile(filePath) &&
fs.existsSync(atom.config.get('latex.sumatraPath'))
}
hasSynctex () {
return true
}
}
## Instruction:
Remove extra quoting of Atom path
## Code After:
/** @babel */
import fs from 'fs-plus'
import Opener from '../opener'
import { isPdfFile } from '../werkzeug'
export default class SumatraOpener extends Opener {
async open (filePath, texPath, lineNumber) {
const sumatraPath = `"${atom.config.get('latex.sumatraPath')}"`
const atomPath = process.argv[0]
const args = [
'-reuse-instance',
'-forward-search',
`"${texPath}"`,
lineNumber,
'-inverse-search',
`"\\"${atomPath}\\" \\"%f:%l\\""`,
`"${filePath}"`
]
const command = `${sumatraPath} ${args.join(' ')}`
await latex.process.executeChildProcess(command)
}
canOpen (filePath) {
return process.platform === 'win32' && isPdfFile(filePath) &&
fs.existsSync(atom.config.get('latex.sumatraPath'))
}
hasSynctex () {
return true
}
}
| /** @babel */
import fs from 'fs-plus'
import Opener from '../opener'
import { isPdfFile } from '../werkzeug'
export default class SumatraOpener extends Opener {
async open (filePath, texPath, lineNumber) {
const sumatraPath = `"${atom.config.get('latex.sumatraPath')}"`
- const atomPath = `"${process.argv[0]}"`
? ---- ---
+ const atomPath = process.argv[0]
const args = [
'-reuse-instance',
'-forward-search',
`"${texPath}"`,
lineNumber,
'-inverse-search',
`"\\"${atomPath}\\" \\"%f:%l\\""`,
`"${filePath}"`
]
const command = `${sumatraPath} ${args.join(' ')}`
-
- console.log(command)
await latex.process.executeChildProcess(command)
}
canOpen (filePath) {
return process.platform === 'win32' && isPdfFile(filePath) &&
fs.existsSync(atom.config.get('latex.sumatraPath'))
}
hasSynctex () {
return true
}
} | 4 | 0.111111 | 1 | 3 |
e97a1ed2015db2eb2d5fe6abe15af6d9020c16d9 | mbuild/tests/test_box.py | mbuild/tests/test_box.py | import pytest
import numpy as np
import mbuild as mb
from mbuild.tests.base_test import BaseTest
class TestBox(BaseTest):
def test_init_lengths(self):
box = mb.Box(lengths=np.ones(3))
assert np.array_equal(box.lengths, np.ones(3))
assert np.array_equal(box.mins, np.zeros(3))
assert np.array_equal(box.maxs, np.ones(3))
def test_init_bounds(self):
box = mb.Box(mins=np.zeros(3), maxs=np.ones(3))
assert np.array_equal(box.lengths, np.ones(3))
assert np.array_equal(box.mins, np.zeros(3))
assert np.array_equal(box.maxs, np.ones(3))
def test_scale(self):
box = mb.Box(lengths=np.ones(3))
scaling_factors = np.array([3, 4, 5])
box.scale(scaling_factors)
assert np.array_equal(box.lengths, scaling_factors)
assert np.array_equal(box.mins, (np.ones(3) / 2) - (scaling_factors / 2))
assert np.array_equal(box.maxs, (scaling_factors / 2) + (np.ones(3) / 2))
def test_center(self):
box = mb.Box(lengths=np.ones(3))
box.center()
assert np.array_equal(box.lengths, np.ones(3))
assert np.array_equal(box.mins, np.ones(3) * -0.5)
assert np.array_equal(box.maxs, np.ones(3) * 0.5)
| import pytest
import numpy as np
import mbuild as mb
from mbuild.tests.base_test import BaseTest
class TestBox(BaseTest):
def test_init_lengths(self):
box = mb.Box(lengths=np.ones(3))
assert np.array_equal(box.lengths, np.ones(3))
assert np.array_equal(box.mins, np.zeros(3))
assert np.array_equal(box.maxs, np.ones(3))
def test_init_bounds(self):
box = mb.Box(mins=np.zeros(3), maxs=np.ones(3))
assert np.array_equal(box.lengths, np.ones(3))
assert np.array_equal(box.mins, np.zeros(3))
assert np.array_equal(box.maxs, np.ones(3))
| Remove box tests for non-existant functionality | Remove box tests for non-existant functionality
| Python | mit | ctk3b/mbuild,iModels/mbuild,iModels/mbuild,ctk3b/mbuild,tcmoore3/mbuild,tcmoore3/mbuild,summeraz/mbuild,summeraz/mbuild | python | ## Code Before:
import pytest
import numpy as np
import mbuild as mb
from mbuild.tests.base_test import BaseTest
class TestBox(BaseTest):
def test_init_lengths(self):
box = mb.Box(lengths=np.ones(3))
assert np.array_equal(box.lengths, np.ones(3))
assert np.array_equal(box.mins, np.zeros(3))
assert np.array_equal(box.maxs, np.ones(3))
def test_init_bounds(self):
box = mb.Box(mins=np.zeros(3), maxs=np.ones(3))
assert np.array_equal(box.lengths, np.ones(3))
assert np.array_equal(box.mins, np.zeros(3))
assert np.array_equal(box.maxs, np.ones(3))
def test_scale(self):
box = mb.Box(lengths=np.ones(3))
scaling_factors = np.array([3, 4, 5])
box.scale(scaling_factors)
assert np.array_equal(box.lengths, scaling_factors)
assert np.array_equal(box.mins, (np.ones(3) / 2) - (scaling_factors / 2))
assert np.array_equal(box.maxs, (scaling_factors / 2) + (np.ones(3) / 2))
def test_center(self):
box = mb.Box(lengths=np.ones(3))
box.center()
assert np.array_equal(box.lengths, np.ones(3))
assert np.array_equal(box.mins, np.ones(3) * -0.5)
assert np.array_equal(box.maxs, np.ones(3) * 0.5)
## Instruction:
Remove box tests for non-existant functionality
## Code After:
import pytest
import numpy as np
import mbuild as mb
from mbuild.tests.base_test import BaseTest
class TestBox(BaseTest):
def test_init_lengths(self):
box = mb.Box(lengths=np.ones(3))
assert np.array_equal(box.lengths, np.ones(3))
assert np.array_equal(box.mins, np.zeros(3))
assert np.array_equal(box.maxs, np.ones(3))
def test_init_bounds(self):
box = mb.Box(mins=np.zeros(3), maxs=np.ones(3))
assert np.array_equal(box.lengths, np.ones(3))
assert np.array_equal(box.mins, np.zeros(3))
assert np.array_equal(box.maxs, np.ones(3))
| import pytest
import numpy as np
import mbuild as mb
from mbuild.tests.base_test import BaseTest
class TestBox(BaseTest):
def test_init_lengths(self):
box = mb.Box(lengths=np.ones(3))
assert np.array_equal(box.lengths, np.ones(3))
assert np.array_equal(box.mins, np.zeros(3))
assert np.array_equal(box.maxs, np.ones(3))
def test_init_bounds(self):
box = mb.Box(mins=np.zeros(3), maxs=np.ones(3))
assert np.array_equal(box.lengths, np.ones(3))
assert np.array_equal(box.mins, np.zeros(3))
assert np.array_equal(box.maxs, np.ones(3))
-
- def test_scale(self):
- box = mb.Box(lengths=np.ones(3))
- scaling_factors = np.array([3, 4, 5])
- box.scale(scaling_factors)
- assert np.array_equal(box.lengths, scaling_factors)
- assert np.array_equal(box.mins, (np.ones(3) / 2) - (scaling_factors / 2))
- assert np.array_equal(box.maxs, (scaling_factors / 2) + (np.ones(3) / 2))
-
- def test_center(self):
- box = mb.Box(lengths=np.ones(3))
- box.center()
- assert np.array_equal(box.lengths, np.ones(3))
- assert np.array_equal(box.mins, np.ones(3) * -0.5)
- assert np.array_equal(box.maxs, np.ones(3) * 0.5) | 15 | 0.441176 | 0 | 15 |
4e7400ae615e537113fba0da228b350cb8496f75 | server/stack.yaml | server/stack.yaml | resolver: lts-12.5
extra-deps:
- github: fimad/prometheus-haskell
commit: e44ab37a5d4e4a570e24cb1ca29d2753a27090c5
subdirs:
- prometheus-client
- prometheus-metrics-ghc
- wai-middleware-prometheus
| resolver: lts-12.5
extra-deps:
- prometheus-client-1.0.0
- prometheus-metrics-ghc-1.0.0
- wai-middleware-prometheus-1.0.0
| Use the newly released prometheus-client-1.0.0 library | Use the newly released prometheus-client-1.0.0 library
Instead of building from git. This still needs some fixes in our code
though, since the types in the library changed.
| YAML | bsd-3-clause | channable/icepeak,channable/icepeak,channable/icepeak | yaml | ## Code Before:
resolver: lts-12.5
extra-deps:
- github: fimad/prometheus-haskell
commit: e44ab37a5d4e4a570e24cb1ca29d2753a27090c5
subdirs:
- prometheus-client
- prometheus-metrics-ghc
- wai-middleware-prometheus
## Instruction:
Use the newly released prometheus-client-1.0.0 library
Instead of building from git. This still needs some fixes in our code
though, since the types in the library changed.
## Code After:
resolver: lts-12.5
extra-deps:
- prometheus-client-1.0.0
- prometheus-metrics-ghc-1.0.0
- wai-middleware-prometheus-1.0.0
| resolver: lts-12.5
extra-deps:
- - github: fimad/prometheus-haskell
- commit: e44ab37a5d4e4a570e24cb1ca29d2753a27090c5
- subdirs:
- - prometheus-client
? ---
+ - prometheus-client-1.0.0
? ++++++
- - prometheus-metrics-ghc
? ---
+ - prometheus-metrics-ghc-1.0.0
? ++++++
- - wai-middleware-prometheus
? ---
+ - wai-middleware-prometheus-1.0.0
? ++++++
| 9 | 1 | 3 | 6 |
7f43a220bd1205b65df98ab7591038e7b1c6d837 | view/index.html.twig | view/index.html.twig | {% extends "base.html.twig" %}
{% block content %}
<h1>Welcome to my recipe book</h1>
<p>Just have a look around</p>
{% include 'recipe/list.html.twig' %}
{% endblock %}
| {% extends "base.html.twig" %}
{% block content %}
<h1>Welcome to my recipe book</h1>
<p>Just have a look around</p>
<div>
<a class="btn btn-primary" href="edit-recipe.php">Add a recipe</a>
</div>
{% include 'recipe/list.html.twig' %}
{% endblock %}
| Add a recipe button in index | Add a recipe button in index
| Twig | mit | webmaster777/recipe-book,webmaster777/recipe-book,webmaster777/recipe-book | twig | ## Code Before:
{% extends "base.html.twig" %}
{% block content %}
<h1>Welcome to my recipe book</h1>
<p>Just have a look around</p>
{% include 'recipe/list.html.twig' %}
{% endblock %}
## Instruction:
Add a recipe button in index
## Code After:
{% extends "base.html.twig" %}
{% block content %}
<h1>Welcome to my recipe book</h1>
<p>Just have a look around</p>
<div>
<a class="btn btn-primary" href="edit-recipe.php">Add a recipe</a>
</div>
{% include 'recipe/list.html.twig' %}
{% endblock %}
| {% extends "base.html.twig" %}
{% block content %}
<h1>Welcome to my recipe book</h1>
<p>Just have a look around</p>
+ <div>
+ <a class="btn btn-primary" href="edit-recipe.php">Add a recipe</a>
+ </div>
+
{% include 'recipe/list.html.twig' %}
{% endblock %} | 4 | 0.444444 | 4 | 0 |
156d84ce4ba53ac46280566681019327ed133de3 | templates/package.json | templates/package.json | {
"name": "<%= appNameSlug %>",
"version": "0.0.0",
"dependencies": {},
"devDependencies": {
"gulp": "^3.6.2",
"connect": "^2.14.5",<% if (includeSass) { %>
"gulp-sass": "^0.7.1",
"gulp-cssbeautify": "^0.1.3",<% } %>
"gulp-livereload": "^1.3.1",
"opn": "^0.1.1",
"connect-livereload": "^0.4.0",
"gulp-jshint": "^1.5.5",
"gulp-useref": "^0.4.2",
"gulp-minify-css": "^0.3.1",
"gulp-uglify": "^0.2.1",
"gulp-if": "^1.2.1",
"rimraf": "^2.2.6",
"wiredep": "^1.4.4"
},
"engines": {
"node": ">=0.10.0"
}
}
| {
"name": "<%= appNameSlug %>",
"version": "0.0.0",
"dependencies": {},
"devDependencies": {
"gulp": "^3.6.2",
"connect": "^2.14.5",<% if (includeSass) { %>
"gulp-sass": "^0.7.1",
"gulp-cssbeautify": "^0.1.3",<% } %>
"gulp-livereload": "^1.3.1",
"opn": "^0.1.1",
"connect-livereload": "^0.4.0",
"gulp-if": "^1.2.1",
"gulp-jshint": "^1.6.3",
"gulp-minify-css": "^0.3.6",
"gulp-uglify": "^0.3.1",
"gulp-useref": "^0.5.0",
"rimraf": "^2.2.6",
"wiredep": "^1.4.4"
},
"engines": {
"node": ">=0.10.0"
}
}
| Update devDependencies for default webapp. | Update devDependencies for default webapp.
| JSON | mit | jonkemp/slush-webapp | json | ## Code Before:
{
"name": "<%= appNameSlug %>",
"version": "0.0.0",
"dependencies": {},
"devDependencies": {
"gulp": "^3.6.2",
"connect": "^2.14.5",<% if (includeSass) { %>
"gulp-sass": "^0.7.1",
"gulp-cssbeautify": "^0.1.3",<% } %>
"gulp-livereload": "^1.3.1",
"opn": "^0.1.1",
"connect-livereload": "^0.4.0",
"gulp-jshint": "^1.5.5",
"gulp-useref": "^0.4.2",
"gulp-minify-css": "^0.3.1",
"gulp-uglify": "^0.2.1",
"gulp-if": "^1.2.1",
"rimraf": "^2.2.6",
"wiredep": "^1.4.4"
},
"engines": {
"node": ">=0.10.0"
}
}
## Instruction:
Update devDependencies for default webapp.
## Code After:
{
"name": "<%= appNameSlug %>",
"version": "0.0.0",
"dependencies": {},
"devDependencies": {
"gulp": "^3.6.2",
"connect": "^2.14.5",<% if (includeSass) { %>
"gulp-sass": "^0.7.1",
"gulp-cssbeautify": "^0.1.3",<% } %>
"gulp-livereload": "^1.3.1",
"opn": "^0.1.1",
"connect-livereload": "^0.4.0",
"gulp-if": "^1.2.1",
"gulp-jshint": "^1.6.3",
"gulp-minify-css": "^0.3.6",
"gulp-uglify": "^0.3.1",
"gulp-useref": "^0.5.0",
"rimraf": "^2.2.6",
"wiredep": "^1.4.4"
},
"engines": {
"node": ">=0.10.0"
}
}
| {
"name": "<%= appNameSlug %>",
"version": "0.0.0",
"dependencies": {},
"devDependencies": {
"gulp": "^3.6.2",
"connect": "^2.14.5",<% if (includeSass) { %>
"gulp-sass": "^0.7.1",
"gulp-cssbeautify": "^0.1.3",<% } %>
"gulp-livereload": "^1.3.1",
"opn": "^0.1.1",
"connect-livereload": "^0.4.0",
- "gulp-jshint": "^1.5.5",
- "gulp-useref": "^0.4.2",
- "gulp-minify-css": "^0.3.1",
- "gulp-uglify": "^0.2.1",
"gulp-if": "^1.2.1",
+ "gulp-jshint": "^1.6.3",
+ "gulp-minify-css": "^0.3.6",
+ "gulp-uglify": "^0.3.1",
+ "gulp-useref": "^0.5.0",
"rimraf": "^2.2.6",
"wiredep": "^1.4.4"
},
"engines": {
"node": ">=0.10.0"
}
} | 8 | 0.333333 | 4 | 4 |
4110b1c3c8eccef14370f48ad5696d04b829ae3d | pillar/elk/prod.sls | pillar/elk/prod.sls | elk:
lumberjack-host-name: dfb10 # used for cert generation
lumberjack-host-ip: 10.172.2.105 # used for cert generation
lumberjack-host: dfb10 # used for actual connect to server
lumberjack-port: 5000
configserver-host: dfb10
configserver-port: 9999
| elk:
lumberjack-host-name: dfb # used for cert generation
lumberjack-host-ip: 171.23.3.39 # used for cert generation
lumberjack-host: dfb # used for actual connect to server
lumberjack-port: 5000
configserver-host: dfb
configserver-port: 9999
| Change dfb10 -> dfb due to network changes. | Change dfb10 -> dfb due to network changes.
| SaltStack | mit | digibib/ls.ext,digibib/ls.ext,digibib/ls.ext,digibib/ls.ext | saltstack | ## Code Before:
elk:
lumberjack-host-name: dfb10 # used for cert generation
lumberjack-host-ip: 10.172.2.105 # used for cert generation
lumberjack-host: dfb10 # used for actual connect to server
lumberjack-port: 5000
configserver-host: dfb10
configserver-port: 9999
## Instruction:
Change dfb10 -> dfb due to network changes.
## Code After:
elk:
lumberjack-host-name: dfb # used for cert generation
lumberjack-host-ip: 171.23.3.39 # used for cert generation
lumberjack-host: dfb # used for actual connect to server
lumberjack-port: 5000
configserver-host: dfb
configserver-port: 9999
| elk:
- lumberjack-host-name: dfb10 # used for cert generation
? --
+ lumberjack-host-name: dfb # used for cert generation
- lumberjack-host-ip: 10.172.2.105 # used for cert generation
? ^^ -- ^^^
+ lumberjack-host-ip: 171.23.3.39 # used for cert generation
? ^ + ^^^^
- lumberjack-host: dfb10 # used for actual connect to server
? --
+ lumberjack-host: dfb # used for actual connect to server
lumberjack-port: 5000
- configserver-host: dfb10
? --
+ configserver-host: dfb
configserver-port: 9999 | 8 | 1.142857 | 4 | 4 |
e2294fe4e3963940d6680d44912211298280f95f | package.json | package.json | {
"name": "hammertime",
"version": "2.0.0",
"description": "U Can't Touch This",
"main": "handler.js",
"scripts": {
"deploy": "node_modules/serverless/bin/serverless deploy --verbose",
"destroy": "node_modules/serverless/bin/serverless remove --verbose",
"lint": "eslint *.js",
"test": "npm run lint && npm run test.unit",
"test.unit": "istanbul cover node_modules/mocha/bin/_mocha -- test/**/*.test.js && istanbul check"
},
"repository": {
"type": "git",
"url": "git@github.com:nib-health-funds/hammertime.git"
},
"author": "nib",
"license": "MIT",
"dependencies": {
"aws-sdk": "^2.7.19",
"istanbul": "^0.4.5",
"promise-retry": "^1.1.1",
"serverless": "^1.4.0"
},
"devDependencies": {
"aws-sdk-mock": "^1.6.1",
"eslint": "^3.15.0",
"eslint-config-airbnb": "^14.1.0",
"eslint-plugin-import": "^2.2.0",
"eslint-plugin-jsx-a11y": "^4.0.0",
"eslint-plugin-mocha": "^4.9.0",
"eslint-plugin-react": "^6.9.0",
"mocha": "^3.2.0"
}
}
| {
"name": "hammertime",
"version": "2.0.0",
"description": "U Can't Touch This",
"main": "handler.js",
"scripts": {
"deploy": "node_modules/serverless/bin/serverless deploy --verbose",
"destroy": "node_modules/serverless/bin/serverless remove --verbose",
"lint": "eslint *.js",
"test": "npm run lint && npm run test.unit",
"test.unit": "istanbul cover node_modules/mocha/bin/_mocha -- test/**/*.test.js && istanbul check"
},
"repository": {
"type": "git",
"url": "git@github.com:nib-health-funds/hammertime.git"
},
"author": "nib",
"license": "MIT",
"dependencies": {
"aws-sdk": "^2.7.19",
"promise-retry": "^1.1.1",
"serverless": "^1.4.0"
},
"devDependencies": {
"istanbul": "^0.4.5",
"aws-sdk-mock": "^1.6.1",
"eslint": "^3.15.0",
"eslint-config-airbnb": "^14.1.0",
"eslint-plugin-import": "^2.2.0",
"eslint-plugin-jsx-a11y": "^4.0.0",
"eslint-plugin-mocha": "^4.9.0",
"eslint-plugin-react": "^6.9.0",
"mocha": "^3.2.0"
}
}
| Move istanbul to dev deps | Move istanbul to dev deps
| JSON | mit | nib-health-funds/hammertime | json | ## Code Before:
{
"name": "hammertime",
"version": "2.0.0",
"description": "U Can't Touch This",
"main": "handler.js",
"scripts": {
"deploy": "node_modules/serverless/bin/serverless deploy --verbose",
"destroy": "node_modules/serverless/bin/serverless remove --verbose",
"lint": "eslint *.js",
"test": "npm run lint && npm run test.unit",
"test.unit": "istanbul cover node_modules/mocha/bin/_mocha -- test/**/*.test.js && istanbul check"
},
"repository": {
"type": "git",
"url": "git@github.com:nib-health-funds/hammertime.git"
},
"author": "nib",
"license": "MIT",
"dependencies": {
"aws-sdk": "^2.7.19",
"istanbul": "^0.4.5",
"promise-retry": "^1.1.1",
"serverless": "^1.4.0"
},
"devDependencies": {
"aws-sdk-mock": "^1.6.1",
"eslint": "^3.15.0",
"eslint-config-airbnb": "^14.1.0",
"eslint-plugin-import": "^2.2.0",
"eslint-plugin-jsx-a11y": "^4.0.0",
"eslint-plugin-mocha": "^4.9.0",
"eslint-plugin-react": "^6.9.0",
"mocha": "^3.2.0"
}
}
## Instruction:
Move istanbul to dev deps
## Code After:
{
"name": "hammertime",
"version": "2.0.0",
"description": "U Can't Touch This",
"main": "handler.js",
"scripts": {
"deploy": "node_modules/serverless/bin/serverless deploy --verbose",
"destroy": "node_modules/serverless/bin/serverless remove --verbose",
"lint": "eslint *.js",
"test": "npm run lint && npm run test.unit",
"test.unit": "istanbul cover node_modules/mocha/bin/_mocha -- test/**/*.test.js && istanbul check"
},
"repository": {
"type": "git",
"url": "git@github.com:nib-health-funds/hammertime.git"
},
"author": "nib",
"license": "MIT",
"dependencies": {
"aws-sdk": "^2.7.19",
"promise-retry": "^1.1.1",
"serverless": "^1.4.0"
},
"devDependencies": {
"istanbul": "^0.4.5",
"aws-sdk-mock": "^1.6.1",
"eslint": "^3.15.0",
"eslint-config-airbnb": "^14.1.0",
"eslint-plugin-import": "^2.2.0",
"eslint-plugin-jsx-a11y": "^4.0.0",
"eslint-plugin-mocha": "^4.9.0",
"eslint-plugin-react": "^6.9.0",
"mocha": "^3.2.0"
}
}
| {
"name": "hammertime",
"version": "2.0.0",
"description": "U Can't Touch This",
"main": "handler.js",
"scripts": {
"deploy": "node_modules/serverless/bin/serverless deploy --verbose",
"destroy": "node_modules/serverless/bin/serverless remove --verbose",
"lint": "eslint *.js",
"test": "npm run lint && npm run test.unit",
"test.unit": "istanbul cover node_modules/mocha/bin/_mocha -- test/**/*.test.js && istanbul check"
},
"repository": {
"type": "git",
"url": "git@github.com:nib-health-funds/hammertime.git"
},
"author": "nib",
"license": "MIT",
"dependencies": {
"aws-sdk": "^2.7.19",
- "istanbul": "^0.4.5",
"promise-retry": "^1.1.1",
"serverless": "^1.4.0"
},
"devDependencies": {
+ "istanbul": "^0.4.5",
"aws-sdk-mock": "^1.6.1",
"eslint": "^3.15.0",
"eslint-config-airbnb": "^14.1.0",
"eslint-plugin-import": "^2.2.0",
"eslint-plugin-jsx-a11y": "^4.0.0",
"eslint-plugin-mocha": "^4.9.0",
"eslint-plugin-react": "^6.9.0",
"mocha": "^3.2.0"
}
} | 2 | 0.057143 | 1 | 1 |
b985c62bee0275cab92e6db044ca390a65eaa774 | app/views/batch_activities/_form.html.haml | app/views/batch_activities/_form.html.haml | = semantic_form_for :batch_activities, :url => batch_activities_path do |f|
= f.inputs do
= f.input :date, :label => t('attributes.date'), :input_html => {:value => @date}
%table{:style => 'width: 100%'}
%thead
%th= t_attr :person, Activity
%th= t_attr :project, Activity
%th{:style => 'width: 5em'}= t_attr :duration, Activity
%th= t_attr :comment, Activity
%tbody
- for activity in @activities
%tr
= fields_for 'activities[]', activity do |a|
%td
= link_to activity.person, activity.person
= a.hidden_field :person_id
%td= a.select :project_id, Project.all.map{|project| [project.to_s, project.id]}, {:prompt => t_select_prompt(Project)}, :style => 'width: 98%'
%td= a.text_field :hours, :style => 'width: 98%'
%td= a.text_field :comment, :style => 'width: 98%'
= f.buttons do
= f.commit_button
| = semantic_form_for :batch_activities, :url => batch_activities_path do |f|
= f.inputs do
= f.input :date, :as => :date_field, :label => t('attributes.date'), :input_html => {:value => @date}
%table{:style => 'width: 100%'}
%thead
%th= t_attr :person, Activity
%th= t_attr :project, Activity
%th{:style => 'width: 5em'}= t_attr :duration, Activity
%th= t_attr :comment, Activity
%tbody
- for activity in @activities
%tr
= fields_for 'activities[]', activity do |a|
%td
= link_to activity.person, activity.person
= a.hidden_field :person_id
%td= a.select :project_id, Project.all.map{|project| [project.to_s, project.id]}, {:prompt => t_select_prompt(Project)}, :style => 'width: 98%'
%td= a.text_field :hours, :style => 'width: 98%'
%td= a.text_field :comment, :style => 'width: 98%'
= f.buttons do
= f.commit_button
| Mark date field in batch activities form for date picker etc. | Mark date field in batch activities form for date picker etc.
| Haml | mit | raskhadafi/bookyt_projects,huerlisi/bookyt_projects,huerlisi/bookyt_projects,raskhadafi/bookyt_projects,huerlisi/bookyt_projects | haml | ## Code Before:
= semantic_form_for :batch_activities, :url => batch_activities_path do |f|
= f.inputs do
= f.input :date, :label => t('attributes.date'), :input_html => {:value => @date}
%table{:style => 'width: 100%'}
%thead
%th= t_attr :person, Activity
%th= t_attr :project, Activity
%th{:style => 'width: 5em'}= t_attr :duration, Activity
%th= t_attr :comment, Activity
%tbody
- for activity in @activities
%tr
= fields_for 'activities[]', activity do |a|
%td
= link_to activity.person, activity.person
= a.hidden_field :person_id
%td= a.select :project_id, Project.all.map{|project| [project.to_s, project.id]}, {:prompt => t_select_prompt(Project)}, :style => 'width: 98%'
%td= a.text_field :hours, :style => 'width: 98%'
%td= a.text_field :comment, :style => 'width: 98%'
= f.buttons do
= f.commit_button
## Instruction:
Mark date field in batch activities form for date picker etc.
## Code After:
= semantic_form_for :batch_activities, :url => batch_activities_path do |f|
= f.inputs do
= f.input :date, :as => :date_field, :label => t('attributes.date'), :input_html => {:value => @date}
%table{:style => 'width: 100%'}
%thead
%th= t_attr :person, Activity
%th= t_attr :project, Activity
%th{:style => 'width: 5em'}= t_attr :duration, Activity
%th= t_attr :comment, Activity
%tbody
- for activity in @activities
%tr
= fields_for 'activities[]', activity do |a|
%td
= link_to activity.person, activity.person
= a.hidden_field :person_id
%td= a.select :project_id, Project.all.map{|project| [project.to_s, project.id]}, {:prompt => t_select_prompt(Project)}, :style => 'width: 98%'
%td= a.text_field :hours, :style => 'width: 98%'
%td= a.text_field :comment, :style => 'width: 98%'
= f.buttons do
= f.commit_button
| = semantic_form_for :batch_activities, :url => batch_activities_path do |f|
= f.inputs do
- = f.input :date, :label => t('attributes.date'), :input_html => {:value => @date}
+ = f.input :date, :as => :date_field, :label => t('attributes.date'), :input_html => {:value => @date}
? ++++++++++++++++++++
%table{:style => 'width: 100%'}
%thead
%th= t_attr :person, Activity
%th= t_attr :project, Activity
%th{:style => 'width: 5em'}= t_attr :duration, Activity
%th= t_attr :comment, Activity
%tbody
- for activity in @activities
%tr
= fields_for 'activities[]', activity do |a|
%td
= link_to activity.person, activity.person
= a.hidden_field :person_id
%td= a.select :project_id, Project.all.map{|project| [project.to_s, project.id]}, {:prompt => t_select_prompt(Project)}, :style => 'width: 98%'
%td= a.text_field :hours, :style => 'width: 98%'
%td= a.text_field :comment, :style => 'width: 98%'
= f.buttons do
= f.commit_button | 2 | 0.083333 | 1 | 1 |
eceb97fd98c93f332c5e182b870177d8549878c5 | .travis.yml | .travis.yml | sudo: false
language: ruby
cache: bundler
branches:
only:
- master
matrix:
allow_failures:
- rvm: 2.4.1
rvm:
- 2.2.7
- 2.3.4
- 2.4.1
| sudo: false
language: ruby
cache: bundler
branches:
only:
- master
rvm:
- 2.3.5
- 2.4.2
| Update Travis config for ruby versions | Update Travis config for ruby versions
Signed-off-by: Tim Smith <764ef62106582a09ed09dfa0b6bff7c05fd7d1e4@chef.io>
| YAML | mit | acrmp/foodcritic-site,acrmp/foodcritic-site,Foodcritic/foodcritic-site,Foodcritic/foodcritic-site | yaml | ## Code Before:
sudo: false
language: ruby
cache: bundler
branches:
only:
- master
matrix:
allow_failures:
- rvm: 2.4.1
rvm:
- 2.2.7
- 2.3.4
- 2.4.1
## Instruction:
Update Travis config for ruby versions
Signed-off-by: Tim Smith <764ef62106582a09ed09dfa0b6bff7c05fd7d1e4@chef.io>
## Code After:
sudo: false
language: ruby
cache: bundler
branches:
only:
- master
rvm:
- 2.3.5
- 2.4.2
| sudo: false
language: ruby
cache: bundler
branches:
only:
- master
- matrix:
- allow_failures:
- - rvm: 2.4.1
rvm:
- - 2.2.7
- - 2.3.4
? ^
+ - 2.3.5
? ^
- - 2.4.1
? ^
+ - 2.4.2
? ^
| 8 | 0.615385 | 2 | 6 |
17ab977cf202a4942cea7513643dce5c69694314 | app/scripts/views/board_view.coffee | app/scripts/views/board_view.coffee | class window.BoardView extends Backbone.View
el: $('.board__tiles')
initialize: ->
@listenTo @collection, 'reset', @render
@collection.view = @
@collection.shuffle()
render: =>
els = []
@$el.html('')
@collection.each (tile) ->
view = new TileView( model: tile )
els.push view.render().el
@$el.append els
return @
| class window.BoardView extends Backbone.View
el: $('.board__tiles')
initialize: ->
@listenTo @collection, 'reset', @render
@collection.shuffle()
@collection.view = @
render: =>
$.fx.off = yes
els = []
@$el.html('')
@collection.each (tile) ->
view = new TileView( model: tile )
els.push view.render().el
@$el.append els
_.delay ->
$.fx.off = no
, 100
@
| Disable animations on new game | Disable animations on new game
| CoffeeScript | mit | okonet/fifteens-trix,okonet/fifteens-trix | coffeescript | ## Code Before:
class window.BoardView extends Backbone.View
el: $('.board__tiles')
initialize: ->
@listenTo @collection, 'reset', @render
@collection.view = @
@collection.shuffle()
render: =>
els = []
@$el.html('')
@collection.each (tile) ->
view = new TileView( model: tile )
els.push view.render().el
@$el.append els
return @
## Instruction:
Disable animations on new game
## Code After:
class window.BoardView extends Backbone.View
el: $('.board__tiles')
initialize: ->
@listenTo @collection, 'reset', @render
@collection.shuffle()
@collection.view = @
render: =>
$.fx.off = yes
els = []
@$el.html('')
@collection.each (tile) ->
view = new TileView( model: tile )
els.push view.render().el
@$el.append els
_.delay ->
$.fx.off = no
, 100
@
| class window.BoardView extends Backbone.View
el: $('.board__tiles')
initialize: ->
@listenTo @collection, 'reset', @render
+ @collection.shuffle()
@collection.view = @
- @collection.shuffle()
render: =>
+ $.fx.off = yes
els = []
@$el.html('')
@collection.each (tile) ->
view = new TileView( model: tile )
els.push view.render().el
@$el.append els
- return @
+ _.delay ->
+ $.fx.off = no
+ , 100
+ @ | 8 | 0.470588 | 6 | 2 |
62b88141b8a5642fabc36e951ea0a15f41b95f2c | .travis.yml | .travis.yml | language: php
sudo: false
cache:
directories:
- $HOME/.composer/cache
matrix:
include:
- php: 5.4
- php: 5.5
- php: 5.6
- php: 5.6
env: COMPOSER_FLAGS='--prefer-lowest --prefer-stable'
- php: nightly
allow_failures:
- php: nightly
fast_finish: true
install: composer update $COMPOSER_FLAGS -n
script: vendor/bin/phpunit
| language: php
sudo: false
cache:
directories:
- $HOME/.composer/cache
matrix:
include:
- php: 5.4
- php: 5.5
- php: 5.6
- php: 5.6
env: COMPOSER_FLAGS='--prefer-lowest --prefer-stable'
- php: 7
- php: hhvm
install: composer update $COMPOSER_FLAGS -n
script: vendor/bin/phpunit
| Add support of PHP7 and HHVM | Add support of PHP7 and HHVM
| YAML | mit | FriendsOfSymfony/FOSMessage | yaml | ## Code Before:
language: php
sudo: false
cache:
directories:
- $HOME/.composer/cache
matrix:
include:
- php: 5.4
- php: 5.5
- php: 5.6
- php: 5.6
env: COMPOSER_FLAGS='--prefer-lowest --prefer-stable'
- php: nightly
allow_failures:
- php: nightly
fast_finish: true
install: composer update $COMPOSER_FLAGS -n
script: vendor/bin/phpunit
## Instruction:
Add support of PHP7 and HHVM
## Code After:
language: php
sudo: false
cache:
directories:
- $HOME/.composer/cache
matrix:
include:
- php: 5.4
- php: 5.5
- php: 5.6
- php: 5.6
env: COMPOSER_FLAGS='--prefer-lowest --prefer-stable'
- php: 7
- php: hhvm
install: composer update $COMPOSER_FLAGS -n
script: vendor/bin/phpunit
| language: php
sudo: false
cache:
directories:
- $HOME/.composer/cache
matrix:
include:
- php: 5.4
- php: 5.5
- php: 5.6
- php: 5.6
env: COMPOSER_FLAGS='--prefer-lowest --prefer-stable'
+ - php: 7
+ - php: hhvm
- - php: nightly
- allow_failures:
- - php: nightly
- fast_finish: true
install: composer update $COMPOSER_FLAGS -n
script: vendor/bin/phpunit | 6 | 0.26087 | 2 | 4 |
bf5224bd99f0e31b9e3e313ce2590cd0fbefb44a | examples/blaze/src/main/scala/com/example/http4s/blaze/BlazeHttp2Example.scala | examples/blaze/src/main/scala/com/example/http4s/blaze/BlazeHttp2Example.scala | package com.example.http4s
package blaze
import cats.effect._
import cats.implicits._
object BlazeHttp2Example extends IOApp {
override def run(args: List[String]): IO[ExitCode] =
BlazeSslExampleApp
.builder[IO]
.enableHttp2(true)
.serve
.compile
.drain
.as(ExitCode.Success)
}
| package com.example.http4s
package blaze
import cats.effect._
import cats.implicits._
object BlazeHttp2Example extends IOApp {
override def run(args: List[String]): IO[ExitCode] =
BlazeSslExampleApp
.builder[IO]
.enableHttp2(true)
.serve
.compile
.lastOrError
}
| Update Example to not ignore App ExitCode | Update Example to not ignore App ExitCode | Scala | apache-2.0 | ChristopherDavenport/http4s,http4s/http4s,ChristopherDavenport/http4s,rossabaker/http4s,ChristopherDavenport/http4s,rossabaker/http4s | scala | ## Code Before:
package com.example.http4s
package blaze
import cats.effect._
import cats.implicits._
object BlazeHttp2Example extends IOApp {
override def run(args: List[String]): IO[ExitCode] =
BlazeSslExampleApp
.builder[IO]
.enableHttp2(true)
.serve
.compile
.drain
.as(ExitCode.Success)
}
## Instruction:
Update Example to not ignore App ExitCode
## Code After:
package com.example.http4s
package blaze
import cats.effect._
import cats.implicits._
object BlazeHttp2Example extends IOApp {
override def run(args: List[String]): IO[ExitCode] =
BlazeSslExampleApp
.builder[IO]
.enableHttp2(true)
.serve
.compile
.lastOrError
}
| package com.example.http4s
package blaze
import cats.effect._
import cats.implicits._
object BlazeHttp2Example extends IOApp {
override def run(args: List[String]): IO[ExitCode] =
BlazeSslExampleApp
.builder[IO]
.enableHttp2(true)
.serve
.compile
+ .lastOrError
- .drain
- .as(ExitCode.Success)
} | 3 | 0.1875 | 1 | 2 |
ac3cc391f02a43e5bee389978f925149d3f1a379 | README.md | README.md |
This project documents the JSON format for the export tool of the Sunriise project.
|
This project documents the JSON format for the export tool of the Sunriise project.
For more information see [wiki](https://github.com/hleofxquotes/mnyjson/wiki)
| Add pointer to wiki page. | Add pointer to wiki page.
| Markdown | apache-2.0 | hleofxquotes/mnyjson,hleofxquotes/mnyjson,hleofxquotes/mnyjson | markdown | ## Code Before:
This project documents the JSON format for the export tool of the Sunriise project.
## Instruction:
Add pointer to wiki page.
## Code After:
This project documents the JSON format for the export tool of the Sunriise project.
For more information see [wiki](https://github.com/hleofxquotes/mnyjson/wiki)
|
This project documents the JSON format for the export tool of the Sunriise project.
+
+ For more information see [wiki](https://github.com/hleofxquotes/mnyjson/wiki) | 2 | 1 | 2 | 0 |
28b88955a277d16e8d49d3d6d65429454c7bf373 | blog-design-cheatsheet.markdown | blog-design-cheatsheet.markdown |
Here are the design tags for blog post frontmatter for common technologies and frameworks that we write about. Copy existing ones whenever possible instead of requesting new ones from design. Feel free to add to this list.
If adding new logos, upload them to the `logos` directory on the CDN. This will ensure that they are easy to find.
### Angular
```
bg_color: "#012C6C"
image: https://cdn.auth0.com/blog/logos/angular.png
```
### Firebase
```
bg_color: "#4236C9"
image: https://cdn.auth0.com/blog/logos/firebase.png
```
### JavaScript
```
bg_color: "#222228"
image: https://cdn.auth0.com/blog/logos/js.png
```
### Node.js
```
bg_color: "#333333"
image: https://cdn.auth0.com/blog/logos/node.png
```
### React
```
bg_color: "#1A1A1A"
image: https://cdn.auth0.com/blog/logos/react.png
```
### Spring Boot
```
bg_color: "#222228"
image: https://cdn.auth0.com/blog/logos/spring-boot.png
``` |
Here are the design tags for blog post frontmatter for common technologies and frameworks that we write about. Copy existing ones whenever possible instead of requesting new ones from design. Feel free to add to this list.
If adding new logos, upload them to the `logos` directory on the CDN. This will ensure that they are easy to find.
### Angular
```
bg_color: "#012C6C"
image: https://cdn.auth0.com/blog/logos/angular.png
```
### Firebase
```
bg_color: "#4236C9"
image: https://cdn.auth0.com/blog/logos/firebase.png
```
### JavaScript
```
bg_color: "#222228"
image: https://cdn.auth0.com/blog/logos/js.png
```
### Node.js
```
bg_color: "#333333"
image: https://cdn.auth0.com/blog/logos/node.png
```
### React
```
bg_color: "#1A1A1A"
image: https://cdn.auth0.com/blog/logos/react.png
```
### Spring Boot
```
bg_color: "#222228"
image: https://cdn.auth0.com/blog/logos/spring-boot.png
```
### PHP
```
bg_color: "#312A4C"
image: https://cdn.auth0.com/blog/logos/php.png
```
| Add PHP image and bg color | Add PHP image and bg color | Markdown | mit | kmaida/blog,mike-casas/blog,auth0/blog,auth0/blog,auth0/blog,mike-casas/blog,auth0/blog,mike-casas/blog,kmaida/blog,kmaida/blog,kmaida/blog | markdown | ## Code Before:
Here are the design tags for blog post frontmatter for common technologies and frameworks that we write about. Copy existing ones whenever possible instead of requesting new ones from design. Feel free to add to this list.
If adding new logos, upload them to the `logos` directory on the CDN. This will ensure that they are easy to find.
### Angular
```
bg_color: "#012C6C"
image: https://cdn.auth0.com/blog/logos/angular.png
```
### Firebase
```
bg_color: "#4236C9"
image: https://cdn.auth0.com/blog/logos/firebase.png
```
### JavaScript
```
bg_color: "#222228"
image: https://cdn.auth0.com/blog/logos/js.png
```
### Node.js
```
bg_color: "#333333"
image: https://cdn.auth0.com/blog/logos/node.png
```
### React
```
bg_color: "#1A1A1A"
image: https://cdn.auth0.com/blog/logos/react.png
```
### Spring Boot
```
bg_color: "#222228"
image: https://cdn.auth0.com/blog/logos/spring-boot.png
```
## Instruction:
Add PHP image and bg color
## Code After:
Here are the design tags for blog post frontmatter for common technologies and frameworks that we write about. Copy existing ones whenever possible instead of requesting new ones from design. Feel free to add to this list.
If adding new logos, upload them to the `logos` directory on the CDN. This will ensure that they are easy to find.
### Angular
```
bg_color: "#012C6C"
image: https://cdn.auth0.com/blog/logos/angular.png
```
### Firebase
```
bg_color: "#4236C9"
image: https://cdn.auth0.com/blog/logos/firebase.png
```
### JavaScript
```
bg_color: "#222228"
image: https://cdn.auth0.com/blog/logos/js.png
```
### Node.js
```
bg_color: "#333333"
image: https://cdn.auth0.com/blog/logos/node.png
```
### React
```
bg_color: "#1A1A1A"
image: https://cdn.auth0.com/blog/logos/react.png
```
### Spring Boot
```
bg_color: "#222228"
image: https://cdn.auth0.com/blog/logos/spring-boot.png
```
### PHP
```
bg_color: "#312A4C"
image: https://cdn.auth0.com/blog/logos/php.png
```
|
Here are the design tags for blog post frontmatter for common technologies and frameworks that we write about. Copy existing ones whenever possible instead of requesting new ones from design. Feel free to add to this list.
If adding new logos, upload them to the `logos` directory on the CDN. This will ensure that they are easy to find.
### Angular
```
bg_color: "#012C6C"
image: https://cdn.auth0.com/blog/logos/angular.png
```
### Firebase
```
bg_color: "#4236C9"
image: https://cdn.auth0.com/blog/logos/firebase.png
```
### JavaScript
```
bg_color: "#222228"
image: https://cdn.auth0.com/blog/logos/js.png
```
### Node.js
```
bg_color: "#333333"
image: https://cdn.auth0.com/blog/logos/node.png
```
### React
```
bg_color: "#1A1A1A"
image: https://cdn.auth0.com/blog/logos/react.png
```
### Spring Boot
```
bg_color: "#222228"
image: https://cdn.auth0.com/blog/logos/spring-boot.png
```
+
+ ### PHP
+
+ ```
+ bg_color: "#312A4C"
+ image: https://cdn.auth0.com/blog/logos/php.png
+ ``` | 7 | 0.152174 | 7 | 0 |
0f791286f1cdd1ef4a6cfcf119c4495087f2d1bb | lib/pug/worker/daemon.rb | lib/pug/worker/daemon.rb | module Pug
module Worker
class Daemon
attr_reader :application
def initialize(application)
@application = application
end
def run
start
sleep
end
private
def start
daemonize if configuration.daemonize?
trap_signals
create_pid_file
application.start
end
def stop
delete_pid_file
application.stop
end
def daemonize
Process.daemon true, true
end
def trap_signals
trap_int_signal
trap_term_signal
end
def trap_int_signal
Signal.trap 'INT' do
Thread.new { stop }.join
exit! 1
end
end
def trap_term_signal
Signal.trap 'TERM' do
Thread.new { stop }.join
exit! 0
end
end
def create_pid_file
pid_file.create pid
end
def delete_pid_file
pid_file.delete
end
def pid_file
@pid_file ||= PidFile.new configuration.pid_path
end
def pid
Process.pid
end
def configuration
Pug::Worker.configuration
end
end
end
end
| module Pug
module Worker
class Daemon
attr_reader :application
def initialize(application)
@application = application
end
def run
start
sleep
end
private
def start
daemonize if daemonize?
trap_signals
create_pid_file if pid_path
application.start
end
def stop
delete_pid_file if pid_path
application.stop
end
def daemonize
Process.daemon true, true
end
def trap_signals
trap_int_signal
trap_term_signal
end
def trap_int_signal
Signal.trap 'INT' do
Thread.new { stop }.join
exit! 1
end
end
def trap_term_signal
Signal.trap 'TERM' do
Thread.new { stop }.join
exit! 0
end
end
def create_pid_file
pid_file.create pid
end
def delete_pid_file
pid_file.delete
end
def pid_file
@pid_file ||= PidFile.new pid_path
end
def pid
Process.pid
end
def pid_path
configuration.pid_path
end
def daemonize?
configuration.daemonize?
end
def configuration
Pug::Worker.configuration
end
end
end
end
| Make pid path configuration optional | Make pid path configuration optional
| Ruby | mit | pug-ci/pug-worker,pug-ci/pug-worker | ruby | ## Code Before:
module Pug
module Worker
class Daemon
attr_reader :application
def initialize(application)
@application = application
end
def run
start
sleep
end
private
def start
daemonize if configuration.daemonize?
trap_signals
create_pid_file
application.start
end
def stop
delete_pid_file
application.stop
end
def daemonize
Process.daemon true, true
end
def trap_signals
trap_int_signal
trap_term_signal
end
def trap_int_signal
Signal.trap 'INT' do
Thread.new { stop }.join
exit! 1
end
end
def trap_term_signal
Signal.trap 'TERM' do
Thread.new { stop }.join
exit! 0
end
end
def create_pid_file
pid_file.create pid
end
def delete_pid_file
pid_file.delete
end
def pid_file
@pid_file ||= PidFile.new configuration.pid_path
end
def pid
Process.pid
end
def configuration
Pug::Worker.configuration
end
end
end
end
## Instruction:
Make pid path configuration optional
## Code After:
module Pug
module Worker
class Daemon
attr_reader :application
def initialize(application)
@application = application
end
def run
start
sleep
end
private
def start
daemonize if daemonize?
trap_signals
create_pid_file if pid_path
application.start
end
def stop
delete_pid_file if pid_path
application.stop
end
def daemonize
Process.daemon true, true
end
def trap_signals
trap_int_signal
trap_term_signal
end
def trap_int_signal
Signal.trap 'INT' do
Thread.new { stop }.join
exit! 1
end
end
def trap_term_signal
Signal.trap 'TERM' do
Thread.new { stop }.join
exit! 0
end
end
def create_pid_file
pid_file.create pid
end
def delete_pid_file
pid_file.delete
end
def pid_file
@pid_file ||= PidFile.new pid_path
end
def pid
Process.pid
end
def pid_path
configuration.pid_path
end
def daemonize?
configuration.daemonize?
end
def configuration
Pug::Worker.configuration
end
end
end
end
| module Pug
module Worker
class Daemon
attr_reader :application
def initialize(application)
@application = application
end
def run
start
sleep
end
private
def start
- daemonize if configuration.daemonize?
? --------------
+ daemonize if daemonize?
trap_signals
- create_pid_file
+ create_pid_file if pid_path
? ++++++++++++
application.start
end
def stop
- delete_pid_file
+ delete_pid_file if pid_path
? ++++++++++++
application.stop
end
def daemonize
Process.daemon true, true
end
def trap_signals
trap_int_signal
trap_term_signal
end
def trap_int_signal
Signal.trap 'INT' do
Thread.new { stop }.join
exit! 1
end
end
def trap_term_signal
Signal.trap 'TERM' do
Thread.new { stop }.join
exit! 0
end
end
def create_pid_file
pid_file.create pid
end
def delete_pid_file
pid_file.delete
end
def pid_file
- @pid_file ||= PidFile.new configuration.pid_path
? --------------
+ @pid_file ||= PidFile.new pid_path
end
def pid
Process.pid
+ end
+
+ def pid_path
+ configuration.pid_path
+ end
+
+ def daemonize?
+ configuration.daemonize?
end
def configuration
Pug::Worker.configuration
end
end
end
end | 16 | 0.219178 | 12 | 4 |
4de72b4bd349ebf16c0046c4ed9034914c03ffb5 | cea/interfaces/dashboard/api/utils.py | cea/interfaces/dashboard/api/utils.py |
from flask import current_app
import cea.config
import cea.inputlocator
def deconstruct_parameters(p):
params = {'name': p.name, 'type': p.typename,
'value': p.get(), 'help': p.help}
if isinstance(p, cea.config.ChoiceParameter):
params['choices'] = p._choices
if p.typename == 'WeatherPathParameter':
config = current_app.cea_config
locator = cea.inputlocator.InputLocator(config.scenario)
params['choices'] = {wn: locator.get_weather(
wn) for wn in locator.get_weather_names()}
elif p.typename == 'DatabasePathParameter':
params['choices'] = p._choices
return params |
from flask import current_app
import cea.config
import cea.inputlocator
def deconstruct_parameters(p: cea.config.Parameter):
params = {'name': p.name, 'type': p.typename, 'help': p.help}
try:
params["value"] = p.get()
except cea.ConfigError as e:
print(e)
params["value"] = ""
if isinstance(p, cea.config.ChoiceParameter):
params['choices'] = p._choices
if p.typename == 'WeatherPathParameter':
config = current_app.cea_config
locator = cea.inputlocator.InputLocator(config.scenario)
params['choices'] = {wn: locator.get_weather(
wn) for wn in locator.get_weather_names()}
elif p.typename == 'DatabasePathParameter':
params['choices'] = p._choices
return params | Fix `weather_helper` bug when creating new scenario | Fix `weather_helper` bug when creating new scenario
| Python | mit | architecture-building-systems/CityEnergyAnalyst,architecture-building-systems/CityEnergyAnalyst,architecture-building-systems/CityEnergyAnalyst | python | ## Code Before:
from flask import current_app
import cea.config
import cea.inputlocator
def deconstruct_parameters(p):
params = {'name': p.name, 'type': p.typename,
'value': p.get(), 'help': p.help}
if isinstance(p, cea.config.ChoiceParameter):
params['choices'] = p._choices
if p.typename == 'WeatherPathParameter':
config = current_app.cea_config
locator = cea.inputlocator.InputLocator(config.scenario)
params['choices'] = {wn: locator.get_weather(
wn) for wn in locator.get_weather_names()}
elif p.typename == 'DatabasePathParameter':
params['choices'] = p._choices
return params
## Instruction:
Fix `weather_helper` bug when creating new scenario
## Code After:
from flask import current_app
import cea.config
import cea.inputlocator
def deconstruct_parameters(p: cea.config.Parameter):
params = {'name': p.name, 'type': p.typename, 'help': p.help}
try:
params["value"] = p.get()
except cea.ConfigError as e:
print(e)
params["value"] = ""
if isinstance(p, cea.config.ChoiceParameter):
params['choices'] = p._choices
if p.typename == 'WeatherPathParameter':
config = current_app.cea_config
locator = cea.inputlocator.InputLocator(config.scenario)
params['choices'] = {wn: locator.get_weather(
wn) for wn in locator.get_weather_names()}
elif p.typename == 'DatabasePathParameter':
params['choices'] = p._choices
return params |
from flask import current_app
import cea.config
import cea.inputlocator
- def deconstruct_parameters(p):
+ def deconstruct_parameters(p: cea.config.Parameter):
- params = {'name': p.name, 'type': p.typename,
+ params = {'name': p.name, 'type': p.typename, 'help': p.help}
? ++++++++++++++++
- 'value': p.get(), 'help': p.help}
+ try:
+ params["value"] = p.get()
+ except cea.ConfigError as e:
+ print(e)
+ params["value"] = ""
if isinstance(p, cea.config.ChoiceParameter):
params['choices'] = p._choices
if p.typename == 'WeatherPathParameter':
config = current_app.cea_config
locator = cea.inputlocator.InputLocator(config.scenario)
params['choices'] = {wn: locator.get_weather(
wn) for wn in locator.get_weather_names()}
elif p.typename == 'DatabasePathParameter':
params['choices'] = p._choices
return params | 10 | 0.454545 | 7 | 3 |
1e0d3c0d0b20f92fd901163a4f2b41627f9e931e | oonib/handlers.py | oonib/handlers.py | from cyclone import web
class OONIBHandler(web.RequestHandler):
pass
class OONIBError(web.HTTPError):
pass
| import types
from cyclone import escape
from cyclone import web
class OONIBHandler(web.RequestHandler):
def write(self, chunk):
"""
This is a monkey patch to RequestHandler to allow us to serialize also
json list objects.
"""
if isinstance(chunk, types.ListType):
chunk = escape.json_encode(chunk)
web.RequestHandler.write(self, chunk)
self.set_header("Content-Type", "application/json")
else:
web.RequestHandler.write(self, chunk)
class OONIBError(web.HTTPError):
pass
| Add support for serializing lists to json via self.write() | Add support for serializing lists to json via self.write()
| Python | bsd-2-clause | DoNotUseThisCodeJUSTFORKS/ooni-backend,dstufft/ooni-backend,DoNotUseThisCodeJUSTFORKS/ooni-backend,dstufft/ooni-backend | python | ## Code Before:
from cyclone import web
class OONIBHandler(web.RequestHandler):
pass
class OONIBError(web.HTTPError):
pass
## Instruction:
Add support for serializing lists to json via self.write()
## Code After:
import types
from cyclone import escape
from cyclone import web
class OONIBHandler(web.RequestHandler):
def write(self, chunk):
"""
This is a monkey patch to RequestHandler to allow us to serialize also
json list objects.
"""
if isinstance(chunk, types.ListType):
chunk = escape.json_encode(chunk)
web.RequestHandler.write(self, chunk)
self.set_header("Content-Type", "application/json")
else:
web.RequestHandler.write(self, chunk)
class OONIBError(web.HTTPError):
pass
| + import types
+
+ from cyclone import escape
from cyclone import web
class OONIBHandler(web.RequestHandler):
- pass
+ def write(self, chunk):
+ """
+ This is a monkey patch to RequestHandler to allow us to serialize also
+ json list objects.
+ """
+ if isinstance(chunk, types.ListType):
+ chunk = escape.json_encode(chunk)
+ web.RequestHandler.write(self, chunk)
+ self.set_header("Content-Type", "application/json")
+ else:
+ web.RequestHandler.write(self, chunk)
class OONIBError(web.HTTPError):
pass | 15 | 2.142857 | 14 | 1 |
2eda67066b9bfb2452c14a7a09f28762d1a1ec4d | app/controllers/donate.js | app/controllers/donate.js | import Ember from 'ember';
export default Ember.Controller.extend({
trey: {
musicMoney: 537,
hostingMoney: 38,
hours: 240,
},
});
| import Ember from 'ember';
export default Ember.Controller.extend({
trey: {
musicMoney: 547,
hostingMoney: 38,
hours: 200 + 79, // Estimated 200 hours spent before I started keeping track
},
});
| Update hours and money spent | Update hours and money spent
| JavaScript | bsd-3-clause | FreeMusicNinja/freemusic.ninja,FreeMusicNinja/freemusic.ninja | javascript | ## Code Before:
import Ember from 'ember';
export default Ember.Controller.extend({
trey: {
musicMoney: 537,
hostingMoney: 38,
hours: 240,
},
});
## Instruction:
Update hours and money spent
## Code After:
import Ember from 'ember';
export default Ember.Controller.extend({
trey: {
musicMoney: 547,
hostingMoney: 38,
hours: 200 + 79, // Estimated 200 hours spent before I started keeping track
},
});
| import Ember from 'ember';
export default Ember.Controller.extend({
trey: {
- musicMoney: 537,
? ^
+ musicMoney: 547,
? ^
hostingMoney: 38,
- hours: 240,
+ hours: 200 + 79, // Estimated 200 hours spent before I started keeping track
},
}); | 4 | 0.444444 | 2 | 2 |
9018d3c4f242c9bd434bc7c2253f9c8e65b052ae | pkgs/tools/networking/tinc/default.nix | pkgs/tools/networking/tinc/default.nix | {stdenv, fetchurl, lzo, openssl, zlib}:
stdenv.mkDerivation rec {
version = "1.0.23";
name = "tinc-${version}";
src = fetchurl {
url = "http://www.tinc-vpn.org/packages/tinc-${version}.tar.gz";
sha256 = "04i88hr46nx3x3s71kasm9qrjhnn35icxh9zwchki47z2vgnpw5w";
};
buildInputs = [ lzo openssl zlib ];
configureFlags = ''
--localstatedir=/var
--sysconfdir=/etc
'';
meta = {
description = "VPN daemon with full mesh routing";
longDescription = ''
tinc is a Virtual Private Network (VPN) daemon that uses tunnelling and
encryption to create a secure private network between hosts on the
Internet. It features full mesh routing, as well as encryption,
authentication, compression and ethernet bridging.
'';
homepage="http://www.tinc-vpn.org/";
license = stdenv.lib.licenses.gpl2Plus;
};
}
| {stdenv, fetchurl, lzo, openssl, zlib}:
stdenv.mkDerivation rec {
version = "1.0.24";
name = "tinc-${version}";
src = fetchurl {
url = "http://www.tinc-vpn.org/packages/tinc-${version}.tar.gz";
sha256 = "11xnz6lz917hq0zb544dvbxl0smlyjx65kv3181j4fcyygwmi3j9";
};
buildInputs = [ lzo openssl zlib ];
configureFlags = ''
--localstatedir=/var
--sysconfdir=/etc
'';
meta = {
description = "VPN daemon with full mesh routing";
longDescription = ''
tinc is a Virtual Private Network (VPN) daemon that uses tunnelling and
encryption to create a secure private network between hosts on the
Internet. It features full mesh routing, as well as encryption,
authentication, compression and ethernet bridging.
'';
homepage="http://www.tinc-vpn.org/";
license = stdenv.lib.licenses.gpl2Plus;
};
}
| Update from 1.0.23 to 1.0.24 | tinc: Update from 1.0.23 to 1.0.24
| Nix | mit | SymbiFlow/nixpkgs,triton/triton,triton/triton,SymbiFlow/nixpkgs,NixOS/nixpkgs,SymbiFlow/nixpkgs,NixOS/nixpkgs,triton/triton,SymbiFlow/nixpkgs,NixOS/nixpkgs,triton/triton,SymbiFlow/nixpkgs,triton/triton,triton/triton,SymbiFlow/nixpkgs,NixOS/nixpkgs,NixOS/nixpkgs,SymbiFlow/nixpkgs,NixOS/nixpkgs,triton/triton,SymbiFlow/nixpkgs,NixOS/nixpkgs,SymbiFlow/nixpkgs,NixOS/nixpkgs,SymbiFlow/nixpkgs,SymbiFlow/nixpkgs,NixOS/nixpkgs,NixOS/nixpkgs,NixOS/nixpkgs,SymbiFlow/nixpkgs,NixOS/nixpkgs,NixOS/nixpkgs,triton/triton,SymbiFlow/nixpkgs,NixOS/nixpkgs,NixOS/nixpkgs | nix | ## Code Before:
{stdenv, fetchurl, lzo, openssl, zlib}:
stdenv.mkDerivation rec {
version = "1.0.23";
name = "tinc-${version}";
src = fetchurl {
url = "http://www.tinc-vpn.org/packages/tinc-${version}.tar.gz";
sha256 = "04i88hr46nx3x3s71kasm9qrjhnn35icxh9zwchki47z2vgnpw5w";
};
buildInputs = [ lzo openssl zlib ];
configureFlags = ''
--localstatedir=/var
--sysconfdir=/etc
'';
meta = {
description = "VPN daemon with full mesh routing";
longDescription = ''
tinc is a Virtual Private Network (VPN) daemon that uses tunnelling and
encryption to create a secure private network between hosts on the
Internet. It features full mesh routing, as well as encryption,
authentication, compression and ethernet bridging.
'';
homepage="http://www.tinc-vpn.org/";
license = stdenv.lib.licenses.gpl2Plus;
};
}
## Instruction:
tinc: Update from 1.0.23 to 1.0.24
## Code After:
{stdenv, fetchurl, lzo, openssl, zlib}:
stdenv.mkDerivation rec {
version = "1.0.24";
name = "tinc-${version}";
src = fetchurl {
url = "http://www.tinc-vpn.org/packages/tinc-${version}.tar.gz";
sha256 = "11xnz6lz917hq0zb544dvbxl0smlyjx65kv3181j4fcyygwmi3j9";
};
buildInputs = [ lzo openssl zlib ];
configureFlags = ''
--localstatedir=/var
--sysconfdir=/etc
'';
meta = {
description = "VPN daemon with full mesh routing";
longDescription = ''
tinc is a Virtual Private Network (VPN) daemon that uses tunnelling and
encryption to create a secure private network between hosts on the
Internet. It features full mesh routing, as well as encryption,
authentication, compression and ethernet bridging.
'';
homepage="http://www.tinc-vpn.org/";
license = stdenv.lib.licenses.gpl2Plus;
};
}
| {stdenv, fetchurl, lzo, openssl, zlib}:
stdenv.mkDerivation rec {
- version = "1.0.23";
? ^
+ version = "1.0.24";
? ^
name = "tinc-${version}";
src = fetchurl {
url = "http://www.tinc-vpn.org/packages/tinc-${version}.tar.gz";
- sha256 = "04i88hr46nx3x3s71kasm9qrjhnn35icxh9zwchki47z2vgnpw5w";
+ sha256 = "11xnz6lz917hq0zb544dvbxl0smlyjx65kv3181j4fcyygwmi3j9";
};
buildInputs = [ lzo openssl zlib ];
configureFlags = ''
--localstatedir=/var
--sysconfdir=/etc
'';
meta = {
description = "VPN daemon with full mesh routing";
longDescription = ''
tinc is a Virtual Private Network (VPN) daemon that uses tunnelling and
encryption to create a secure private network between hosts on the
Internet. It features full mesh routing, as well as encryption,
authentication, compression and ethernet bridging.
'';
homepage="http://www.tinc-vpn.org/";
license = stdenv.lib.licenses.gpl2Plus;
};
} | 4 | 0.133333 | 2 | 2 |
7294030ab5f8cde285ebc3c68c7e0d28d841d083 | lib/tentd-admin/views/_profile_type_fields.slim | lib/tentd-admin/views/_profile_type_fields.slim | - input_name = "#{type}[#{name}]"
- if value.kind_of? Array
- input_name << '[]'
label.control-label for=input_name = name
.controls
select.input.input-xxlarge name=input_name multiple=true
- value.each do |option|
option value=option selected=true = option
- elsif value.kind_of?(Hash)
- value.each_pair do |name, value|
.control-group.span8
legend = name
== slim :_profile_type_fields, :locals => { :type => type, :name => name, :value => value }
- elsif value.kind_of?(TrueClass) || value.kind_of?(FalseClass)
label.checkbox
input type='checkbox' name=input_name value='true' checked=value = name
- else
label.control-label for=input_name = name
.controls
input.input.input-xxlarge type='text' name=input_name value=value
| - input_name = "#{type}[#{name}]"
- if value.kind_of? Array
- input_name << '[]'
label.control-label for=input_name = name
.controls
select.input.input-xxlarge name=input_name multiple=true
- value.each do |option|
option value=option selected=true = option
- elsif value.kind_of?(Hash)
- value.each_pair do |name, value|
- next if %w( groups entities public ).include?(name.to_s)
.control-group.span8
legend = name
== slim :_profile_type_fields, :locals => { :type => type, :name => name, :value => value }
- elsif value.kind_of?(TrueClass) || value.kind_of?(FalseClass)
- else
label.control-label for=input_name = name
.controls
input.input.input-xxlarge type='text' name=input_name value=value
| Remove fields from update profile (public entities groups) | Remove fields from update profile (public entities groups)
| Slim | bsd-3-clause | tent/tentd-admin,tent/tentd-admin | slim | ## Code Before:
- input_name = "#{type}[#{name}]"
- if value.kind_of? Array
- input_name << '[]'
label.control-label for=input_name = name
.controls
select.input.input-xxlarge name=input_name multiple=true
- value.each do |option|
option value=option selected=true = option
- elsif value.kind_of?(Hash)
- value.each_pair do |name, value|
.control-group.span8
legend = name
== slim :_profile_type_fields, :locals => { :type => type, :name => name, :value => value }
- elsif value.kind_of?(TrueClass) || value.kind_of?(FalseClass)
label.checkbox
input type='checkbox' name=input_name value='true' checked=value = name
- else
label.control-label for=input_name = name
.controls
input.input.input-xxlarge type='text' name=input_name value=value
## Instruction:
Remove fields from update profile (public entities groups)
## Code After:
- input_name = "#{type}[#{name}]"
- if value.kind_of? Array
- input_name << '[]'
label.control-label for=input_name = name
.controls
select.input.input-xxlarge name=input_name multiple=true
- value.each do |option|
option value=option selected=true = option
- elsif value.kind_of?(Hash)
- value.each_pair do |name, value|
- next if %w( groups entities public ).include?(name.to_s)
.control-group.span8
legend = name
== slim :_profile_type_fields, :locals => { :type => type, :name => name, :value => value }
- elsif value.kind_of?(TrueClass) || value.kind_of?(FalseClass)
- else
label.control-label for=input_name = name
.controls
input.input.input-xxlarge type='text' name=input_name value=value
| - input_name = "#{type}[#{name}]"
- if value.kind_of? Array
- input_name << '[]'
label.control-label for=input_name = name
.controls
select.input.input-xxlarge name=input_name multiple=true
- value.each do |option|
option value=option selected=true = option
- elsif value.kind_of?(Hash)
- value.each_pair do |name, value|
+ - next if %w( groups entities public ).include?(name.to_s)
.control-group.span8
legend = name
== slim :_profile_type_fields, :locals => { :type => type, :name => name, :value => value }
- elsif value.kind_of?(TrueClass) || value.kind_of?(FalseClass)
- label.checkbox
- input type='checkbox' name=input_name value='true' checked=value = name
- else
label.control-label for=input_name = name
.controls
input.input.input-xxlarge type='text' name=input_name value=value | 3 | 0.15 | 1 | 2 |
56e3cd1b22cdbb9f0af836cc7b44803c4c45a5ca | app/views/accounts/admin.html.erb | app/views/accounts/admin.html.erb | <div class="container order form">
<h1 class="page-title">Manage Accounts / Admin</h1>
<%= form_tag(accounts_path, method: "get", id: "account-form") do %>
<%= select("account", "account_id", Account.all.collect {|a| [ "#{a.first_name} #{a.last_name} #{ '| ' + a.company if a.company.present?}", a.id ] } + [["View All Accounts"]], {prompt: 'Select Account'}) %>
<%= submit_tag "View Account", :name => nil, class: "btn btn-primary"%>
<% end %>
</div>
<div class="js-account"></div>
<div class="js-order"></div>
| <div class="container order form">
<h1 class="page-title">Manage Accounts</h1>
<%= form_tag(accounts_path, method: "get", id: "account-form") do %>
<%= select("account", "account_id", Account.all.collect {|a| [ "#{a.first_name} #{a.last_name} #{ '| ' + a.company if a.company.present?}", a.id ] } + [["View All Accounts"]], {prompt: 'Select Account'}) %>
<%= submit_tag "View Account", :name => nil, class: "btn btn-primary"%>
<% end %>
<div id="js-panel">
<div class="js-account"></div>
</div>
</div>
| Change title add classes and ids for displaying js | Change title add classes and ids for displaying js
| HTML+ERB | mit | Dom-Mc/fetch_it,Dom-Mc/fetch_it,Dom-Mc/fetch_it | html+erb | ## Code Before:
<div class="container order form">
<h1 class="page-title">Manage Accounts / Admin</h1>
<%= form_tag(accounts_path, method: "get", id: "account-form") do %>
<%= select("account", "account_id", Account.all.collect {|a| [ "#{a.first_name} #{a.last_name} #{ '| ' + a.company if a.company.present?}", a.id ] } + [["View All Accounts"]], {prompt: 'Select Account'}) %>
<%= submit_tag "View Account", :name => nil, class: "btn btn-primary"%>
<% end %>
</div>
<div class="js-account"></div>
<div class="js-order"></div>
## Instruction:
Change title add classes and ids for displaying js
## Code After:
<div class="container order form">
<h1 class="page-title">Manage Accounts</h1>
<%= form_tag(accounts_path, method: "get", id: "account-form") do %>
<%= select("account", "account_id", Account.all.collect {|a| [ "#{a.first_name} #{a.last_name} #{ '| ' + a.company if a.company.present?}", a.id ] } + [["View All Accounts"]], {prompt: 'Select Account'}) %>
<%= submit_tag "View Account", :name => nil, class: "btn btn-primary"%>
<% end %>
<div id="js-panel">
<div class="js-account"></div>
</div>
</div>
| <div class="container order form">
- <h1 class="page-title">Manage Accounts / Admin</h1>
? --------
+ <h1 class="page-title">Manage Accounts</h1>
<%= form_tag(accounts_path, method: "get", id: "account-form") do %>
- <%= select("account", "account_id", Account.all.collect {|a| [ "#{a.first_name} #{a.last_name} #{ '| ' + a.company if a.company.present?}", a.id ] } + [["View All Accounts"]], {prompt: 'Select Account'}) %>
+ <%= select("account", "account_id", Account.all.collect {|a| [ "#{a.first_name} #{a.last_name} #{ '| ' + a.company if a.company.present?}", a.id ] } + [["View All Accounts"]], {prompt: 'Select Account'}) %>
? ++
- <%= submit_tag "View Account", :name => nil, class: "btn btn-primary"%>
+ <%= submit_tag "View Account", :name => nil, class: "btn btn-primary"%>
? ++
<% end %>
+ <div id="js-panel">
+ <div class="js-account"></div>
+ </div>
+
</div>
-
- <div class="js-account"></div>
- <div class="js-order"></div> | 13 | 0.928571 | 7 | 6 |
48c43c16db64fbfcb8b52161e0850f9b31bead04 | spec/unit/veritas/immutable/module_methods/memoize_spec.rb | spec/unit/veritas/immutable/module_methods/memoize_spec.rb | require 'spec_helper'
require File.expand_path('../../fixtures/classes', __FILE__)
shared_examples_for 'memoizes method' do
it 'memoizes the instance method' do
subject
object = klass.new
object.send(method).should equal(object.send(method))
end
end
describe 'Veritas::Immutable::ModuleMethods#memoize' do
subject { klass.memoize(method) }
let(:klass) { Class.new(ImmutableSpecs::Object) }
context 'public method' do
let(:method) { :public_method }
it { should equal(klass) }
it_should_behave_like 'memoizes method'
it 'is still a public method' do
should be_public_method_defined(method)
end
end
context 'protected method' do
let(:method) { :protected_method }
it { should equal(klass) }
it_should_behave_like 'memoizes method'
it 'is still a protected method' do
should be_protected_method_defined(method)
end
end
context 'private method' do
let(:method) { :private_method }
it { should equal(klass) }
it_should_behave_like 'memoizes method'
it 'is still a private method' do
should be_private_method_defined(method)
end
end
end
| require 'spec_helper'
require File.expand_path('../../fixtures/classes', __FILE__)
shared_examples_for 'memoizes method' do
it 'memoizes the instance method' do
subject
object = klass.new
object.send(method).should equal(object.send(method))
end
it 'adds a private method' do
count = klass.private_instance_methods.count
expect { subject }.to change { klass.private_instance_methods.count }.from(count).to(count + 1)
end
end
describe 'Veritas::Immutable::ModuleMethods#memoize' do
subject { klass.memoize(method) }
let(:klass) { Class.new(ImmutableSpecs::Object) }
context 'public method' do
let(:method) { :public_method }
it { should equal(klass) }
it_should_behave_like 'memoizes method'
it 'is still a public method' do
should be_public_method_defined(method)
end
end
context 'protected method' do
let(:method) { :protected_method }
it { should equal(klass) }
it_should_behave_like 'memoizes method'
it 'is still a protected method' do
should be_protected_method_defined(method)
end
end
context 'private method' do
let(:method) { :private_method }
it { should equal(klass) }
it_should_behave_like 'memoizes method'
it 'is still a private method' do
should be_private_method_defined(method)
end
end
end
| Add a spec to ensure that only one private method is added by memoization | Add a spec to ensure that only one private method is added by memoization
| Ruby | mit | dkubb/axiom | ruby | ## Code Before:
require 'spec_helper'
require File.expand_path('../../fixtures/classes', __FILE__)
shared_examples_for 'memoizes method' do
it 'memoizes the instance method' do
subject
object = klass.new
object.send(method).should equal(object.send(method))
end
end
describe 'Veritas::Immutable::ModuleMethods#memoize' do
subject { klass.memoize(method) }
let(:klass) { Class.new(ImmutableSpecs::Object) }
context 'public method' do
let(:method) { :public_method }
it { should equal(klass) }
it_should_behave_like 'memoizes method'
it 'is still a public method' do
should be_public_method_defined(method)
end
end
context 'protected method' do
let(:method) { :protected_method }
it { should equal(klass) }
it_should_behave_like 'memoizes method'
it 'is still a protected method' do
should be_protected_method_defined(method)
end
end
context 'private method' do
let(:method) { :private_method }
it { should equal(klass) }
it_should_behave_like 'memoizes method'
it 'is still a private method' do
should be_private_method_defined(method)
end
end
end
## Instruction:
Add a spec to ensure that only one private method is added by memoization
## Code After:
require 'spec_helper'
require File.expand_path('../../fixtures/classes', __FILE__)
shared_examples_for 'memoizes method' do
it 'memoizes the instance method' do
subject
object = klass.new
object.send(method).should equal(object.send(method))
end
it 'adds a private method' do
count = klass.private_instance_methods.count
expect { subject }.to change { klass.private_instance_methods.count }.from(count).to(count + 1)
end
end
describe 'Veritas::Immutable::ModuleMethods#memoize' do
subject { klass.memoize(method) }
let(:klass) { Class.new(ImmutableSpecs::Object) }
context 'public method' do
let(:method) { :public_method }
it { should equal(klass) }
it_should_behave_like 'memoizes method'
it 'is still a public method' do
should be_public_method_defined(method)
end
end
context 'protected method' do
let(:method) { :protected_method }
it { should equal(klass) }
it_should_behave_like 'memoizes method'
it 'is still a protected method' do
should be_protected_method_defined(method)
end
end
context 'private method' do
let(:method) { :private_method }
it { should equal(klass) }
it_should_behave_like 'memoizes method'
it 'is still a private method' do
should be_private_method_defined(method)
end
end
end
| require 'spec_helper'
require File.expand_path('../../fixtures/classes', __FILE__)
shared_examples_for 'memoizes method' do
it 'memoizes the instance method' do
subject
object = klass.new
object.send(method).should equal(object.send(method))
+ end
+
+ it 'adds a private method' do
+ count = klass.private_instance_methods.count
+ expect { subject }.to change { klass.private_instance_methods.count }.from(count).to(count + 1)
end
end
describe 'Veritas::Immutable::ModuleMethods#memoize' do
subject { klass.memoize(method) }
let(:klass) { Class.new(ImmutableSpecs::Object) }
context 'public method' do
let(:method) { :public_method }
it { should equal(klass) }
it_should_behave_like 'memoizes method'
it 'is still a public method' do
should be_public_method_defined(method)
end
end
context 'protected method' do
let(:method) { :protected_method }
it { should equal(klass) }
it_should_behave_like 'memoizes method'
it 'is still a protected method' do
should be_protected_method_defined(method)
end
end
context 'private method' do
let(:method) { :private_method }
it { should equal(klass) }
it_should_behave_like 'memoizes method'
it 'is still a private method' do
should be_private_method_defined(method)
end
end
end | 5 | 0.096154 | 5 | 0 |
e598ae5f6d08227171a7c5d16688db769a4748ad | lib/tasks/cleanup.rake | lib/tasks/cleanup.rake | namespace :cleanup do
desc 'Clean up all message redelivery and destroy actions from the holding pen to make admin actions there faster'
task :holding_pen => :environment do
dryrun = ENV['DRYRUN'] != '0' if ENV['DRYRUN']
if dryrun
$stderr.puts "This is a dryrun - nothing will be deleted"
end
holding_pen = InfoRequest.find_by_url_title('holding_pen')
old_events = holding_pen.info_request_events.find_each(:conditions => ['event_type in (?)',
['redeliver_incoming',
'destroy_incoming']]) do |event|
puts event.inspect
if not dryrun
event.destroy
end
end
end
end
| namespace :cleanup do
desc 'Clean up all message redelivery and destroy actions from the holding pen to make admin actions there faster'
task :holding_pen => :environment do
dryrun = ENV['DRYRUN'] != '0' if ENV['DRYRUN']
if dryrun
$stderr.puts "This is a dryrun - nothing will be deleted"
end
holding_pen = InfoRequest.holding_pen_request
old_events = holding_pen.info_request_events.find_each(:conditions => ['event_type in (?)',
['redeliver_incoming',
'destroy_incoming']]) do |event|
puts event.inspect
if not dryrun
event.destroy
end
end
end
end
| Use InfoRequest.holding_pen_request to find the holding pen | Use InfoRequest.holding_pen_request to find the holding pen
Then it is automatically created if it does not already exist
| Ruby | agpl-3.0 | andreicristianpetcu/alaveteli_old,andreicristianpetcu/alaveteli_old,andreicristianpetcu/alaveteli,andreicristianpetcu/alaveteli,andreicristianpetcu/alaveteli,andreicristianpetcu/alaveteli_old,andreicristianpetcu/alaveteli_old,andreicristianpetcu/alaveteli,andreicristianpetcu/alaveteli,andreicristianpetcu/alaveteli_old | ruby | ## Code Before:
namespace :cleanup do
desc 'Clean up all message redelivery and destroy actions from the holding pen to make admin actions there faster'
task :holding_pen => :environment do
dryrun = ENV['DRYRUN'] != '0' if ENV['DRYRUN']
if dryrun
$stderr.puts "This is a dryrun - nothing will be deleted"
end
holding_pen = InfoRequest.find_by_url_title('holding_pen')
old_events = holding_pen.info_request_events.find_each(:conditions => ['event_type in (?)',
['redeliver_incoming',
'destroy_incoming']]) do |event|
puts event.inspect
if not dryrun
event.destroy
end
end
end
end
## Instruction:
Use InfoRequest.holding_pen_request to find the holding pen
Then it is automatically created if it does not already exist
## Code After:
namespace :cleanup do
desc 'Clean up all message redelivery and destroy actions from the holding pen to make admin actions there faster'
task :holding_pen => :environment do
dryrun = ENV['DRYRUN'] != '0' if ENV['DRYRUN']
if dryrun
$stderr.puts "This is a dryrun - nothing will be deleted"
end
holding_pen = InfoRequest.holding_pen_request
old_events = holding_pen.info_request_events.find_each(:conditions => ['event_type in (?)',
['redeliver_incoming',
'destroy_incoming']]) do |event|
puts event.inspect
if not dryrun
event.destroy
end
end
end
end
| namespace :cleanup do
desc 'Clean up all message redelivery and destroy actions from the holding pen to make admin actions there faster'
task :holding_pen => :environment do
dryrun = ENV['DRYRUN'] != '0' if ENV['DRYRUN']
if dryrun
$stderr.puts "This is a dryrun - nothing will be deleted"
end
- holding_pen = InfoRequest.find_by_url_title('holding_pen')
+ holding_pen = InfoRequest.holding_pen_request
old_events = holding_pen.info_request_events.find_each(:conditions => ['event_type in (?)',
['redeliver_incoming',
'destroy_incoming']]) do |event|
puts event.inspect
if not dryrun
event.destroy
end
end
end
end | 2 | 0.1 | 1 | 1 |
64b61f5f83da81ab3f1fb33afb24ca8fc2eeec0b | check-all.sh | check-all.sh | set -e
dartanalyzer lib/postgresql.dart
dartanalyzer lib/pool.dart
dartanalyzer test/postgresql_test.dart
dartanalyzer test/postgresql_pool_test.dart
dartanalyzer test/substitute_test.dart
dart --checked test/substitute_test.dart
dart --checked test/settings_test.dart
dart --checked test/postgresql_test.dart
#dart --checked test/postgresql_mock_test.dart
dart --checked test/postgresql_pool_test.dart
| set -e
dartanalyzer lib/postgresql.dart
dartanalyzer lib/pool.dart
dartanalyzer test/postgresql_test.dart
dartanalyzer test/postgresql_pool_test.dart
dartanalyzer test/substitute_test.dart
dart --checked test/substitute_test.dart
dart --checked test/settings_test.dart
dart --checked test/type_converter_test.dart
dart --checked test/postgresql_test.dart
#dart --checked test/postgresql_mock_test.dart
dart --checked test/postgresql_pool_test.dart
| Add type converter test to automated build | Add type converter test to automated build
| Shell | bsd-2-clause | xxgreg/dart_postgresql,xxgreg/dart_postgresql,tomyeh/postgresql,tomyeh/postgresql | shell | ## Code Before:
set -e
dartanalyzer lib/postgresql.dart
dartanalyzer lib/pool.dart
dartanalyzer test/postgresql_test.dart
dartanalyzer test/postgresql_pool_test.dart
dartanalyzer test/substitute_test.dart
dart --checked test/substitute_test.dart
dart --checked test/settings_test.dart
dart --checked test/postgresql_test.dart
#dart --checked test/postgresql_mock_test.dart
dart --checked test/postgresql_pool_test.dart
## Instruction:
Add type converter test to automated build
## Code After:
set -e
dartanalyzer lib/postgresql.dart
dartanalyzer lib/pool.dart
dartanalyzer test/postgresql_test.dart
dartanalyzer test/postgresql_pool_test.dart
dartanalyzer test/substitute_test.dart
dart --checked test/substitute_test.dart
dart --checked test/settings_test.dart
dart --checked test/type_converter_test.dart
dart --checked test/postgresql_test.dart
#dart --checked test/postgresql_mock_test.dart
dart --checked test/postgresql_pool_test.dart
| set -e
dartanalyzer lib/postgresql.dart
dartanalyzer lib/pool.dart
dartanalyzer test/postgresql_test.dart
dartanalyzer test/postgresql_pool_test.dart
dartanalyzer test/substitute_test.dart
dart --checked test/substitute_test.dart
dart --checked test/settings_test.dart
+ dart --checked test/type_converter_test.dart
dart --checked test/postgresql_test.dart
#dart --checked test/postgresql_mock_test.dart
dart --checked test/postgresql_pool_test.dart
| 1 | 0.066667 | 1 | 0 |
58c37d761a62100016a92786ae8620006143cf3d | cypress/integration/mentions.ts | cypress/integration/mentions.ts | const DELAY = 500;
context("Mentions", () => {
beforeEach(() => {
cy.visit("http://localhost:4000");
});
it("'@ + Enter' should select the first suggestion", () => {
cy.get(".mde-text")
.type("{selectall}{backspace}")
.type("@")
.wait(DELAY)
.type("{enter}")
.should("have.value", "@andre ");
});
it("'@ + Arrow down + Enter' should select the first suggestion", () => {
cy.get(".mde-text")
.type("{selectall}{backspace}")
.type("@")
.wait(DELAY)
.type("{downArrow}")
.type("{enter}")
.should("have.value", "@angela ");
});
it("'@ + 4x Arrow down + Enter' should cycle and select the first suggestion", () => {
cy.get(".mde-text")
.type("{selectall}{backspace}")
.type("@")
.wait(DELAY)
.type("{downArrow}{downArrow}{downArrow}{downArrow}")
.type("{enter}")
.should("have.value", "@andre ");
});
});
| const DELAY = 500;
context("Suggestions", () => {
beforeEach(() => {
cy.visit("http://localhost:4000");
});
it("'@' should display the suggestions box", () => {
cy.get(".mde-suggestions").should("not.exist");
cy.get(".mde-text")
.type("{selectall}{backspace}")
.type("@");
cy.get(".mde-suggestions").should("exist");
});
it("'@ + Enter' should select the first suggestion", () => {
cy.get(".mde-text")
.type("{selectall}{backspace}")
.type("@")
.wait(DELAY)
.type("{enter}")
.should("have.value", "@andre ");
});
it("'@ + Arrow down + Enter' should select the first suggestion", () => {
cy.get(".mde-text")
.type("{selectall}{backspace}")
.type("@")
.wait(DELAY)
.type("{downArrow}")
.type("{enter}")
.should("have.value", "@angela ");
});
it("'@ + 4x Arrow down + Enter' should cycle and select the first suggestion", () => {
cy.get(".mde-text")
.type("{selectall}{backspace}")
.type("@")
.wait(DELAY)
.type("{downArrow}{downArrow}{downArrow}{downArrow}")
.type("{enter}")
.should("have.value", "@andre ");
});
});
| Add a test for suggestions | Add a test for suggestions
| TypeScript | mit | andrerpena/react-mde,andrerpena/react-mde,andrerpena/react-mde | typescript | ## Code Before:
const DELAY = 500;
context("Mentions", () => {
beforeEach(() => {
cy.visit("http://localhost:4000");
});
it("'@ + Enter' should select the first suggestion", () => {
cy.get(".mde-text")
.type("{selectall}{backspace}")
.type("@")
.wait(DELAY)
.type("{enter}")
.should("have.value", "@andre ");
});
it("'@ + Arrow down + Enter' should select the first suggestion", () => {
cy.get(".mde-text")
.type("{selectall}{backspace}")
.type("@")
.wait(DELAY)
.type("{downArrow}")
.type("{enter}")
.should("have.value", "@angela ");
});
it("'@ + 4x Arrow down + Enter' should cycle and select the first suggestion", () => {
cy.get(".mde-text")
.type("{selectall}{backspace}")
.type("@")
.wait(DELAY)
.type("{downArrow}{downArrow}{downArrow}{downArrow}")
.type("{enter}")
.should("have.value", "@andre ");
});
});
## Instruction:
Add a test for suggestions
## Code After:
const DELAY = 500;
context("Suggestions", () => {
beforeEach(() => {
cy.visit("http://localhost:4000");
});
it("'@' should display the suggestions box", () => {
cy.get(".mde-suggestions").should("not.exist");
cy.get(".mde-text")
.type("{selectall}{backspace}")
.type("@");
cy.get(".mde-suggestions").should("exist");
});
it("'@ + Enter' should select the first suggestion", () => {
cy.get(".mde-text")
.type("{selectall}{backspace}")
.type("@")
.wait(DELAY)
.type("{enter}")
.should("have.value", "@andre ");
});
it("'@ + Arrow down + Enter' should select the first suggestion", () => {
cy.get(".mde-text")
.type("{selectall}{backspace}")
.type("@")
.wait(DELAY)
.type("{downArrow}")
.type("{enter}")
.should("have.value", "@angela ");
});
it("'@ + 4x Arrow down + Enter' should cycle and select the first suggestion", () => {
cy.get(".mde-text")
.type("{selectall}{backspace}")
.type("@")
.wait(DELAY)
.type("{downArrow}{downArrow}{downArrow}{downArrow}")
.type("{enter}")
.should("have.value", "@andre ");
});
});
| const DELAY = 500;
- context("Mentions", () => {
? ^ ^
+ context("Suggestions", () => {
? ^^^^ ^
beforeEach(() => {
cy.visit("http://localhost:4000");
+ });
+ it("'@' should display the suggestions box", () => {
+ cy.get(".mde-suggestions").should("not.exist");
+ cy.get(".mde-text")
+ .type("{selectall}{backspace}")
+ .type("@");
+ cy.get(".mde-suggestions").should("exist");
});
it("'@ + Enter' should select the first suggestion", () => {
cy.get(".mde-text")
.type("{selectall}{backspace}")
.type("@")
.wait(DELAY)
.type("{enter}")
.should("have.value", "@andre ");
});
it("'@ + Arrow down + Enter' should select the first suggestion", () => {
cy.get(".mde-text")
.type("{selectall}{backspace}")
.type("@")
.wait(DELAY)
.type("{downArrow}")
.type("{enter}")
.should("have.value", "@angela ");
});
it("'@ + 4x Arrow down + Enter' should cycle and select the first suggestion", () => {
cy.get(".mde-text")
.type("{selectall}{backspace}")
.type("@")
.wait(DELAY)
.type("{downArrow}{downArrow}{downArrow}{downArrow}")
.type("{enter}")
.should("have.value", "@andre ");
});
}); | 9 | 0.272727 | 8 | 1 |
d74a77463bac22cf76a8f5dd99202fb1c4b4f556 | Mk/dports.tea.mk | Mk/dports.tea.mk | .c.o:
${CC} -c -DUSE_TCL_STUBS ${CFLAGS} ${TCL_DEFS} ${SHLIB_CFLAGS} $< -o $@
$(SHLIB_NAME):: ${OBJS}
${SHLIB_LD} ${OBJS} -o ${SHLIB_NAME} ${TCL_STUB_LIB_SPEC} ${LIBS}
all:: ${SHLIB_NAME}
clean::
rm -f ${OBJS} ${SHLIB_NAME} so_locations
distclean:: clean
install:: all
$(INSTALL) -d -o ${DSTUSR} -g ${DSTGRP} -m 775 ${INSTALLDIR}
$(INSTALL) -o ${DSTUSR} -g ${DSTGRP} -m 444 ${SHLIB_NAME} ${INSTALLDIR}
$(SILENT) $(TCLSH) ../pkg_mkindex.tcl ${INSTALLDIR}
| .c.o:
${CC} -c -DUSE_TCL_STUBS ${CFLAGS} ${TCL_DEFS} ${SHLIB_CFLAGS} $< -o $@
$(SHLIB_NAME):: ${OBJS}
${SHLIB_LD} ${OBJS} -o ${SHLIB_NAME} ${TCL_STUB_LIB_SPEC} ${SHLIB_LDFLAGS} ${LIBS}
all:: ${SHLIB_NAME}
clean::
rm -f ${OBJS} ${SHLIB_NAME} so_locations
distclean:: clean
install:: all
$(INSTALL) -d -o ${DSTUSR} -g ${DSTGRP} -m 775 ${INSTALLDIR}
$(INSTALL) -o ${DSTUSR} -g ${DSTGRP} -m 444 ${SHLIB_NAME} ${INSTALLDIR}
$(SILENT) $(TCLSH) ../pkg_mkindex.tcl ${INSTALLDIR}
| Use SHLIB_LDFLAGS when linking Pextlib | Use SHLIB_LDFLAGS when linking Pextlib
git-svn-id: 620571fa9b4bd0cbce9a0cf901e91ef896adbf27@4215 d073be05-634f-4543-b044-5fe20cf6d1d6
| Makefile | bsd-3-clause | neverpanic/macports-base,danchr/macports-base,neverpanic/macports-base,macports/macports-base,macports/macports-base,danchr/macports-base | makefile | ## Code Before:
.c.o:
${CC} -c -DUSE_TCL_STUBS ${CFLAGS} ${TCL_DEFS} ${SHLIB_CFLAGS} $< -o $@
$(SHLIB_NAME):: ${OBJS}
${SHLIB_LD} ${OBJS} -o ${SHLIB_NAME} ${TCL_STUB_LIB_SPEC} ${LIBS}
all:: ${SHLIB_NAME}
clean::
rm -f ${OBJS} ${SHLIB_NAME} so_locations
distclean:: clean
install:: all
$(INSTALL) -d -o ${DSTUSR} -g ${DSTGRP} -m 775 ${INSTALLDIR}
$(INSTALL) -o ${DSTUSR} -g ${DSTGRP} -m 444 ${SHLIB_NAME} ${INSTALLDIR}
$(SILENT) $(TCLSH) ../pkg_mkindex.tcl ${INSTALLDIR}
## Instruction:
Use SHLIB_LDFLAGS when linking Pextlib
git-svn-id: 620571fa9b4bd0cbce9a0cf901e91ef896adbf27@4215 d073be05-634f-4543-b044-5fe20cf6d1d6
## Code After:
.c.o:
${CC} -c -DUSE_TCL_STUBS ${CFLAGS} ${TCL_DEFS} ${SHLIB_CFLAGS} $< -o $@
$(SHLIB_NAME):: ${OBJS}
${SHLIB_LD} ${OBJS} -o ${SHLIB_NAME} ${TCL_STUB_LIB_SPEC} ${SHLIB_LDFLAGS} ${LIBS}
all:: ${SHLIB_NAME}
clean::
rm -f ${OBJS} ${SHLIB_NAME} so_locations
distclean:: clean
install:: all
$(INSTALL) -d -o ${DSTUSR} -g ${DSTGRP} -m 775 ${INSTALLDIR}
$(INSTALL) -o ${DSTUSR} -g ${DSTGRP} -m 444 ${SHLIB_NAME} ${INSTALLDIR}
$(SILENT) $(TCLSH) ../pkg_mkindex.tcl ${INSTALLDIR}
| .c.o:
${CC} -c -DUSE_TCL_STUBS ${CFLAGS} ${TCL_DEFS} ${SHLIB_CFLAGS} $< -o $@
$(SHLIB_NAME):: ${OBJS}
- ${SHLIB_LD} ${OBJS} -o ${SHLIB_NAME} ${TCL_STUB_LIB_SPEC} ${LIBS}
+ ${SHLIB_LD} ${OBJS} -o ${SHLIB_NAME} ${TCL_STUB_LIB_SPEC} ${SHLIB_LDFLAGS} ${LIBS}
? +++++++++++++++++
all:: ${SHLIB_NAME}
clean::
rm -f ${OBJS} ${SHLIB_NAME} so_locations
distclean:: clean
install:: all
$(INSTALL) -d -o ${DSTUSR} -g ${DSTGRP} -m 775 ${INSTALLDIR}
$(INSTALL) -o ${DSTUSR} -g ${DSTGRP} -m 444 ${SHLIB_NAME} ${INSTALLDIR}
$(SILENT) $(TCLSH) ../pkg_mkindex.tcl ${INSTALLDIR} | 2 | 0.117647 | 1 | 1 |
b39f6714bb4b8d8faa9a91e5efd719c54473556c | alfred/install.sh | alfred/install.sh | set -e
dirs=(appearence features)
for dir in "${dirs[@]}"
do
rm -rf ~/Library/Application\ Support/Alfred/Alfred.alfredpreferences/preferences/$dir || true
ln -s $ZSH/alfred/appearence ~/Library/Application\ Support/Alfred/Alfred.alfredpreferences/preferences/$dir
done
localdir=$(/bin/ls -1 ~/Library/Application\ Support/Alfred/Alfred.alfredpreferences/preferences/local)
rm -rf ~/Library/Application\ Support/Alfred/Alfred.alfredpreferences/preferences/local/$localdir/*
dirs=(appearence features hotkey)
for dir in "${dirs[@]}"
do
ln -s $ZSH/alfred/local/appearence ~/Library/Application\ Support/Alfred/Alfred.alfredpreferences/preferences/local/$localdir/$dir
done
| set -e
mkdir -p ~/Library/Application\ Support/Alfred/Alfred.alfredpreferences/preferences
dirs=(appearence features)
for dir in "${dirs[@]}"
do
rm -rf ~/Library/Application\ Support/Alfred/Alfred.alfredpreferences/preferences/$dir || true
ln -s $ZSH/alfred/appearence ~/Library/Application\ Support/Alfred/Alfred.alfredpreferences/preferences/$dir
done
localdir=$(/bin/ls -1 ~/Library/Application\ Support/Alfred/Alfred.alfredpreferences/preferences/local)
rm -rf ~/Library/Application\ Support/Alfred/Alfred.alfredpreferences/preferences/local/$localdir/*
dirs=(appearence features hotkey)
for dir in "${dirs[@]}"
do
ln -s $ZSH/alfred/local/appearence ~/Library/Application\ Support/Alfred/Alfred.alfredpreferences/preferences/local/$localdir/$dir
done
| Create alfred pref dir if missing | Create alfred pref dir if missing
| Shell | mit | eexit/dotfiles,eexit/dotfiles,eexit/dotfiles | shell | ## Code Before:
set -e
dirs=(appearence features)
for dir in "${dirs[@]}"
do
rm -rf ~/Library/Application\ Support/Alfred/Alfred.alfredpreferences/preferences/$dir || true
ln -s $ZSH/alfred/appearence ~/Library/Application\ Support/Alfred/Alfred.alfredpreferences/preferences/$dir
done
localdir=$(/bin/ls -1 ~/Library/Application\ Support/Alfred/Alfred.alfredpreferences/preferences/local)
rm -rf ~/Library/Application\ Support/Alfred/Alfred.alfredpreferences/preferences/local/$localdir/*
dirs=(appearence features hotkey)
for dir in "${dirs[@]}"
do
ln -s $ZSH/alfred/local/appearence ~/Library/Application\ Support/Alfred/Alfred.alfredpreferences/preferences/local/$localdir/$dir
done
## Instruction:
Create alfred pref dir if missing
## Code After:
set -e
mkdir -p ~/Library/Application\ Support/Alfred/Alfred.alfredpreferences/preferences
dirs=(appearence features)
for dir in "${dirs[@]}"
do
rm -rf ~/Library/Application\ Support/Alfred/Alfred.alfredpreferences/preferences/$dir || true
ln -s $ZSH/alfred/appearence ~/Library/Application\ Support/Alfred/Alfred.alfredpreferences/preferences/$dir
done
localdir=$(/bin/ls -1 ~/Library/Application\ Support/Alfred/Alfred.alfredpreferences/preferences/local)
rm -rf ~/Library/Application\ Support/Alfred/Alfred.alfredpreferences/preferences/local/$localdir/*
dirs=(appearence features hotkey)
for dir in "${dirs[@]}"
do
ln -s $ZSH/alfred/local/appearence ~/Library/Application\ Support/Alfred/Alfred.alfredpreferences/preferences/local/$localdir/$dir
done
| set -e
+
+ mkdir -p ~/Library/Application\ Support/Alfred/Alfred.alfredpreferences/preferences
dirs=(appearence features)
for dir in "${dirs[@]}"
do
rm -rf ~/Library/Application\ Support/Alfred/Alfred.alfredpreferences/preferences/$dir || true
ln -s $ZSH/alfred/appearence ~/Library/Application\ Support/Alfred/Alfred.alfredpreferences/preferences/$dir
done
localdir=$(/bin/ls -1 ~/Library/Application\ Support/Alfred/Alfred.alfredpreferences/preferences/local)
rm -rf ~/Library/Application\ Support/Alfred/Alfred.alfredpreferences/preferences/local/$localdir/*
dirs=(appearence features hotkey)
for dir in "${dirs[@]}"
do
ln -s $ZSH/alfred/local/appearence ~/Library/Application\ Support/Alfred/Alfred.alfredpreferences/preferences/local/$localdir/$dir
done | 2 | 0.1 | 2 | 0 |
9c89fe69ba805abc5a66f3d6ec8597ec24467c47 | configurations/vsphere/cloud-config.yml | configurations/vsphere/cloud-config.yml | ---
azs:
- name: z1
cloud_properties:
datacenters:
- clusters: [((vcenter_cluster)): ((vcenter_rp))]
networks:
- name: &cloud_network ((deployments_network))
type: manual
subnets:
- range: ((internal_cidr))
gateway: ((internal_gw))
dns: [((internal_ip))]
azs: [z1]
cloud_properties:
name: ((network_name))
reserved: ((reserved_ips))
vm_types:
- name: common
cloud_properties:
ram: 4096
cpu: 1
disk: 20480
- name: master
cloud_properties:
ram: 4096
cpu: 1
disk: 20480
- name: worker
cloud_properties:
ram: 8192
cpu: 1
disk: 102400
disk_types:
- name: 10240
disk_size: 10240
- name: 5120
disk_size: 5120
compilation:
workers: 4
network: *cloud_network
az: z1
reuse_compilation_vms: true
vm_type: worker
| ---
azs:
- name: z1
cloud_properties:
datacenters:
- clusters: [((vcenter_cluster)): { resource_pool: ((vcenter_rp))}]
networks:
- name: &cloud_network ((deployments_network))
type: manual
subnets:
- range: ((internal_cidr))
gateway: ((internal_gw))
dns: [((internal_ip))]
azs: [z1]
cloud_properties:
name: ((network_name))
reserved: ((reserved_ips))
vm_types:
- name: common
cloud_properties:
ram: 4096
cpu: 1
disk: 20480
- name: master
cloud_properties:
ram: 4096
cpu: 1
disk: 20480
- name: worker
cloud_properties:
ram: 8192
cpu: 1
disk: 102400
disk_types:
- name: 10240
disk_size: 10240
- name: 5120
disk_size: 5120
compilation:
workers: 4
network: *cloud_network
az: z1
reuse_compilation_vms: true
vm_type: worker
| Use new configuration for resource pool for vSphere clusters | Use new configuration for resource pool for vSphere clusters
| YAML | apache-2.0 | pivotal-cf-experimental/kubo-deployment,pivotal-cf-experimental/kubo-deployment,jaimegag/kubo-deployment,jaimegag/kubo-deployment | yaml | ## Code Before:
---
azs:
- name: z1
cloud_properties:
datacenters:
- clusters: [((vcenter_cluster)): ((vcenter_rp))]
networks:
- name: &cloud_network ((deployments_network))
type: manual
subnets:
- range: ((internal_cidr))
gateway: ((internal_gw))
dns: [((internal_ip))]
azs: [z1]
cloud_properties:
name: ((network_name))
reserved: ((reserved_ips))
vm_types:
- name: common
cloud_properties:
ram: 4096
cpu: 1
disk: 20480
- name: master
cloud_properties:
ram: 4096
cpu: 1
disk: 20480
- name: worker
cloud_properties:
ram: 8192
cpu: 1
disk: 102400
disk_types:
- name: 10240
disk_size: 10240
- name: 5120
disk_size: 5120
compilation:
workers: 4
network: *cloud_network
az: z1
reuse_compilation_vms: true
vm_type: worker
## Instruction:
Use new configuration for resource pool for vSphere clusters
## Code After:
---
azs:
- name: z1
cloud_properties:
datacenters:
- clusters: [((vcenter_cluster)): { resource_pool: ((vcenter_rp))}]
networks:
- name: &cloud_network ((deployments_network))
type: manual
subnets:
- range: ((internal_cidr))
gateway: ((internal_gw))
dns: [((internal_ip))]
azs: [z1]
cloud_properties:
name: ((network_name))
reserved: ((reserved_ips))
vm_types:
- name: common
cloud_properties:
ram: 4096
cpu: 1
disk: 20480
- name: master
cloud_properties:
ram: 4096
cpu: 1
disk: 20480
- name: worker
cloud_properties:
ram: 8192
cpu: 1
disk: 102400
disk_types:
- name: 10240
disk_size: 10240
- name: 5120
disk_size: 5120
compilation:
workers: 4
network: *cloud_network
az: z1
reuse_compilation_vms: true
vm_type: worker
| ---
azs:
- name: z1
cloud_properties:
datacenters:
- - clusters: [((vcenter_cluster)): ((vcenter_rp))]
+ - clusters: [((vcenter_cluster)): { resource_pool: ((vcenter_rp))}]
? +++++++++++++++++ +
networks:
- name: &cloud_network ((deployments_network))
type: manual
subnets:
- range: ((internal_cidr))
gateway: ((internal_gw))
dns: [((internal_ip))]
azs: [z1]
cloud_properties:
name: ((network_name))
reserved: ((reserved_ips))
vm_types:
- name: common
cloud_properties:
ram: 4096
cpu: 1
disk: 20480
- name: master
cloud_properties:
ram: 4096
cpu: 1
disk: 20480
- name: worker
cloud_properties:
ram: 8192
cpu: 1
disk: 102400
disk_types:
- name: 10240
disk_size: 10240
- name: 5120
disk_size: 5120
compilation:
workers: 4
network: *cloud_network
az: z1
reuse_compilation_vms: true
vm_type: worker | 2 | 0.04 | 1 | 1 |
ecff4d938b84614526075600d7d3df35555970b5 | AwesomeButtonExample/AwesomeButtonExample/ViewController.swift | AwesomeButtonExample/AwesomeButtonExample/ViewController.swift | //
// ViewController.swift
// AwesomeButtonExample
//
// Created by Панов Андрей on 26.12.15.
// Copyright © 2015 Панов Андрей. All rights reserved.
//
import UIKit
import AwesomeButton
class ViewController: UIViewController {
@IBOutlet weak var buttonNew: AwesomeButton!
override func viewDidLoad() {
super.viewDidLoad()
// image only
let button = AwesomeButton(type: .Custom)
button.frame = CGRectMake(10, 200, 200, 40)
button.buttonWithIcon(UIImage(named: "example-arrow")!, highlightedImage: UIImage(named: "arrowNextDark"), title: "loadadl")
self.view.addSubview(button)
}
override func didReceiveMemoryWarning() {
super.didReceiveMemoryWarning()
// Dispose of any resources that can be recreated.
}
}
| //
// ViewController.swift
// AwesomeButtonExample
//
// Created by Панов Андрей on 26.12.15.
// Copyright © 2015 Панов Андрей. All rights reserved.
//
import UIKit
import AwesomeButton
class ViewController: UIViewController {
@IBOutlet weak var buttonNew: AwesomeButton!
override func viewDidLoad() {
super.viewDidLoad()
// image only
let button = AwesomeButton(type: .custom)
button.frame = CGRect(x: 10, y: 200, width: 200, height: 40)
button.buttonWithIcon(UIImage(named: "example-arrow")!, highlightedImage: UIImage(named: "arrowNextDark"), title: "loadadl")
self.view.addSubview(button)
}
override func didReceiveMemoryWarning() {
super.didReceiveMemoryWarning()
// Dispose of any resources that can be recreated.
}
}
| Fix example to function as expected | Fix example to function as expected
| Swift | mit | AndreyPanov/AwesomeButton,AndreyPanov/AwesomeButton,AndreyPanov/AwesomeButton | swift | ## Code Before:
//
// ViewController.swift
// AwesomeButtonExample
//
// Created by Панов Андрей on 26.12.15.
// Copyright © 2015 Панов Андрей. All rights reserved.
//
import UIKit
import AwesomeButton
class ViewController: UIViewController {
@IBOutlet weak var buttonNew: AwesomeButton!
override func viewDidLoad() {
super.viewDidLoad()
// image only
let button = AwesomeButton(type: .Custom)
button.frame = CGRectMake(10, 200, 200, 40)
button.buttonWithIcon(UIImage(named: "example-arrow")!, highlightedImage: UIImage(named: "arrowNextDark"), title: "loadadl")
self.view.addSubview(button)
}
override func didReceiveMemoryWarning() {
super.didReceiveMemoryWarning()
// Dispose of any resources that can be recreated.
}
}
## Instruction:
Fix example to function as expected
## Code After:
//
// ViewController.swift
// AwesomeButtonExample
//
// Created by Панов Андрей on 26.12.15.
// Copyright © 2015 Панов Андрей. All rights reserved.
//
import UIKit
import AwesomeButton
class ViewController: UIViewController {
@IBOutlet weak var buttonNew: AwesomeButton!
override func viewDidLoad() {
super.viewDidLoad()
// image only
let button = AwesomeButton(type: .custom)
button.frame = CGRect(x: 10, y: 200, width: 200, height: 40)
button.buttonWithIcon(UIImage(named: "example-arrow")!, highlightedImage: UIImage(named: "arrowNextDark"), title: "loadadl")
self.view.addSubview(button)
}
override func didReceiveMemoryWarning() {
super.didReceiveMemoryWarning()
// Dispose of any resources that can be recreated.
}
}
| //
// ViewController.swift
// AwesomeButtonExample
//
// Created by Панов Андрей on 26.12.15.
// Copyright © 2015 Панов Андрей. All rights reserved.
//
import UIKit
import AwesomeButton
class ViewController: UIViewController {
@IBOutlet weak var buttonNew: AwesomeButton!
override func viewDidLoad() {
super.viewDidLoad()
// image only
- let button = AwesomeButton(type: .Custom)
? ^
+ let button = AwesomeButton(type: .custom)
? ^
- button.frame = CGRectMake(10, 200, 200, 40)
? ----
+ button.frame = CGRect(x: 10, y: 200, width: 200, height: 40)
? +++ +++ +++++++ ++++++++
button.buttonWithIcon(UIImage(named: "example-arrow")!, highlightedImage: UIImage(named: "arrowNextDark"), title: "loadadl")
self.view.addSubview(button)
}
override func didReceiveMemoryWarning() {
super.didReceiveMemoryWarning()
// Dispose of any resources that can be recreated.
}
}
| 4 | 0.117647 | 2 | 2 |
8e35c9b1cc2db303a2b31394bf8a470dd2e6edae | src/AuraIsHere/LaravelMultiTenant/LaravelMultiTenantServiceProvider.php | src/AuraIsHere/LaravelMultiTenant/LaravelMultiTenantServiceProvider.php | <?php namespace AuraIsHere\LaravelMultiTenant;
use Illuminate\Foundation\AliasLoader;
use Illuminate\Support\ServiceProvider;
class LaravelMultiTenantServiceProvider extends ServiceProvider
{
/**
* Indicates if loading of the provider is deferred.
*
* @var bool
*/
protected $defer = false;
/**
* Bootstrap the application events.
*
* @return void
*/
public function boot()
{
$this->package('aura-is-here/laravel-multi-tenant');
}
/**
* Register the service provider.
*
* @return void
*/
public function register()
{
// Register our tenant scope instance
$this->app->bindshared('AuraIsHere\LaravelMultiTenant\TenantScope', function ($app) {
return new TenantScope();
});
// Define alias 'TenantScope'
$this->app->booting(function () {
$loader = AliasLoader::getInstance();
$loader->alias('TenantScope', 'AuraIsHere\LaravelMultiTenant\Facades\TenantScopeFacade');
});
// Register our config
$this->app['config']->package('aura-is-here/laravel-multi-tenant', __DIR__.'/../../config');
}
/**
* Get the services provided by the provider.
*
* @return array
*/
public function provides()
{
return [];
}
}
| <?php namespace AuraIsHere\LaravelMultiTenant;
use Illuminate\Foundation\AliasLoader;
use Illuminate\Support\ServiceProvider;
class LaravelMultiTenantServiceProvider extends ServiceProvider
{
/**
* Indicates if loading of the provider is deferred.
*
* @var bool
*/
protected $defer = false;
/**
* Bootstrap the application events.
*
* @return void
*/
public function boot()
{
$this->package('aura-is-here/laravel-multi-tenant');
}
/**
* Register the service provider.
*
* @return void
*/
public function register()
{
// Register our tenant scope instance
$this->app->singleton('AuraIsHere\LaravelMultiTenant\TenantScope', function ($app) {
return new TenantScope();
});
// Define alias 'TenantScope'
$this->app->booting(function () {
$loader = AliasLoader::getInstance();
$loader->alias('TenantScope', 'AuraIsHere\LaravelMultiTenant\Facades\TenantScopeFacade');
});
// Register our config
$this->app['config']->package('aura-is-here/laravel-multi-tenant', __DIR__.'/../../config');
}
/**
* Get the services provided by the provider.
*
* @return array
*/
public function provides()
{
return [];
}
}
| Make it a singleton instead | Make it a singleton instead
| PHP | mit | Feijs/laravel-multi-tenant,JamesForks/laravel-multi-tenant,HipsterJazzbo/Landlord,alexw23/laravel-multi-tenant,HipsterJazzbo/laravel-multi-tenant,BlueBayTravel/laravel-multi-tenant,mpociot/laravel-multi-tenant,AuraEQ/laravel-multi-tenant,rikvdlooi/Landlord | php | ## Code Before:
<?php namespace AuraIsHere\LaravelMultiTenant;
use Illuminate\Foundation\AliasLoader;
use Illuminate\Support\ServiceProvider;
class LaravelMultiTenantServiceProvider extends ServiceProvider
{
/**
* Indicates if loading of the provider is deferred.
*
* @var bool
*/
protected $defer = false;
/**
* Bootstrap the application events.
*
* @return void
*/
public function boot()
{
$this->package('aura-is-here/laravel-multi-tenant');
}
/**
* Register the service provider.
*
* @return void
*/
public function register()
{
// Register our tenant scope instance
$this->app->bindshared('AuraIsHere\LaravelMultiTenant\TenantScope', function ($app) {
return new TenantScope();
});
// Define alias 'TenantScope'
$this->app->booting(function () {
$loader = AliasLoader::getInstance();
$loader->alias('TenantScope', 'AuraIsHere\LaravelMultiTenant\Facades\TenantScopeFacade');
});
// Register our config
$this->app['config']->package('aura-is-here/laravel-multi-tenant', __DIR__.'/../../config');
}
/**
* Get the services provided by the provider.
*
* @return array
*/
public function provides()
{
return [];
}
}
## Instruction:
Make it a singleton instead
## Code After:
<?php namespace AuraIsHere\LaravelMultiTenant;
use Illuminate\Foundation\AliasLoader;
use Illuminate\Support\ServiceProvider;
class LaravelMultiTenantServiceProvider extends ServiceProvider
{
/**
* Indicates if loading of the provider is deferred.
*
* @var bool
*/
protected $defer = false;
/**
* Bootstrap the application events.
*
* @return void
*/
public function boot()
{
$this->package('aura-is-here/laravel-multi-tenant');
}
/**
* Register the service provider.
*
* @return void
*/
public function register()
{
// Register our tenant scope instance
$this->app->singleton('AuraIsHere\LaravelMultiTenant\TenantScope', function ($app) {
return new TenantScope();
});
// Define alias 'TenantScope'
$this->app->booting(function () {
$loader = AliasLoader::getInstance();
$loader->alias('TenantScope', 'AuraIsHere\LaravelMultiTenant\Facades\TenantScopeFacade');
});
// Register our config
$this->app['config']->package('aura-is-here/laravel-multi-tenant', __DIR__.'/../../config');
}
/**
* Get the services provided by the provider.
*
* @return array
*/
public function provides()
{
return [];
}
}
| <?php namespace AuraIsHere\LaravelMultiTenant;
use Illuminate\Foundation\AliasLoader;
use Illuminate\Support\ServiceProvider;
class LaravelMultiTenantServiceProvider extends ServiceProvider
{
/**
* Indicates if loading of the provider is deferred.
*
* @var bool
*/
protected $defer = false;
/**
* Bootstrap the application events.
*
* @return void
*/
public function boot()
{
$this->package('aura-is-here/laravel-multi-tenant');
}
- /**
+ /**
- * Register the service provider.
? ^^^^
+ * Register the service provider.
? ^
- *
+ *
- * @return void
? ^^^^
+ * @return void
? ^
- */
+ */
- public function register()
? ^^^^
+ public function register()
? ^
- {
+ {
- // Register our tenant scope instance
? ^^^^^^^^
+ // Register our tenant scope instance
? ^^
- $this->app->bindshared('AuraIsHere\LaravelMultiTenant\TenantScope', function ($app) {
? ^^^^^^^^ ^ ^^^^^ ^
+ $this->app->singleton('AuraIsHere\LaravelMultiTenant\TenantScope', function ($app) {
? ^^ ^ ^^ ^^^
return new TenantScope();
});
// Define alias 'TenantScope'
$this->app->booting(function () {
$loader = AliasLoader::getInstance();
$loader->alias('TenantScope', 'AuraIsHere\LaravelMultiTenant\Facades\TenantScopeFacade');
});
// Register our config
$this->app['config']->package('aura-is-here/laravel-multi-tenant', __DIR__.'/../../config');
}
/**
* Get the services provided by the provider.
*
* @return array
*/
public function provides()
{
return [];
}
} | 18 | 0.321429 | 9 | 9 |
165221441313bcce4769eda91a1b0b1fd62188bd | _people/daniel_weyman.md | _people/daniel_weyman.md | ---
title: Daniel Weyman
gender: male
headshot: X6dqbwJ
submitted: false
careers:
- Actor
---
Daniel has a prolific career both on the stage and in front of the camera. Daniels theatre credits include performances at the Sheffield Crucible and the Open Air Theatre in Regents Park. His Television and Film credits include Foyle's War for ITV and the film version of Great Expectations. | ---
title: Daniel Weyman
gender: male
headshot: X6dqbwJ
graduated: 2000
submitted: false
careers:
- Actor
---
Daniel has a prolific career both on the stage and in front of the camera. Daniels theatre credits include performances at the Sheffield Crucible and the Open Air Theatre in Regents Park. His Television and Film credits include Foyle's War for ITV and the film version of Great Expectations.
| Add Daniel Weyman graduation year | Add Daniel Weyman graduation year
Source: Uni Alum Relations FB Post
https://www.facebook.com/universityofnottinghamalumni/photos/a.186558471388162.42076.153229431387733/894191290624873/?type=1&theater
| Markdown | mit | newtheatre/history-project,newtheatre/history-project,johnathan99j/history-project,johnathan99j/history-project,johnathan99j/history-project,johnathan99j/history-project,johnathan99j/history-project,newtheatre/history-project,newtheatre/history-project,newtheatre/history-project | markdown | ## Code Before:
---
title: Daniel Weyman
gender: male
headshot: X6dqbwJ
submitted: false
careers:
- Actor
---
Daniel has a prolific career both on the stage and in front of the camera. Daniels theatre credits include performances at the Sheffield Crucible and the Open Air Theatre in Regents Park. His Television and Film credits include Foyle's War for ITV and the film version of Great Expectations.
## Instruction:
Add Daniel Weyman graduation year
Source: Uni Alum Relations FB Post
https://www.facebook.com/universityofnottinghamalumni/photos/a.186558471388162.42076.153229431387733/894191290624873/?type=1&theater
## Code After:
---
title: Daniel Weyman
gender: male
headshot: X6dqbwJ
graduated: 2000
submitted: false
careers:
- Actor
---
Daniel has a prolific career both on the stage and in front of the camera. Daniels theatre credits include performances at the Sheffield Crucible and the Open Air Theatre in Regents Park. His Television and Film credits include Foyle's War for ITV and the film version of Great Expectations.
| ---
title: Daniel Weyman
gender: male
headshot: X6dqbwJ
+ graduated: 2000
submitted: false
careers:
- Actor
---
Daniel has a prolific career both on the stage and in front of the camera. Daniels theatre credits include performances at the Sheffield Crucible and the Open Air Theatre in Regents Park. His Television and Film credits include Foyle's War for ITV and the film version of Great Expectations. | 1 | 0.1 | 1 | 0 |
025c3f6b73c97fdb58b1a492efcb6efe44cfdab0 | twisted/plugins/caldav.py | twisted/plugins/caldav.py | from zope.interface import implements
from twisted.plugin import IPlugin
from twisted.application.service import IServiceMaker
from twisted.python import reflect
def serviceMakerProperty(propname):
def getProperty(self):
return getattr(reflect.namedClass(self.serviceMakerClass), propname)
return property(getProperty)
class TAP(object):
implements(IPlugin, IServiceMaker)
def __init__(self, serviceMakerClass):
self.serviceMakerClass = serviceMakerClass
self._serviceMaker = None
options = serviceMakerProperty("options")
tapname = serviceMakerProperty("tapname")
description = serviceMakerProperty("description")
def makeService(self, options):
if self._serviceMaker is None:
self._serviceMaker = reflect.namedClass(self.serviceMakerClass)()
return self._serviceMaker.makeService(options)
TwistedCalDAV = TAP("calendarserver.tap.caldav.CalDAVServiceMaker")
CalDAVNotifier = TAP("twistedcaldav.notify.NotificationServiceMaker")
CalDAVMailGateway = TAP("twistedcaldav.mail.MailGatewayServiceMaker")
| from zope.interface import implements
from twisted.plugin import IPlugin
from twisted.application.service import IServiceMaker
from twisted.python import reflect
from twisted.internet.protocol import Factory
Factory.noisy = False
def serviceMakerProperty(propname):
def getProperty(self):
return getattr(reflect.namedClass(self.serviceMakerClass), propname)
return property(getProperty)
class TAP(object):
implements(IPlugin, IServiceMaker)
def __init__(self, serviceMakerClass):
self.serviceMakerClass = serviceMakerClass
self._serviceMaker = None
options = serviceMakerProperty("options")
tapname = serviceMakerProperty("tapname")
description = serviceMakerProperty("description")
def makeService(self, options):
if self._serviceMaker is None:
self._serviceMaker = reflect.namedClass(self.serviceMakerClass)()
return self._serviceMaker.makeService(options)
TwistedCalDAV = TAP("calendarserver.tap.caldav.CalDAVServiceMaker")
CalDAVNotifier = TAP("twistedcaldav.notify.NotificationServiceMaker")
CalDAVMailGateway = TAP("twistedcaldav.mail.MailGatewayServiceMaker")
| Set Factory.noisy to False by default | Set Factory.noisy to False by default
git-svn-id: 81e381228600e5752b80483efd2b45b26c451ea2@3933 e27351fd-9f3e-4f54-a53b-843176b1656c
| Python | apache-2.0 | trevor/calendarserver,trevor/calendarserver,trevor/calendarserver | python | ## Code Before:
from zope.interface import implements
from twisted.plugin import IPlugin
from twisted.application.service import IServiceMaker
from twisted.python import reflect
def serviceMakerProperty(propname):
def getProperty(self):
return getattr(reflect.namedClass(self.serviceMakerClass), propname)
return property(getProperty)
class TAP(object):
implements(IPlugin, IServiceMaker)
def __init__(self, serviceMakerClass):
self.serviceMakerClass = serviceMakerClass
self._serviceMaker = None
options = serviceMakerProperty("options")
tapname = serviceMakerProperty("tapname")
description = serviceMakerProperty("description")
def makeService(self, options):
if self._serviceMaker is None:
self._serviceMaker = reflect.namedClass(self.serviceMakerClass)()
return self._serviceMaker.makeService(options)
TwistedCalDAV = TAP("calendarserver.tap.caldav.CalDAVServiceMaker")
CalDAVNotifier = TAP("twistedcaldav.notify.NotificationServiceMaker")
CalDAVMailGateway = TAP("twistedcaldav.mail.MailGatewayServiceMaker")
## Instruction:
Set Factory.noisy to False by default
git-svn-id: 81e381228600e5752b80483efd2b45b26c451ea2@3933 e27351fd-9f3e-4f54-a53b-843176b1656c
## Code After:
from zope.interface import implements
from twisted.plugin import IPlugin
from twisted.application.service import IServiceMaker
from twisted.python import reflect
from twisted.internet.protocol import Factory
Factory.noisy = False
def serviceMakerProperty(propname):
def getProperty(self):
return getattr(reflect.namedClass(self.serviceMakerClass), propname)
return property(getProperty)
class TAP(object):
implements(IPlugin, IServiceMaker)
def __init__(self, serviceMakerClass):
self.serviceMakerClass = serviceMakerClass
self._serviceMaker = None
options = serviceMakerProperty("options")
tapname = serviceMakerProperty("tapname")
description = serviceMakerProperty("description")
def makeService(self, options):
if self._serviceMaker is None:
self._serviceMaker = reflect.namedClass(self.serviceMakerClass)()
return self._serviceMaker.makeService(options)
TwistedCalDAV = TAP("calendarserver.tap.caldav.CalDAVServiceMaker")
CalDAVNotifier = TAP("twistedcaldav.notify.NotificationServiceMaker")
CalDAVMailGateway = TAP("twistedcaldav.mail.MailGatewayServiceMaker")
| from zope.interface import implements
from twisted.plugin import IPlugin
from twisted.application.service import IServiceMaker
from twisted.python import reflect
+
+ from twisted.internet.protocol import Factory
+ Factory.noisy = False
+
def serviceMakerProperty(propname):
def getProperty(self):
return getattr(reflect.namedClass(self.serviceMakerClass), propname)
return property(getProperty)
class TAP(object):
implements(IPlugin, IServiceMaker)
+
def __init__(self, serviceMakerClass):
self.serviceMakerClass = serviceMakerClass
self._serviceMaker = None
- options = serviceMakerProperty("options")
+ options = serviceMakerProperty("options")
? ++++
- tapname = serviceMakerProperty("tapname")
+ tapname = serviceMakerProperty("tapname")
? ++++
description = serviceMakerProperty("description")
def makeService(self, options):
if self._serviceMaker is None:
self._serviceMaker = reflect.namedClass(self.serviceMakerClass)()
return self._serviceMaker.makeService(options)
TwistedCalDAV = TAP("calendarserver.tap.caldav.CalDAVServiceMaker")
CalDAVNotifier = TAP("twistedcaldav.notify.NotificationServiceMaker")
CalDAVMailGateway = TAP("twistedcaldav.mail.MailGatewayServiceMaker") | 9 | 0.272727 | 7 | 2 |
cf0033ee371925218237efdb613bcd312c0faefa | forms.py | forms.py | from flask_wtf import FlaskForm
from wtforms import StringField, TextAreaField, PasswordField, HiddenField
from wtforms.validators import DataRequired, Length, URL, Optional
class LoginForm(FlaskForm):
username = StringField('Username', [Length(max=255)], render_kw={'inputmode': 'verbatim'})
password = PasswordField('Password', [Length(max=1024)])
class SignupForm(FlaskForm):
username = StringField('Username', [Length(min=3, max=255)], render_kw={'inputmode': 'verbatim'})
password = PasswordField('Password', [Length(min=8, max=1024)])
class AddForm(FlaskForm):
title = StringField('Title', [Length(max=255)])
url = StringField('URL', [Optional(), URL(), Length(max=255)])
username = StringField(
'Username',
[Length(max=255)],
render_kw={'inputmode': 'verbatim'},
)
other = TextAreaField('Other')
class EditForm(AddForm):
id = HiddenField()
| from flask_wtf import FlaskForm
from wtforms import StringField, TextAreaField, PasswordField, HiddenField
from wtforms.validators import DataRequired, Length, URL, Optional
class LoginForm(FlaskForm):
username = StringField('Username', [Length(max=255)], render_kw={'inputmode': 'verbatim'})
password = PasswordField('Password', [Length(max=1024)])
class SignupForm(FlaskForm):
username = StringField(
'Username',
[Length(min=3, max=255)],
render_kw={'inputmode': 'verbatim'},
id='newusername',
)
password = PasswordField('Password', [Length(min=8, max=1024)])
class AddForm(FlaskForm):
title = StringField('Title', [Length(max=255)])
url = StringField('URL', [Optional(), URL(), Length(max=255)])
username = StringField(
'Username',
[Length(max=255)],
render_kw={'inputmode': 'verbatim'},
)
other = TextAreaField('Other')
class EditForm(AddForm):
id = HiddenField()
| Add 'newusername' id to username on signup page | Add 'newusername' id to username on signup page
Needed for js frontend.
| Python | mit | tortxof/flask-password,tortxof/flask-password,tortxof/flask-password | python | ## Code Before:
from flask_wtf import FlaskForm
from wtforms import StringField, TextAreaField, PasswordField, HiddenField
from wtforms.validators import DataRequired, Length, URL, Optional
class LoginForm(FlaskForm):
username = StringField('Username', [Length(max=255)], render_kw={'inputmode': 'verbatim'})
password = PasswordField('Password', [Length(max=1024)])
class SignupForm(FlaskForm):
username = StringField('Username', [Length(min=3, max=255)], render_kw={'inputmode': 'verbatim'})
password = PasswordField('Password', [Length(min=8, max=1024)])
class AddForm(FlaskForm):
title = StringField('Title', [Length(max=255)])
url = StringField('URL', [Optional(), URL(), Length(max=255)])
username = StringField(
'Username',
[Length(max=255)],
render_kw={'inputmode': 'verbatim'},
)
other = TextAreaField('Other')
class EditForm(AddForm):
id = HiddenField()
## Instruction:
Add 'newusername' id to username on signup page
Needed for js frontend.
## Code After:
from flask_wtf import FlaskForm
from wtforms import StringField, TextAreaField, PasswordField, HiddenField
from wtforms.validators import DataRequired, Length, URL, Optional
class LoginForm(FlaskForm):
username = StringField('Username', [Length(max=255)], render_kw={'inputmode': 'verbatim'})
password = PasswordField('Password', [Length(max=1024)])
class SignupForm(FlaskForm):
username = StringField(
'Username',
[Length(min=3, max=255)],
render_kw={'inputmode': 'verbatim'},
id='newusername',
)
password = PasswordField('Password', [Length(min=8, max=1024)])
class AddForm(FlaskForm):
title = StringField('Title', [Length(max=255)])
url = StringField('URL', [Optional(), URL(), Length(max=255)])
username = StringField(
'Username',
[Length(max=255)],
render_kw={'inputmode': 'verbatim'},
)
other = TextAreaField('Other')
class EditForm(AddForm):
id = HiddenField()
| from flask_wtf import FlaskForm
from wtforms import StringField, TextAreaField, PasswordField, HiddenField
from wtforms.validators import DataRequired, Length, URL, Optional
class LoginForm(FlaskForm):
username = StringField('Username', [Length(max=255)], render_kw={'inputmode': 'verbatim'})
password = PasswordField('Password', [Length(max=1024)])
class SignupForm(FlaskForm):
- username = StringField('Username', [Length(min=3, max=255)], render_kw={'inputmode': 'verbatim'})
+ username = StringField(
+ 'Username',
+ [Length(min=3, max=255)],
+ render_kw={'inputmode': 'verbatim'},
+ id='newusername',
+ )
password = PasswordField('Password', [Length(min=8, max=1024)])
class AddForm(FlaskForm):
title = StringField('Title', [Length(max=255)])
url = StringField('URL', [Optional(), URL(), Length(max=255)])
username = StringField(
'Username',
[Length(max=255)],
render_kw={'inputmode': 'verbatim'},
)
other = TextAreaField('Other')
class EditForm(AddForm):
id = HiddenField() | 7 | 0.291667 | 6 | 1 |
62c24f6edaa91834d4a7b2a3f9b99b8b96322230 | nova/policies/hide_server_addresses.py | nova/policies/hide_server_addresses.py |
from oslo_policy import policy
BASE_POLICY_NAME = 'os_compute_api:os-hide-server-addresses'
hide_server_addresses_policies = [
policy.RuleDefault(
name=BASE_POLICY_NAME,
check_str='is_admin:False'),
]
def list_rules():
return hide_server_addresses_policies
|
from nova.policies import base
BASE_POLICY_NAME = 'os_compute_api:os-hide-server-addresses'
hide_server_addresses_policies = [
base.create_rule_default(
BASE_POLICY_NAME,
'is_admin:False',
"""Hide server's 'addresses' key in the server response.
This set the 'addresses' key in the server response to an empty dictionary
when the server is in a specific set of states as defined in
CONF.api.hide_server_address_states.
By default 'addresses' is hidden only when the server is in 'BUILDING'
state.""",
[
{
'method': 'GET',
'path': '/servers/{id}'
},
{
'method': 'GET',
'path': '/servers/detail'
}
]),
]
def list_rules():
return hide_server_addresses_policies
| Add policy description for 'os-hide-server-addresses' | Add policy description for 'os-hide-server-addresses'
This commit adds policy doc for 'os-hide-server-addresses' policies.
Partial implement blueprint policy-docs
Change-Id: I98edbd8579f052c74283bde2ec4f85d301a0807a
| Python | apache-2.0 | rahulunair/nova,gooddata/openstack-nova,mikalstill/nova,mahak/nova,Juniper/nova,mahak/nova,mikalstill/nova,Juniper/nova,rahulunair/nova,gooddata/openstack-nova,vmturbo/nova,openstack/nova,Juniper/nova,jianghuaw/nova,openstack/nova,gooddata/openstack-nova,klmitch/nova,gooddata/openstack-nova,jianghuaw/nova,klmitch/nova,klmitch/nova,mahak/nova,vmturbo/nova,klmitch/nova,jianghuaw/nova,phenoxim/nova,mikalstill/nova,vmturbo/nova,openstack/nova,vmturbo/nova,jianghuaw/nova,rahulunair/nova,phenoxim/nova,Juniper/nova | python | ## Code Before:
from oslo_policy import policy
BASE_POLICY_NAME = 'os_compute_api:os-hide-server-addresses'
hide_server_addresses_policies = [
policy.RuleDefault(
name=BASE_POLICY_NAME,
check_str='is_admin:False'),
]
def list_rules():
return hide_server_addresses_policies
## Instruction:
Add policy description for 'os-hide-server-addresses'
This commit adds policy doc for 'os-hide-server-addresses' policies.
Partial implement blueprint policy-docs
Change-Id: I98edbd8579f052c74283bde2ec4f85d301a0807a
## Code After:
from nova.policies import base
BASE_POLICY_NAME = 'os_compute_api:os-hide-server-addresses'
hide_server_addresses_policies = [
base.create_rule_default(
BASE_POLICY_NAME,
'is_admin:False',
"""Hide server's 'addresses' key in the server response.
This set the 'addresses' key in the server response to an empty dictionary
when the server is in a specific set of states as defined in
CONF.api.hide_server_address_states.
By default 'addresses' is hidden only when the server is in 'BUILDING'
state.""",
[
{
'method': 'GET',
'path': '/servers/{id}'
},
{
'method': 'GET',
'path': '/servers/detail'
}
]),
]
def list_rules():
return hide_server_addresses_policies
|
+ from nova.policies import base
- from oslo_policy import policy
-
BASE_POLICY_NAME = 'os_compute_api:os-hide-server-addresses'
hide_server_addresses_policies = [
- policy.RuleDefault(
+ base.create_rule_default(
- name=BASE_POLICY_NAME,
? -----
+ BASE_POLICY_NAME,
- check_str='is_admin:False'),
? ---------- -
+ 'is_admin:False',
+ """Hide server's 'addresses' key in the server response.
+
+ This set the 'addresses' key in the server response to an empty dictionary
+ when the server is in a specific set of states as defined in
+ CONF.api.hide_server_address_states.
+ By default 'addresses' is hidden only when the server is in 'BUILDING'
+ state.""",
+ [
+ {
+ 'method': 'GET',
+ 'path': '/servers/{id}'
+ },
+ {
+ 'method': 'GET',
+ 'path': '/servers/detail'
+ }
+ ]),
]
def list_rules():
return hide_server_addresses_policies | 26 | 1.625 | 21 | 5 |
04fbda52733afb41536d9e9127c0dcb0e3287b03 | Cargo.toml | Cargo.toml | [package]
name = "grafen"
version = "0.3.1"
authors = ["Petter Johansson <pettjoha@kth.se>"]
description = "Create graphene and other substrates for use in molecular dynamics simulations."
repository = "https://github.com/pjohansson/grafen"
travis-ci = { repository = "https://travis-ci.org/pjohansson/grafen", branch = "master"}
categories = ["command-line-utilities", "science"]
keywords = ["gromacs", "molecular-dynamics", "substrates"]
readme = "README.md"
license = "Unlicense"
[dependencies]
ansi_term = "^0.9.0"
clap = "~2.23.0"
| [package]
name = "grafen"
version = "0.3.3"
authors = ["Petter Johansson <pettjoha@kth.se>"]
description = "Create graphene and other substrates for use in molecular dynamics simulations."
repository = "https://github.com/pjohansson/grafen"
categories = ["command-line-utilities", "science"]
keywords = ["gromacs", "molecular-dynamics", "substrates"]
readme = "README.md"
license = "Unlicense"
[badges]
travis-ci = { repository = "pjohansson/grafen", branch = "master"}
[dependencies]
ansi_term = "^0.9.0"
clap = "~2.23.0"
| Correct [badges] section in manifest | Correct [badges] section in manifest
| TOML | unlicense | pjohansson/grafen | toml | ## Code Before:
[package]
name = "grafen"
version = "0.3.1"
authors = ["Petter Johansson <pettjoha@kth.se>"]
description = "Create graphene and other substrates for use in molecular dynamics simulations."
repository = "https://github.com/pjohansson/grafen"
travis-ci = { repository = "https://travis-ci.org/pjohansson/grafen", branch = "master"}
categories = ["command-line-utilities", "science"]
keywords = ["gromacs", "molecular-dynamics", "substrates"]
readme = "README.md"
license = "Unlicense"
[dependencies]
ansi_term = "^0.9.0"
clap = "~2.23.0"
## Instruction:
Correct [badges] section in manifest
## Code After:
[package]
name = "grafen"
version = "0.3.3"
authors = ["Petter Johansson <pettjoha@kth.se>"]
description = "Create graphene and other substrates for use in molecular dynamics simulations."
repository = "https://github.com/pjohansson/grafen"
categories = ["command-line-utilities", "science"]
keywords = ["gromacs", "molecular-dynamics", "substrates"]
readme = "README.md"
license = "Unlicense"
[badges]
travis-ci = { repository = "pjohansson/grafen", branch = "master"}
[dependencies]
ansi_term = "^0.9.0"
clap = "~2.23.0"
| [package]
name = "grafen"
- version = "0.3.1"
? ^
+ version = "0.3.3"
? ^
authors = ["Petter Johansson <pettjoha@kth.se>"]
description = "Create graphene and other substrates for use in molecular dynamics simulations."
repository = "https://github.com/pjohansson/grafen"
- travis-ci = { repository = "https://travis-ci.org/pjohansson/grafen", branch = "master"}
categories = ["command-line-utilities", "science"]
keywords = ["gromacs", "molecular-dynamics", "substrates"]
readme = "README.md"
license = "Unlicense"
+ [badges]
+ travis-ci = { repository = "pjohansson/grafen", branch = "master"}
+
[dependencies]
ansi_term = "^0.9.0"
clap = "~2.23.0" | 6 | 0.4 | 4 | 2 |
4d640b5e1027dd03905140862aae86543829b9cd | package.json | package.json | {
"scripts": {
"initSubModules": "sh ./scripts/initialiseSubModules.sh",
"initLatest": "npm run initSubModules && sh ./scripts/switchToBranch.sh latest",
"initMaster": "npm run initSubModules && sh ./scripts/switchToBranch.sh master",
"updateDependencies": "lerna bootstrap",
"build": "lerna run build",
"buildCore": "lerna run build --scope ag-grid --scope ag-grid-enterprise --scope ag-grid-angular --scope ag-grid-react --scope ag-grid-vue --scope ag-grid-aurelia",
"updateAndRebuild": "npm run updateDependencies && npm run build",
"initIntelliJ": "cp .idea/workspace.xml.default .idea/workspace.xml",
"init": "npm run initLatest && npm run updateAndRebuild"
},
"devDependencies": {
"lerna": "^2.11.0"
}
}
| {
"scripts": {
"initSubModules": "sh ./scripts/initialiseSubModules.sh",
"initLatest": "npm run initSubModules && sh ./scripts/switchToBranch.sh latest",
"initMaster": "npm run initSubModules && sh ./scripts/switchToBranch.sh master",
"updateExamples": "sh ./scripts/updateSubModules.sh",
"updateDependencies": "lerna bootstrap",
"build": "lerna run build",
"buildCore": "lerna run build --scope ag-grid --scope ag-grid-enterprise --scope ag-grid-angular --scope ag-grid-react --scope ag-grid-vue --scope ag-grid-aurelia",
"updateAndRebuild": "npm run updateDependencies && npm run build",
"initIntelliJ": "cp .idea/workspace.xml.default .idea/workspace.xml",
"init": "npm run initLatest && npm run updateAndRebuild"
},
"devDependencies": {
"lerna": "^2.11.0"
}
}
| Add utility to update examples/submodules | Add utility to update examples/submodules
| JSON | mit | ceolter/ag-grid,ceolter/ag-grid,ceolter/angular-grid,ceolter/angular-grid | json | ## Code Before:
{
"scripts": {
"initSubModules": "sh ./scripts/initialiseSubModules.sh",
"initLatest": "npm run initSubModules && sh ./scripts/switchToBranch.sh latest",
"initMaster": "npm run initSubModules && sh ./scripts/switchToBranch.sh master",
"updateDependencies": "lerna bootstrap",
"build": "lerna run build",
"buildCore": "lerna run build --scope ag-grid --scope ag-grid-enterprise --scope ag-grid-angular --scope ag-grid-react --scope ag-grid-vue --scope ag-grid-aurelia",
"updateAndRebuild": "npm run updateDependencies && npm run build",
"initIntelliJ": "cp .idea/workspace.xml.default .idea/workspace.xml",
"init": "npm run initLatest && npm run updateAndRebuild"
},
"devDependencies": {
"lerna": "^2.11.0"
}
}
## Instruction:
Add utility to update examples/submodules
## Code After:
{
"scripts": {
"initSubModules": "sh ./scripts/initialiseSubModules.sh",
"initLatest": "npm run initSubModules && sh ./scripts/switchToBranch.sh latest",
"initMaster": "npm run initSubModules && sh ./scripts/switchToBranch.sh master",
"updateExamples": "sh ./scripts/updateSubModules.sh",
"updateDependencies": "lerna bootstrap",
"build": "lerna run build",
"buildCore": "lerna run build --scope ag-grid --scope ag-grid-enterprise --scope ag-grid-angular --scope ag-grid-react --scope ag-grid-vue --scope ag-grid-aurelia",
"updateAndRebuild": "npm run updateDependencies && npm run build",
"initIntelliJ": "cp .idea/workspace.xml.default .idea/workspace.xml",
"init": "npm run initLatest && npm run updateAndRebuild"
},
"devDependencies": {
"lerna": "^2.11.0"
}
}
| {
"scripts": {
"initSubModules": "sh ./scripts/initialiseSubModules.sh",
"initLatest": "npm run initSubModules && sh ./scripts/switchToBranch.sh latest",
"initMaster": "npm run initSubModules && sh ./scripts/switchToBranch.sh master",
+ "updateExamples": "sh ./scripts/updateSubModules.sh",
"updateDependencies": "lerna bootstrap",
"build": "lerna run build",
"buildCore": "lerna run build --scope ag-grid --scope ag-grid-enterprise --scope ag-grid-angular --scope ag-grid-react --scope ag-grid-vue --scope ag-grid-aurelia",
"updateAndRebuild": "npm run updateDependencies && npm run build",
"initIntelliJ": "cp .idea/workspace.xml.default .idea/workspace.xml",
"init": "npm run initLatest && npm run updateAndRebuild"
},
"devDependencies": {
"lerna": "^2.11.0"
}
} | 1 | 0.0625 | 1 | 0 |
39622abd496cea11a92e25aac8dd68f1c5528e5b | spec/helpers.rb | spec/helpers.rb | module Helpers
def setup_reporter_spec(reporter, reporter_options = {})
before { ProgressHandler.configure {|config| config.reporters = {reporter => reporter_options} } }
let(:items) { 5.times.map &:to_s }
let(:progress_handler) { ProgressHandler.new name: 'Specs', report_gap: 2 }
end
end
| module Helpers
def setup_reporter_spec(reporter)
before { ProgressHandler.configure {|config| config.reporters = {reporter => reporter_options} } }
let(:reporter_options) { {} }
let(:items) { 5.times.map &:to_s }
let(:progress_handler) { ProgressHandler.new name: 'Specs', report_gap: 2 }
end
end
| Allow reporter parameters (re)definition in simple let statements | Allow reporter parameters (re)definition in simple let statements
| Ruby | mit | gandralf/progress_handler,gandralf/progress_handler | ruby | ## Code Before:
module Helpers
def setup_reporter_spec(reporter, reporter_options = {})
before { ProgressHandler.configure {|config| config.reporters = {reporter => reporter_options} } }
let(:items) { 5.times.map &:to_s }
let(:progress_handler) { ProgressHandler.new name: 'Specs', report_gap: 2 }
end
end
## Instruction:
Allow reporter parameters (re)definition in simple let statements
## Code After:
module Helpers
def setup_reporter_spec(reporter)
before { ProgressHandler.configure {|config| config.reporters = {reporter => reporter_options} } }
let(:reporter_options) { {} }
let(:items) { 5.times.map &:to_s }
let(:progress_handler) { ProgressHandler.new name: 'Specs', report_gap: 2 }
end
end
| module Helpers
- def setup_reporter_spec(reporter, reporter_options = {})
? -----------------------
+ def setup_reporter_spec(reporter)
before { ProgressHandler.configure {|config| config.reporters = {reporter => reporter_options} } }
-
+ let(:reporter_options) { {} }
let(:items) { 5.times.map &:to_s }
let(:progress_handler) { ProgressHandler.new name: 'Specs', report_gap: 2 }
end
end | 4 | 0.5 | 2 | 2 |
91d0ab7af1489ed71980408aeb30734811474121 | pkgs/development/libraries/ldns/default.nix | pkgs/development/libraries/ldns/default.nix | {stdenv, fetchurl, openssl, perl}:
stdenv.mkDerivation {
name = "ldns-1.6.11";
src = fetchurl {
url = "http://www.nlnetlabs.nl/downloads/ldns/ldns-1.6.11.tar.gz";
sha256 = "1248c9gkgfmjdmpp3lfd56vvln94ii54kbxa5iykxvcxivmbi4b8";
};
patchPhase = ''
sed -i 's,\$(srcdir)/doc/doxyparse.pl,perl $(srcdir)/doc/doxyparse.pl,' Makefile.in
'';
buildInputs = [ openssl perl ];
configureFlags = [ "--with-ssl=${openssl}" ];
meta = {
description = "Library with the aim of simplifying DNS programming in C";
license = "BSD";
homepage = "http://www.nlnetlabs.nl/projects/ldns/";
};
}
| {stdenv, fetchurl, openssl, perl}:
stdenv.mkDerivation rec {
name = "ldns-1.6.16";
src = fetchurl {
url = "http://www.nlnetlabs.nl/downloads/ldns/${name}.tar.gz";
sha256 = "15gn9m95r6sq2n55dw4r87p2aljb5lvy1w0y0br70wbr0p5zkci4";
};
patchPhase = ''
sed -i 's,\$(srcdir)/doc/doxyparse.pl,perl $(srcdir)/doc/doxyparse.pl,' Makefile.in
'';
buildInputs = [ openssl perl ];
configureFlags = [ "--with-ssl=${openssl}" "--with-drill" ];
meta = {
description = "Library with the aim of simplifying DNS programming in C";
license = "BSD";
homepage = "http://www.nlnetlabs.nl/projects/ldns/";
};
}
| Update to 1.6.16 and enable drill. | ldns: Update to 1.6.16 and enable drill.
Drill is a lightweight DNS lookup command using ldns, which we are going to use
in Gajim, to avoid a dependency on bind.
Signed-off-by: aszlig <ee1aa092358634f9c53f01b5a783726c9e21b35a@redmoonstudios.org>
| Nix | mit | NixOS/nixpkgs,NixOS/nixpkgs,SymbiFlow/nixpkgs,SymbiFlow/nixpkgs,triton/triton,NixOS/nixpkgs,NixOS/nixpkgs,NixOS/nixpkgs,NixOS/nixpkgs,SymbiFlow/nixpkgs,triton/triton,NixOS/nixpkgs,SymbiFlow/nixpkgs,NixOS/nixpkgs,NixOS/nixpkgs,triton/triton,SymbiFlow/nixpkgs,SymbiFlow/nixpkgs,NixOS/nixpkgs,SymbiFlow/nixpkgs,NixOS/nixpkgs,SymbiFlow/nixpkgs,triton/triton,NixOS/nixpkgs,triton/triton,NixOS/nixpkgs,NixOS/nixpkgs,SymbiFlow/nixpkgs,triton/triton,NixOS/nixpkgs,SymbiFlow/nixpkgs,SymbiFlow/nixpkgs,triton/triton,SymbiFlow/nixpkgs,SymbiFlow/nixpkgs,triton/triton | nix | ## Code Before:
{stdenv, fetchurl, openssl, perl}:
stdenv.mkDerivation {
name = "ldns-1.6.11";
src = fetchurl {
url = "http://www.nlnetlabs.nl/downloads/ldns/ldns-1.6.11.tar.gz";
sha256 = "1248c9gkgfmjdmpp3lfd56vvln94ii54kbxa5iykxvcxivmbi4b8";
};
patchPhase = ''
sed -i 's,\$(srcdir)/doc/doxyparse.pl,perl $(srcdir)/doc/doxyparse.pl,' Makefile.in
'';
buildInputs = [ openssl perl ];
configureFlags = [ "--with-ssl=${openssl}" ];
meta = {
description = "Library with the aim of simplifying DNS programming in C";
license = "BSD";
homepage = "http://www.nlnetlabs.nl/projects/ldns/";
};
}
## Instruction:
ldns: Update to 1.6.16 and enable drill.
Drill is a lightweight DNS lookup command using ldns, which we are going to use
in Gajim, to avoid a dependency on bind.
Signed-off-by: aszlig <ee1aa092358634f9c53f01b5a783726c9e21b35a@redmoonstudios.org>
## Code After:
{stdenv, fetchurl, openssl, perl}:
stdenv.mkDerivation rec {
name = "ldns-1.6.16";
src = fetchurl {
url = "http://www.nlnetlabs.nl/downloads/ldns/${name}.tar.gz";
sha256 = "15gn9m95r6sq2n55dw4r87p2aljb5lvy1w0y0br70wbr0p5zkci4";
};
patchPhase = ''
sed -i 's,\$(srcdir)/doc/doxyparse.pl,perl $(srcdir)/doc/doxyparse.pl,' Makefile.in
'';
buildInputs = [ openssl perl ];
configureFlags = [ "--with-ssl=${openssl}" "--with-drill" ];
meta = {
description = "Library with the aim of simplifying DNS programming in C";
license = "BSD";
homepage = "http://www.nlnetlabs.nl/projects/ldns/";
};
}
| {stdenv, fetchurl, openssl, perl}:
- stdenv.mkDerivation {
+ stdenv.mkDerivation rec {
? ++++
- name = "ldns-1.6.11";
? ^
+ name = "ldns-1.6.16";
? ^
+
src = fetchurl {
- url = "http://www.nlnetlabs.nl/downloads/ldns/ldns-1.6.11.tar.gz";
? ^^ ^^^^^^^^
+ url = "http://www.nlnetlabs.nl/downloads/ldns/${name}.tar.gz";
? ^^ ^^^^
- sha256 = "1248c9gkgfmjdmpp3lfd56vvln94ii54kbxa5iykxvcxivmbi4b8";
+ sha256 = "15gn9m95r6sq2n55dw4r87p2aljb5lvy1w0y0br70wbr0p5zkci4";
};
patchPhase = ''
sed -i 's,\$(srcdir)/doc/doxyparse.pl,perl $(srcdir)/doc/doxyparse.pl,' Makefile.in
'';
buildInputs = [ openssl perl ];
- configureFlags = [ "--with-ssl=${openssl}" ];
+ configureFlags = [ "--with-ssl=${openssl}" "--with-drill" ];
? +++++++++++++++
meta = {
description = "Library with the aim of simplifying DNS programming in C";
license = "BSD";
homepage = "http://www.nlnetlabs.nl/projects/ldns/";
};
} | 11 | 0.478261 | 6 | 5 |
17655f4b099ac840712dd95ad989f7b41301b83c | case_conversion/__init__.py | case_conversion/__init__.py | from case_conversion import (
camelcase, pascalcase, snakecase, dashcase, kebabcase, spinalcase,
constcase, dotcase, separate_words, slashcase, backslashcase)
| from __future__ import absolute_import
from .case_conversion import (
camelcase, pascalcase, snakecase, dashcase, kebabcase, spinalcase,
constcase, dotcase, separate_words, slashcase, backslashcase)
| Fix import errors from implicit-relative imports. | Fix import errors from implicit-relative imports.
| Python | mit | AlejandroFrias/case-conversion | python | ## Code Before:
from case_conversion import (
camelcase, pascalcase, snakecase, dashcase, kebabcase, spinalcase,
constcase, dotcase, separate_words, slashcase, backslashcase)
## Instruction:
Fix import errors from implicit-relative imports.
## Code After:
from __future__ import absolute_import
from .case_conversion import (
camelcase, pascalcase, snakecase, dashcase, kebabcase, spinalcase,
constcase, dotcase, separate_words, slashcase, backslashcase)
| + from __future__ import absolute_import
+
- from case_conversion import (
+ from .case_conversion import (
? +
camelcase, pascalcase, snakecase, dashcase, kebabcase, spinalcase,
constcase, dotcase, separate_words, slashcase, backslashcase) | 4 | 1.333333 | 3 | 1 |
62a43fd244436dfa79d7ca6de90835a7eeb9ceb7 | README.md | README.md |
[](https://codeclimate.com/github/ddollar/github-backup)
Back up your Github repositories locally.
To install it as a Gem, just run:
$ gem install github-backup
To use it:
$ github-backup ddollar /path/to/my/backup/root
## Authenticated Usage
### Seamless authentication using .gitconfig defaults
You will need a `~/.gitconfig` file in place with a `[github]` section
See: http://github.com/guides/tell-git-your-user-name-and-email-address
### Specify authentication at the command line
If you don't have a `[github]` section set up in your `~/.gitconfig` file, you
can provide your Github OAuth access token at the command line.
## License
MIT License
## Authorship
Created by David Dollar
Fixed for Github v3 by [Gareth Rees](https://github.com/garethrees)
[Other Contributors](https://github.com/ddollar/github-backup/graphs/contributors)
## Copyright
Copyright (c) 2010 David Dollar.
|
[](https://travis-ci.org/ddollar/github-backup)
[](https://codeclimate.com/github/ddollar/github-backup)
Back up your Github repositories locally.
To install it as a Gem, just run:
$ gem install github-backup
To use it:
$ github-backup ddollar /path/to/my/backup/root
## Authenticated Usage
### Seamless authentication using .gitconfig defaults
You will need a `~/.gitconfig` file in place with a `[github]` section
See: http://github.com/guides/tell-git-your-user-name-and-email-address
### Specify authentication at the command line
If you don't have a `[github]` section set up in your `~/.gitconfig` file, you
can provide your Github OAuth access token at the command line.
## License
MIT License
## Authorship
Created by David Dollar
Fixed for Github v3 by [Gareth Rees](https://github.com/garethrees)
[Other Contributors](https://github.com/ddollar/github-backup/graphs/contributors)
## Copyright
Copyright (c) 2010 David Dollar.
| Add Travis CI status badge | Add Travis CI status badge | Markdown | mit | ddollar/github-backup,garethrees/github-backup | markdown | ## Code Before:
[](https://codeclimate.com/github/ddollar/github-backup)
Back up your Github repositories locally.
To install it as a Gem, just run:
$ gem install github-backup
To use it:
$ github-backup ddollar /path/to/my/backup/root
## Authenticated Usage
### Seamless authentication using .gitconfig defaults
You will need a `~/.gitconfig` file in place with a `[github]` section
See: http://github.com/guides/tell-git-your-user-name-and-email-address
### Specify authentication at the command line
If you don't have a `[github]` section set up in your `~/.gitconfig` file, you
can provide your Github OAuth access token at the command line.
## License
MIT License
## Authorship
Created by David Dollar
Fixed for Github v3 by [Gareth Rees](https://github.com/garethrees)
[Other Contributors](https://github.com/ddollar/github-backup/graphs/contributors)
## Copyright
Copyright (c) 2010 David Dollar.
## Instruction:
Add Travis CI status badge
## Code After:
[](https://travis-ci.org/ddollar/github-backup)
[](https://codeclimate.com/github/ddollar/github-backup)
Back up your Github repositories locally.
To install it as a Gem, just run:
$ gem install github-backup
To use it:
$ github-backup ddollar /path/to/my/backup/root
## Authenticated Usage
### Seamless authentication using .gitconfig defaults
You will need a `~/.gitconfig` file in place with a `[github]` section
See: http://github.com/guides/tell-git-your-user-name-and-email-address
### Specify authentication at the command line
If you don't have a `[github]` section set up in your `~/.gitconfig` file, you
can provide your Github OAuth access token at the command line.
## License
MIT License
## Authorship
Created by David Dollar
Fixed for Github v3 by [Gareth Rees](https://github.com/garethrees)
[Other Contributors](https://github.com/ddollar/github-backup/graphs/contributors)
## Copyright
Copyright (c) 2010 David Dollar.
|
+ [](https://travis-ci.org/ddollar/github-backup)
[](https://codeclimate.com/github/ddollar/github-backup)
Back up your Github repositories locally.
To install it as a Gem, just run:
$ gem install github-backup
To use it:
$ github-backup ddollar /path/to/my/backup/root
## Authenticated Usage
### Seamless authentication using .gitconfig defaults
You will need a `~/.gitconfig` file in place with a `[github]` section
See: http://github.com/guides/tell-git-your-user-name-and-email-address
### Specify authentication at the command line
If you don't have a `[github]` section set up in your `~/.gitconfig` file, you
can provide your Github OAuth access token at the command line.
## License
MIT License
## Authorship
Created by David Dollar
Fixed for Github v3 by [Gareth Rees](https://github.com/garethrees)
[Other Contributors](https://github.com/ddollar/github-backup/graphs/contributors)
## Copyright
Copyright (c) 2010 David Dollar. | 1 | 0.02439 | 1 | 0 |
330a81b9b555162ab64e254090433b4ea300542a | README.md | README.md | glacier-cmd
===========
AWS Glacier command line client
## Configuration
The configuration for this tool needs to be located in a file called
.glaciercmd in your home directory with the following format:
```
[configuration]
aws_key:<your aws api key>
aws_secret:<your aws secret>
dynamodb_table:<DynamoDB table to use>
```
## Options
- `list vaults` List available vaults.
- `create vault <vault name>` Create a vault with the given name.
- `delete vault <vault name>` Delete the vault with the given name.
- `vault info <vault name>` Get information on the vault with the given name.
- `upload file <file name> to vault <vault name>` Upload a file to the given vault.
- `delete archive <archive id> from <vault name>` Delete an archive from the given vault.
- `list known archives for vault <vault name>` List known archives from the given vault.
## License
MIT - See the LICENSE file
| glacier-cmd
===========
AWS Glacier command line client
## Configuration
The configuration for this tool needs to be located in a file called
.glaciercmd in your home directory with the following format:
```
[configuration]
aws_key:<your aws api key>
aws_secret:<your aws secret>
dynamodb_table:<DynamoDB table to use>
```
## Options
- `list vaults` List available vaults.
- `create vault <vault name>` Create a vault with the given name.
- `delete vault <vault name>` Delete the vault with the given name.
- `vault info <vault name>` Get information on the vault with the given name.
- `upload file <file name> to vault <vault name>` Upload a file to the given vault.
- `delete archive <archive id> from <vault name>` Delete an archive from the given vault.
- `list known archives for vault <vault name>` List known archives from the given vault.
- `create inventory job for vault <vault name>` Create an inventory job for the given vault.
## License
MIT - See the LICENSE file
| Document create inventory job for vault | Document create inventory job for vault
| Markdown | mit | carsonmcdonald/glacier-cmd | markdown | ## Code Before:
glacier-cmd
===========
AWS Glacier command line client
## Configuration
The configuration for this tool needs to be located in a file called
.glaciercmd in your home directory with the following format:
```
[configuration]
aws_key:<your aws api key>
aws_secret:<your aws secret>
dynamodb_table:<DynamoDB table to use>
```
## Options
- `list vaults` List available vaults.
- `create vault <vault name>` Create a vault with the given name.
- `delete vault <vault name>` Delete the vault with the given name.
- `vault info <vault name>` Get information on the vault with the given name.
- `upload file <file name> to vault <vault name>` Upload a file to the given vault.
- `delete archive <archive id> from <vault name>` Delete an archive from the given vault.
- `list known archives for vault <vault name>` List known archives from the given vault.
## License
MIT - See the LICENSE file
## Instruction:
Document create inventory job for vault
## Code After:
glacier-cmd
===========
AWS Glacier command line client
## Configuration
The configuration for this tool needs to be located in a file called
.glaciercmd in your home directory with the following format:
```
[configuration]
aws_key:<your aws api key>
aws_secret:<your aws secret>
dynamodb_table:<DynamoDB table to use>
```
## Options
- `list vaults` List available vaults.
- `create vault <vault name>` Create a vault with the given name.
- `delete vault <vault name>` Delete the vault with the given name.
- `vault info <vault name>` Get information on the vault with the given name.
- `upload file <file name> to vault <vault name>` Upload a file to the given vault.
- `delete archive <archive id> from <vault name>` Delete an archive from the given vault.
- `list known archives for vault <vault name>` List known archives from the given vault.
- `create inventory job for vault <vault name>` Create an inventory job for the given vault.
## License
MIT - See the LICENSE file
| glacier-cmd
===========
AWS Glacier command line client
## Configuration
The configuration for this tool needs to be located in a file called
.glaciercmd in your home directory with the following format:
```
[configuration]
aws_key:<your aws api key>
aws_secret:<your aws secret>
dynamodb_table:<DynamoDB table to use>
```
## Options
- `list vaults` List available vaults.
- `create vault <vault name>` Create a vault with the given name.
- `delete vault <vault name>` Delete the vault with the given name.
- `vault info <vault name>` Get information on the vault with the given name.
- `upload file <file name> to vault <vault name>` Upload a file to the given vault.
- `delete archive <archive id> from <vault name>` Delete an archive from the given vault.
- `list known archives for vault <vault name>` List known archives from the given vault.
+ - `create inventory job for vault <vault name>` Create an inventory job for the given vault.
## License
MIT - See the LICENSE file | 1 | 0.032258 | 1 | 0 |
a285c3e9af1979ed4ad5a82bfdc3da6d4d56fc9d | pkgs/os-specific/linux/alsa-firmware/default.nix | pkgs/os-specific/linux/alsa-firmware/default.nix | { stdenv, buildPackages, autoreconfHook, fetchurl, fetchpatch }:
stdenv.mkDerivation rec {
name = "alsa-firmware-1.2.4";
src = fetchurl {
url = "mirror://alsa/firmware/${name}.tar.bz2";
sha256 = "sha256-tnttfQi8/CR+9v8KuIqZwYgwWjz1euLf0LzZpbNs1bs=";
};
nativeBuildInputs = [ autoreconfHook buildPackages.stdenv.cc ];
configureFlags = [
"--with-hotplug-dir=$(out)/lib/firmware"
];
dontStrip = true;
postInstall = ''
# These are lifted from the Arch PKGBUILD
# remove files which conflicts with linux-firmware
rm -rf $out/lib/firmware/{ct{efx,speq}.bin,ess,korg,sb16,yamaha}
# remove broken symlinks (broken upstream)
rm -rf $out/lib/firmware/turtlebeach
# remove empty dir
rm -rf $out/bin
'';
meta = {
homepage = "http://www.alsa-project.org/";
description = "Soundcard firmwares from the alsa project";
license = stdenv.lib.licenses.gpl2Plus;
platforms = stdenv.lib.platforms.linux;
};
}
| { stdenv, buildPackages, autoreconfHook, fetchurl, fetchpatch }:
stdenv.mkDerivation rec {
name = "alsa-firmware-1.2.1";
src = fetchurl {
url = "mirror://alsa/firmware/${name}.tar.bz2";
sha256 = "1aq8z8ajpjvcx7bwhwp36bh5idzximyn77ygk3ifs0my3mbpr8mf";
};
patches = [ (fetchpatch {
url = "https://github.com/alsa-project/alsa-firmware/commit/a8a478485a999ff9e4a8d8098107d3b946b70288.patch";
sha256 = "0zd7vrgz00hn02va5bkv7qj2395a1rl6f8jq1mwbryxs7hiysb78";
}) ];
nativeBuildInputs = [ autoreconfHook buildPackages.stdenv.cc ];
configureFlags = [
"--with-hotplug-dir=$(out)/lib/firmware"
];
dontStrip = true;
postInstall = ''
# These are lifted from the Arch PKGBUILD
# remove files which conflicts with linux-firmware
rm -rf $out/lib/firmware/{ct{efx,speq}.bin,ess,korg,sb16,yamaha}
# remove broken symlinks (broken upstream)
rm -rf $out/lib/firmware/turtlebeach
# remove empty dir
rm -rf $out/bin
'';
meta = {
homepage = "http://www.alsa-project.org/";
description = "Soundcard firmwares from the alsa project";
license = stdenv.lib.licenses.gpl2Plus;
platforms = stdenv.lib.platforms.linux;
};
}
| Revert "alsa-firmware: 1.2.1 -> 1.2.4" | Revert "alsa-firmware: 1.2.1 -> 1.2.4"
This was pushed to master without a PR and broke cross compilation of a
small cross-NixOS:
```
configure: error: in `/build/alsa-firmware-1.2.4':
configure: error: C compiler cannot create executables
See `config.log' for more details
```
Let's re-roll this in a PR.
This reverts commit 72f71e907135f250b656090eeb70d9f95f8e653d.
| Nix | mit | NixOS/nixpkgs,NixOS/nixpkgs,NixOS/nixpkgs,NixOS/nixpkgs,NixOS/nixpkgs,NixOS/nixpkgs,NixOS/nixpkgs,NixOS/nixpkgs,NixOS/nixpkgs,NixOS/nixpkgs,NixOS/nixpkgs,NixOS/nixpkgs,NixOS/nixpkgs,NixOS/nixpkgs,NixOS/nixpkgs | nix | ## Code Before:
{ stdenv, buildPackages, autoreconfHook, fetchurl, fetchpatch }:
stdenv.mkDerivation rec {
name = "alsa-firmware-1.2.4";
src = fetchurl {
url = "mirror://alsa/firmware/${name}.tar.bz2";
sha256 = "sha256-tnttfQi8/CR+9v8KuIqZwYgwWjz1euLf0LzZpbNs1bs=";
};
nativeBuildInputs = [ autoreconfHook buildPackages.stdenv.cc ];
configureFlags = [
"--with-hotplug-dir=$(out)/lib/firmware"
];
dontStrip = true;
postInstall = ''
# These are lifted from the Arch PKGBUILD
# remove files which conflicts with linux-firmware
rm -rf $out/lib/firmware/{ct{efx,speq}.bin,ess,korg,sb16,yamaha}
# remove broken symlinks (broken upstream)
rm -rf $out/lib/firmware/turtlebeach
# remove empty dir
rm -rf $out/bin
'';
meta = {
homepage = "http://www.alsa-project.org/";
description = "Soundcard firmwares from the alsa project";
license = stdenv.lib.licenses.gpl2Plus;
platforms = stdenv.lib.platforms.linux;
};
}
## Instruction:
Revert "alsa-firmware: 1.2.1 -> 1.2.4"
This was pushed to master without a PR and broke cross compilation of a
small cross-NixOS:
```
configure: error: in `/build/alsa-firmware-1.2.4':
configure: error: C compiler cannot create executables
See `config.log' for more details
```
Let's re-roll this in a PR.
This reverts commit 72f71e907135f250b656090eeb70d9f95f8e653d.
## Code After:
{ stdenv, buildPackages, autoreconfHook, fetchurl, fetchpatch }:
stdenv.mkDerivation rec {
name = "alsa-firmware-1.2.1";
src = fetchurl {
url = "mirror://alsa/firmware/${name}.tar.bz2";
sha256 = "1aq8z8ajpjvcx7bwhwp36bh5idzximyn77ygk3ifs0my3mbpr8mf";
};
patches = [ (fetchpatch {
url = "https://github.com/alsa-project/alsa-firmware/commit/a8a478485a999ff9e4a8d8098107d3b946b70288.patch";
sha256 = "0zd7vrgz00hn02va5bkv7qj2395a1rl6f8jq1mwbryxs7hiysb78";
}) ];
nativeBuildInputs = [ autoreconfHook buildPackages.stdenv.cc ];
configureFlags = [
"--with-hotplug-dir=$(out)/lib/firmware"
];
dontStrip = true;
postInstall = ''
# These are lifted from the Arch PKGBUILD
# remove files which conflicts with linux-firmware
rm -rf $out/lib/firmware/{ct{efx,speq}.bin,ess,korg,sb16,yamaha}
# remove broken symlinks (broken upstream)
rm -rf $out/lib/firmware/turtlebeach
# remove empty dir
rm -rf $out/bin
'';
meta = {
homepage = "http://www.alsa-project.org/";
description = "Soundcard firmwares from the alsa project";
license = stdenv.lib.licenses.gpl2Plus;
platforms = stdenv.lib.platforms.linux;
};
}
| { stdenv, buildPackages, autoreconfHook, fetchurl, fetchpatch }:
stdenv.mkDerivation rec {
- name = "alsa-firmware-1.2.4";
? ^
+ name = "alsa-firmware-1.2.1";
? ^
src = fetchurl {
url = "mirror://alsa/firmware/${name}.tar.bz2";
- sha256 = "sha256-tnttfQi8/CR+9v8KuIqZwYgwWjz1euLf0LzZpbNs1bs=";
+ sha256 = "1aq8z8ajpjvcx7bwhwp36bh5idzximyn77ygk3ifs0my3mbpr8mf";
};
+
+ patches = [ (fetchpatch {
+ url = "https://github.com/alsa-project/alsa-firmware/commit/a8a478485a999ff9e4a8d8098107d3b946b70288.patch";
+ sha256 = "0zd7vrgz00hn02va5bkv7qj2395a1rl6f8jq1mwbryxs7hiysb78";
+ }) ];
nativeBuildInputs = [ autoreconfHook buildPackages.stdenv.cc ];
configureFlags = [
"--with-hotplug-dir=$(out)/lib/firmware"
];
dontStrip = true;
postInstall = ''
# These are lifted from the Arch PKGBUILD
# remove files which conflicts with linux-firmware
rm -rf $out/lib/firmware/{ct{efx,speq}.bin,ess,korg,sb16,yamaha}
# remove broken symlinks (broken upstream)
rm -rf $out/lib/firmware/turtlebeach
# remove empty dir
rm -rf $out/bin
'';
meta = {
homepage = "http://www.alsa-project.org/";
description = "Soundcard firmwares from the alsa project";
license = stdenv.lib.licenses.gpl2Plus;
platforms = stdenv.lib.platforms.linux;
};
} | 9 | 0.257143 | 7 | 2 |
503e549878e577048af54eb8f227877c7833f99e | docker/start-hecuba.sh | docker/start-hecuba.sh |
DATA_DIR="/data/hecuba/${HOSTNAME}/"
mkdir -p ${DATA_DIR}
java -jar -Djava.io.tmpdir=${DATA_DIR} /uberjar.jar
|
CONFIG_FILE=/.hecuba.edn
DATA_DIR="/data/hecuba/${HOSTNAME}/"
mkdir -p ${DATA_DIR}
CASSANDRA_SESSION_KEYSPACE=${HECUBA_KEYSPACE?NOT DEFINED}
S3_FILE_BUCKET=${HECUBA_FILE_BUCKET?NOT DEFINED}
S3_STATUS_BUCKET=${HECUBA_STATUS_BUCKET?NOT DEFINED}
echo "DATA_DIR is ${DATA_DIR}"
echo "CASSANDRA_SESSION_KEYSPACE is ${CASSANDRA_SESSION_KEYSPACE}"
echo "S3_FILE_BUCKET is ${S3_FILE_BUCKET}"
echo "S3_STATUS_BUCKET is ${S3_STATUS_BUCKET}"
echo <<EOF > ${CONFIG_FILE}
{
:cassandra-session {:keyspace :${CASSANDRA_SESSION_KEYSPACE}
:hecuba-session {:keyspace ${CASSANDRA_SESSION_KEYSPACE}}
:search-session {:host "${ES01_TCP_PORT_9200_ADDR}" }
:s3 {:access-key "${AWS_ACCESS_KEY_ID:?NOT DEFINED}"
:secret-key "${AWS_SECRET_ACCESS_KEY:?NOT DEFINED}"
:file-bucket "${S3_FILE_BUCKET}"
:status-bucket "${S3_STATUS_BUCKET}"
:download-dir "${DATA_DIR}"}
}
EOF
export HOME=/
java -jar -Djava.io.tmpdir=${DATA_DIR} /uberjar.jar
| Add some checks and debug output to docker start script. | Add some checks and debug output to docker start script.
| Shell | epl-1.0 | MastodonC/kixi.hecuba,MastodonC/kixi.hecuba,MastodonC/kixi.hecuba,MastodonC/kixi.hecuba,MastodonC/kixi.hecuba | shell | ## Code Before:
DATA_DIR="/data/hecuba/${HOSTNAME}/"
mkdir -p ${DATA_DIR}
java -jar -Djava.io.tmpdir=${DATA_DIR} /uberjar.jar
## Instruction:
Add some checks and debug output to docker start script.
## Code After:
CONFIG_FILE=/.hecuba.edn
DATA_DIR="/data/hecuba/${HOSTNAME}/"
mkdir -p ${DATA_DIR}
CASSANDRA_SESSION_KEYSPACE=${HECUBA_KEYSPACE?NOT DEFINED}
S3_FILE_BUCKET=${HECUBA_FILE_BUCKET?NOT DEFINED}
S3_STATUS_BUCKET=${HECUBA_STATUS_BUCKET?NOT DEFINED}
echo "DATA_DIR is ${DATA_DIR}"
echo "CASSANDRA_SESSION_KEYSPACE is ${CASSANDRA_SESSION_KEYSPACE}"
echo "S3_FILE_BUCKET is ${S3_FILE_BUCKET}"
echo "S3_STATUS_BUCKET is ${S3_STATUS_BUCKET}"
echo <<EOF > ${CONFIG_FILE}
{
:cassandra-session {:keyspace :${CASSANDRA_SESSION_KEYSPACE}
:hecuba-session {:keyspace ${CASSANDRA_SESSION_KEYSPACE}}
:search-session {:host "${ES01_TCP_PORT_9200_ADDR}" }
:s3 {:access-key "${AWS_ACCESS_KEY_ID:?NOT DEFINED}"
:secret-key "${AWS_SECRET_ACCESS_KEY:?NOT DEFINED}"
:file-bucket "${S3_FILE_BUCKET}"
:status-bucket "${S3_STATUS_BUCKET}"
:download-dir "${DATA_DIR}"}
}
EOF
export HOME=/
java -jar -Djava.io.tmpdir=${DATA_DIR} /uberjar.jar
| +
+ CONFIG_FILE=/.hecuba.edn
DATA_DIR="/data/hecuba/${HOSTNAME}/"
mkdir -p ${DATA_DIR}
+ CASSANDRA_SESSION_KEYSPACE=${HECUBA_KEYSPACE?NOT DEFINED}
+ S3_FILE_BUCKET=${HECUBA_FILE_BUCKET?NOT DEFINED}
+ S3_STATUS_BUCKET=${HECUBA_STATUS_BUCKET?NOT DEFINED}
+
+ echo "DATA_DIR is ${DATA_DIR}"
+ echo "CASSANDRA_SESSION_KEYSPACE is ${CASSANDRA_SESSION_KEYSPACE}"
+ echo "S3_FILE_BUCKET is ${S3_FILE_BUCKET}"
+ echo "S3_STATUS_BUCKET is ${S3_STATUS_BUCKET}"
+
+ echo <<EOF > ${CONFIG_FILE}
+
+ {
+ :cassandra-session {:keyspace :${CASSANDRA_SESSION_KEYSPACE}
+ :hecuba-session {:keyspace ${CASSANDRA_SESSION_KEYSPACE}}
+ :search-session {:host "${ES01_TCP_PORT_9200_ADDR}" }
+ :s3 {:access-key "${AWS_ACCESS_KEY_ID:?NOT DEFINED}"
+ :secret-key "${AWS_SECRET_ACCESS_KEY:?NOT DEFINED}"
+ :file-bucket "${S3_FILE_BUCKET}"
+ :status-bucket "${S3_STATUS_BUCKET}"
+ :download-dir "${DATA_DIR}"}
+ }
+
+ EOF
+
+ export HOME=/
+
java -jar -Djava.io.tmpdir=${DATA_DIR} /uberjar.jar | 28 | 5.6 | 28 | 0 |
8065830ff2163bf1178b204a5873b7d56b40a228 | src/scripts/commons/d3-utils.js | src/scripts/commons/d3-utils.js | // External
import * as d3 from 'd3';
export function mergeSelections (selections) {
// Create a new empty selection
const mergedSelection = d3.selectAll('.d3-list-graph-not-existent');
function pushSelection (selection) {
selection.each(function pushDomNode () {
mergedSelection[0].push(this);
});
}
for (let i = selections.length; i--;) {
pushSelection(selections[i]);
}
return mergedSelection;
}
export function allTransitionsEnded (transition, callback) {
if (transition.size() === 0) {
callback();
}
let n = 0;
transition
.each(() => ++n)
.on('end', function (...args) {
if (!--n) callback.apply(this, args);
});
}
| // External
import * as d3 from 'd3';
export function mergeSelections (selections) {
// Create a new empty selection
const mergedSelection = d3.selectAll('.d3-list-graph-not-existent');
for (let i = selections.length; i--;) {
mergedSelection._groups = mergedSelection._groups.concat(
selections[i]._groups
);
}
return mergedSelection;
}
export function allTransitionsEnded (transition, callback) {
if (transition.size() === 0) {
callback();
}
let n = 0;
transition
.each(() => ++n)
.on('end', function (...args) {
if (!--n) callback.apply(this, args);
});
}
| Adjust selection merge method to D3 v4. | Adjust selection merge method to D3 v4.
| JavaScript | mit | flekschas/d3-list-graph,flekschas/d3-list-graph | javascript | ## Code Before:
// External
import * as d3 from 'd3';
export function mergeSelections (selections) {
// Create a new empty selection
const mergedSelection = d3.selectAll('.d3-list-graph-not-existent');
function pushSelection (selection) {
selection.each(function pushDomNode () {
mergedSelection[0].push(this);
});
}
for (let i = selections.length; i--;) {
pushSelection(selections[i]);
}
return mergedSelection;
}
export function allTransitionsEnded (transition, callback) {
if (transition.size() === 0) {
callback();
}
let n = 0;
transition
.each(() => ++n)
.on('end', function (...args) {
if (!--n) callback.apply(this, args);
});
}
## Instruction:
Adjust selection merge method to D3 v4.
## Code After:
// External
import * as d3 from 'd3';
export function mergeSelections (selections) {
// Create a new empty selection
const mergedSelection = d3.selectAll('.d3-list-graph-not-existent');
for (let i = selections.length; i--;) {
mergedSelection._groups = mergedSelection._groups.concat(
selections[i]._groups
);
}
return mergedSelection;
}
export function allTransitionsEnded (transition, callback) {
if (transition.size() === 0) {
callback();
}
let n = 0;
transition
.each(() => ++n)
.on('end', function (...args) {
if (!--n) callback.apply(this, args);
});
}
| // External
import * as d3 from 'd3';
export function mergeSelections (selections) {
// Create a new empty selection
const mergedSelection = d3.selectAll('.d3-list-graph-not-existent');
- function pushSelection (selection) {
- selection.each(function pushDomNode () {
- mergedSelection[0].push(this);
- });
- }
-
for (let i = selections.length; i--;) {
- pushSelection(selections[i]);
+ mergedSelection._groups = mergedSelection._groups.concat(
+ selections[i]._groups
+ );
}
return mergedSelection;
}
export function allTransitionsEnded (transition, callback) {
if (transition.size() === 0) {
callback();
}
let n = 0;
transition
.each(() => ++n)
.on('end', function (...args) {
if (!--n) callback.apply(this, args);
});
} | 10 | 0.322581 | 3 | 7 |
7edaa535d61991f57f2b6f94fae9381bdd1c819f | app/models/bookmark.rb | app/models/bookmark.rb | class Bookmark < ApplicationRecord
has_many :bookmark_tags
has_many :tags, through: :bookmark_tags
end
| class Bookmark < ApplicationRecord
has_many :bookmark_tags
has_many :tags, through: :bookmark_tags
enum source_type: { "Post" => 0, "Publication" => 1, "Video" => 2, "Thread" => 3, "Other" => 4}
end
| Add enum source_type in Bookmark model | Add enum source_type in Bookmark model
| Ruby | mit | Dom-Mc/shareMark,Dom-Mc/shareMark,Dom-Mc/shareMark | ruby | ## Code Before:
class Bookmark < ApplicationRecord
has_many :bookmark_tags
has_many :tags, through: :bookmark_tags
end
## Instruction:
Add enum source_type in Bookmark model
## Code After:
class Bookmark < ApplicationRecord
has_many :bookmark_tags
has_many :tags, through: :bookmark_tags
enum source_type: { "Post" => 0, "Publication" => 1, "Video" => 2, "Thread" => 3, "Other" => 4}
end
| class Bookmark < ApplicationRecord
has_many :bookmark_tags
has_many :tags, through: :bookmark_tags
+
+ enum source_type: { "Post" => 0, "Publication" => 1, "Video" => 2, "Thread" => 3, "Other" => 4}
end | 2 | 0.5 | 2 | 0 |
ecb53b917a5447928d7fd41eb6d1cf19f7e77c94 | app/views/quotes/index.html.haml | app/views/quotes/index.html.haml | .hero-unit
.pull-right
= image_tag 'http://www.gravatar.com/avatar/9100faec8b9af029823b0a0ff208b942', class: 'img-circle'
%h1 Quoty
%h3 Quotes worth spreading. Words worth remembering.
%ul#quotes= render @quotes
= link_to 'Upload one?', new_quote_path, :class => 'btn btn-primary'
| .hero-unit
.pull-right
= image_tag 'http://www.gravatar.com/avatar/9100faec8b9af029823b0a0ff208b942', class: 'img-circle'
%h1 Quoty
%h3 Quotes that worth spreading. Words that worth remembering.
%p Here are my collections of meaningful words or quotes, some by great people already passed away, some by cool guys around me.
%ul#quotes= render @quotes
= link_to 'Upload one?', new_quote_path, :class => 'btn btn-primary'
| Add top page tag line | Add top page tag line
| Haml | mit | kinopyo/quoty,dddaisuke/quoty,kinopyo/quoty,dddaisuke/quoty,dddaisuke/quoty | haml | ## Code Before:
.hero-unit
.pull-right
= image_tag 'http://www.gravatar.com/avatar/9100faec8b9af029823b0a0ff208b942', class: 'img-circle'
%h1 Quoty
%h3 Quotes worth spreading. Words worth remembering.
%ul#quotes= render @quotes
= link_to 'Upload one?', new_quote_path, :class => 'btn btn-primary'
## Instruction:
Add top page tag line
## Code After:
.hero-unit
.pull-right
= image_tag 'http://www.gravatar.com/avatar/9100faec8b9af029823b0a0ff208b942', class: 'img-circle'
%h1 Quoty
%h3 Quotes that worth spreading. Words that worth remembering.
%p Here are my collections of meaningful words or quotes, some by great people already passed away, some by cool guys around me.
%ul#quotes= render @quotes
= link_to 'Upload one?', new_quote_path, :class => 'btn btn-primary'
| .hero-unit
.pull-right
= image_tag 'http://www.gravatar.com/avatar/9100faec8b9af029823b0a0ff208b942', class: 'img-circle'
%h1 Quoty
- %h3 Quotes worth spreading. Words worth remembering.
+ %h3 Quotes that worth spreading. Words that worth remembering.
? +++++ +++++
+ %p Here are my collections of meaningful words or quotes, some by great people already passed away, some by cool guys around me.
%ul#quotes= render @quotes
= link_to 'Upload one?', new_quote_path, :class => 'btn btn-primary' | 3 | 0.3 | 2 | 1 |
9a2f7fcac723a43655a7aa89df1a40bc8dbe31e0 | src/Perl6/Metamodel/NativeHOW.pm | src/Perl6/Metamodel/NativeHOW.pm | class Perl6::Metamodel::NativeHOW
does Perl6::Metamodel::Naming
does Perl6::Metamodel::Documenting
does Perl6::Metamodel::Versioning
does Perl6::Metamodel::Stashing
does Perl6::Metamodel::MultipleInheritance
does Perl6::Metamodel::C3MRO
{
has $!composed;
my $archetypes := Perl6::Metamodel::Archetypes.new( :nominal(1), :inheritable(1) );
method archetypes() {
$archetypes
}
method new_type(:$name = '<anon>', :$repr = 'P6opaque', :$ver, :$auth) {
my $metaclass := self.new(:name($name), :ver($ver), :auth($auth));
self.add_stash(pir::repr_type_object_for__PPS($metaclass, $repr));
}
method compose($obj) {
$!composed := 1;
}
method is_composed($obj) {
$!composed
}
method type_check($obj, $checkee) {
# The only time we end up in here is if the type check cache was
# not yet published, which means the class isn't yet fully composed.
# Just hunt through MRO.
for self.mro($obj) {
if $_ =:= $checkee {
return 1;
}
}
0
}
}
| class Perl6::Metamodel::NativeHOW
does Perl6::Metamodel::Naming
does Perl6::Metamodel::Documenting
does Perl6::Metamodel::Versioning
does Perl6::Metamodel::Stashing
does Perl6::Metamodel::MultipleInheritance
does Perl6::Metamodel::C3MRO
does Perl6::Metamodel::MROBasedMethodDispatch
{
has $!composed;
my $archetypes := Perl6::Metamodel::Archetypes.new( :nominal(1), :inheritable(1) );
method archetypes() {
$archetypes
}
method new_type(:$name = '<anon>', :$repr = 'P6opaque', :$ver, :$auth) {
my $metaclass := self.new(:name($name), :ver($ver), :auth($auth));
self.add_stash(pir::repr_type_object_for__PPS($metaclass, $repr));
}
method compose($obj) {
self.publish_method_cache($obj);
$!composed := 1;
}
method is_composed($obj) {
$!composed
}
method method_table($obj) { nqp::hash() }
method submethod_table($obj) { nqp::hash() }
method type_check($obj, $checkee) {
# The only time we end up in here is if the type check cache was
# not yet published, which means the class isn't yet fully composed.
# Just hunt through MRO.
for self.mro($obj) {
if $_ =:= $checkee {
return 1;
}
}
0
}
}
| Make Int ~~ int and various other things on native type objects not explode. | Make Int ~~ int and various other things on native type objects not explode.
| Perl | artistic-2.0 | samcv/rakudo,rjbs/rakudo,awwaiid/rakudo,salortiz/rakudo,pmurias/rakudo,tony-o/rakudo,niner/rakudo,tony-o/rakudo,raydiak/rakudo,raydiak/rakudo,rakudo/rakudo,zostay/rakudo,ungrim97/rakudo,retupmoca/rakudo,ugexe/rakudo,cygx/rakudo,rjbs/rakudo,pmurias/rakudo,sergot/rakudo,b2gills/rakudo,tony-o/deb-rakudodaily,tony-o/rakudo,tbrowder/rakudo,tony-o/rakudo,laben/rakudo,teodozjan/rakudo,tony-o/rakudo,cygx/rakudo,samcv/rakudo,jonathanstowe/rakudo,niner/rakudo,Gnouc/rakudo,teodozjan/rakudo,cognominal/rakudo,b2gills/rakudo,jonathanstowe/rakudo,sjn/rakudo,usev6/rakudo,MasterDuke17/rakudo,awwaiid/rakudo,sjn/rakudo,salortiz/rakudo,nbrown/rakudo,usev6/rakudo,ugexe/rakudo,dwarring/rakudo,MasterDuke17/rakudo,rjbs/rakudo,sergot/rakudo,tbrowder/rakudo,rakudo/rakudo,labster/rakudo,ugexe/rakudo,rakudo/rakudo,Gnouc/rakudo,lucasbuchala/rakudo,raydiak/rakudo,teodozjan/rakudo,nunorc/rakudo,ab5tract/rakudo,Gnouc/rakudo,skids/rakudo,zostay/rakudo,cygx/rakudo,skids/rakudo,cognominal/rakudo,samcv/rakudo,tbrowder/rakudo,nbrown/rakudo,dankogai/rakudo,sergot/rakudo,tbrowder/rakudo,b2gills/rakudo,LLFourn/rakudo,salortiz/rakudo,azawawi/rakudo,ab5tract/rakudo,retupmoca/rakudo,zhuomingliang/rakudo,Gnouc/rakudo,dankogai/rakudo,Gnouc/rakudo,LLFourn/rakudo,nunorc/rakudo,Leont/rakudo,zostay/rakudo,zhuomingliang/rakudo,cognominal/rakudo,MasterDuke17/rakudo,lucasbuchala/rakudo,softmoth/rakudo,paultcochrane/rakudo,jonathanstowe/rakudo,Leont/rakudo,ab5tract/rakudo,salortiz/rakudo,retupmoca/rakudo,tony-o/deb-rakudodaily,niner/rakudo,awwaiid/rakudo,ab5tract/rakudo,zhuomingliang/rakudo,azawawi/rakudo,tony-o/deb-rakudodaily,niner/rakudo,salortiz/rakudo,salortiz/rakudo,cognominal/rakudo,ugexe/rakudo,ungrim97/rakudo,nbrown/rakudo,pmurias/rakudo,laben/rakudo,zostay/rakudo,dankogai/rakudo,paultcochrane/rakudo,teodozjan/rakudo,pmurias/rakudo,sjn/rakudo,rakudo/rakudo,nbrown/rakudo,lucasbuchala/rakudo,nunorc/rakudo,labster/rakudo,MasterDuke17/rakudo,tony-o/rakudo,sjn/rakudo,samcv/rakudo,Gnouc/rakudo,b2gills/rakudo,samcv/rakudo,lucasbuchala/rakudo,nunorc/rakudo,tbrowder/rakudo,Leont/rakudo,cygx/rakudo,zhuomingliang/rakudo,dwarring/rakudo,laben/rakudo,awwaiid/rakudo,paultcochrane/rakudo,skids/rakudo,rakudo/rakudo,tony-o/deb-rakudodaily,dwarring/rakudo,ugexe/rakudo,laben/rakudo,softmoth/rakudo,tony-o/deb-rakudodaily,LLFourn/rakudo,usev6/rakudo,azawawi/rakudo,labster/rakudo,azawawi/rakudo,dwarring/rakudo,raydiak/rakudo,ab5tract/rakudo,cognominal/rakudo,MasterDuke17/rakudo,softmoth/rakudo,rakudo/rakudo,tbrowder/rakudo,softmoth/rakudo,labster/rakudo,cygx/rakudo,MasterDuke17/rakudo,nbrown/rakudo,skids/rakudo,ungrim97/rakudo,LLFourn/rakudo,tony-o/deb-rakudodaily,ungrim97/rakudo,ungrim97/rakudo,jonathanstowe/rakudo,dankogai/rakudo,skids/rakudo,awwaiid/rakudo,labster/rakudo,b2gills/rakudo,lucasbuchala/rakudo,dankogai/rakudo,sergot/rakudo,LLFourn/rakudo,usev6/rakudo,sjn/rakudo,azawawi/rakudo,paultcochrane/rakudo,softmoth/rakudo,jonathanstowe/rakudo,nbrown/rakudo,tony-o/deb-rakudodaily,labster/rakudo,paultcochrane/rakudo,retupmoca/rakudo,usev6/rakudo,Leont/rakudo | perl | ## Code Before:
class Perl6::Metamodel::NativeHOW
does Perl6::Metamodel::Naming
does Perl6::Metamodel::Documenting
does Perl6::Metamodel::Versioning
does Perl6::Metamodel::Stashing
does Perl6::Metamodel::MultipleInheritance
does Perl6::Metamodel::C3MRO
{
has $!composed;
my $archetypes := Perl6::Metamodel::Archetypes.new( :nominal(1), :inheritable(1) );
method archetypes() {
$archetypes
}
method new_type(:$name = '<anon>', :$repr = 'P6opaque', :$ver, :$auth) {
my $metaclass := self.new(:name($name), :ver($ver), :auth($auth));
self.add_stash(pir::repr_type_object_for__PPS($metaclass, $repr));
}
method compose($obj) {
$!composed := 1;
}
method is_composed($obj) {
$!composed
}
method type_check($obj, $checkee) {
# The only time we end up in here is if the type check cache was
# not yet published, which means the class isn't yet fully composed.
# Just hunt through MRO.
for self.mro($obj) {
if $_ =:= $checkee {
return 1;
}
}
0
}
}
## Instruction:
Make Int ~~ int and various other things on native type objects not explode.
## Code After:
class Perl6::Metamodel::NativeHOW
does Perl6::Metamodel::Naming
does Perl6::Metamodel::Documenting
does Perl6::Metamodel::Versioning
does Perl6::Metamodel::Stashing
does Perl6::Metamodel::MultipleInheritance
does Perl6::Metamodel::C3MRO
does Perl6::Metamodel::MROBasedMethodDispatch
{
has $!composed;
my $archetypes := Perl6::Metamodel::Archetypes.new( :nominal(1), :inheritable(1) );
method archetypes() {
$archetypes
}
method new_type(:$name = '<anon>', :$repr = 'P6opaque', :$ver, :$auth) {
my $metaclass := self.new(:name($name), :ver($ver), :auth($auth));
self.add_stash(pir::repr_type_object_for__PPS($metaclass, $repr));
}
method compose($obj) {
self.publish_method_cache($obj);
$!composed := 1;
}
method is_composed($obj) {
$!composed
}
method method_table($obj) { nqp::hash() }
method submethod_table($obj) { nqp::hash() }
method type_check($obj, $checkee) {
# The only time we end up in here is if the type check cache was
# not yet published, which means the class isn't yet fully composed.
# Just hunt through MRO.
for self.mro($obj) {
if $_ =:= $checkee {
return 1;
}
}
0
}
}
| class Perl6::Metamodel::NativeHOW
does Perl6::Metamodel::Naming
does Perl6::Metamodel::Documenting
does Perl6::Metamodel::Versioning
does Perl6::Metamodel::Stashing
- does Perl6::Metamodel::MultipleInheritance
+ does Perl6::Metamodel::MultipleInheritance
? ++++
does Perl6::Metamodel::C3MRO
+ does Perl6::Metamodel::MROBasedMethodDispatch
{
has $!composed;
my $archetypes := Perl6::Metamodel::Archetypes.new( :nominal(1), :inheritable(1) );
method archetypes() {
$archetypes
}
method new_type(:$name = '<anon>', :$repr = 'P6opaque', :$ver, :$auth) {
my $metaclass := self.new(:name($name), :ver($ver), :auth($auth));
self.add_stash(pir::repr_type_object_for__PPS($metaclass, $repr));
}
method compose($obj) {
+ self.publish_method_cache($obj);
$!composed := 1;
}
method is_composed($obj) {
$!composed
}
+
+ method method_table($obj) { nqp::hash() }
+ method submethod_table($obj) { nqp::hash() }
method type_check($obj, $checkee) {
# The only time we end up in here is if the type check cache was
# not yet published, which means the class isn't yet fully composed.
# Just hunt through MRO.
for self.mro($obj) {
if $_ =:= $checkee {
return 1;
}
}
0
}
} | 7 | 0.175 | 6 | 1 |
7c38ab1f739e90f79f692b714beb30a86d2b0473 | Gruntfile.js | Gruntfile.js | module.exports = function(grunt) {
grunt.initConfig({
pkg: grunt.file.readJSON('package.json'),
// contrib-watch
watch: {
all: {
files: ['src/**/*.js', 'lib/**/*.js', 'test/**/*.js', 'Gruntfile.js'],
tasks: ['test', 'build']
}
},
// contrib-jshint
jshint: {
all: ['Gruntfile.js', 'src/jquery.countdown.js', 'test/**/*.js']
},
// contrib-qunit
qunit: {
all: 'test/**/*.html'
},
// contrib-uglify
uglify: {
options: {
preserveComments: function(node, comment) {
// Preserve the license banner
return comment.col === 0 && comment.pos === 0;
}
},
all: {
files: {
'src/jquery.countdown.min.js': ['src/jquery.countdown.js']
}
}
}
});
grunt.loadNpmTasks('grunt-contrib-watch');
grunt.loadNpmTasks('grunt-contrib-qunit');
grunt.loadNpmTasks('grunt-contrib-jshint');
grunt.loadNpmTasks('grunt-contrib-uglify');
// Project tasks
grunt.registerTask('license', function() {
});
grunt.registerTask('test', ['jshint', 'qunit']);
grunt.registerTask('build', ['license', 'uglify']);
grunt.registerTask('default', ['test', 'build', 'watch']);
}; | module.exports = function(grunt) {
grunt.initConfig({
pkg: grunt.file.readJSON('package.json'),
// contrib-watch
watch: {
all: {
files: ['src/**/*.js', 'lib/**/*.js', 'test/**/*.js', 'Gruntfile.js'],
tasks: ['test', 'build']
}
},
// contrib-jshint
jshint: {
all: ['Gruntfile.js', 'src/jquery.countdown.js', 'test/**/*.js']
},
// contrib-qunit
qunit: {
all: 'test/**/*.html',
dev: 'test/scenario-jquery-1.9.1.html'
},
// contrib-uglify
uglify: {
options: {
preserveComments: function(node, comment) {
// Preserve the license banner
return comment.col === 0 && comment.pos === 0;
}
},
all: {
files: {
'src/jquery.countdown.min.js': ['src/jquery.countdown.js']
}
}
}
});
grunt.loadNpmTasks('grunt-contrib-watch');
grunt.loadNpmTasks('grunt-contrib-qunit');
grunt.loadNpmTasks('grunt-contrib-jshint');
grunt.loadNpmTasks('grunt-contrib-uglify');
// Project tasks
grunt.registerTask('license', function() {
});
grunt.registerTask('test', ['jshint', 'qunit:all']);
grunt.registerTask('build', ['qunit:all', 'uglify']);
grunt.registerTask('build:dev', ['qunit:dev', 'uglify']);
grunt.registerTask('default', ['build:dev', 'watch']);
}; | Add taks to speed development testing | Add taks to speed development testing
| JavaScript | mit | heriwahyudianto/jQuery.countdown,davidBearinTokyo/jQuery.countdown,adrianha/jQuery.countdown,nsekecharles/count-down,fashionsun/jQuery.countdown,shiyamkumar/jQuery.countdown,hilios/jQuery.countdown,baffolobill/jQuery.countdown,tbilous/jQuery.countdown,loki315zx/jQuery.countdown,DigitalCoder/jQuery.countdown,Renji/jQuery.countdown,masterweb121/jQuery.countdown,ayann/jQuery.countdown | javascript | ## Code Before:
module.exports = function(grunt) {
grunt.initConfig({
pkg: grunt.file.readJSON('package.json'),
// contrib-watch
watch: {
all: {
files: ['src/**/*.js', 'lib/**/*.js', 'test/**/*.js', 'Gruntfile.js'],
tasks: ['test', 'build']
}
},
// contrib-jshint
jshint: {
all: ['Gruntfile.js', 'src/jquery.countdown.js', 'test/**/*.js']
},
// contrib-qunit
qunit: {
all: 'test/**/*.html'
},
// contrib-uglify
uglify: {
options: {
preserveComments: function(node, comment) {
// Preserve the license banner
return comment.col === 0 && comment.pos === 0;
}
},
all: {
files: {
'src/jquery.countdown.min.js': ['src/jquery.countdown.js']
}
}
}
});
grunt.loadNpmTasks('grunt-contrib-watch');
grunt.loadNpmTasks('grunt-contrib-qunit');
grunt.loadNpmTasks('grunt-contrib-jshint');
grunt.loadNpmTasks('grunt-contrib-uglify');
// Project tasks
grunt.registerTask('license', function() {
});
grunt.registerTask('test', ['jshint', 'qunit']);
grunt.registerTask('build', ['license', 'uglify']);
grunt.registerTask('default', ['test', 'build', 'watch']);
};
## Instruction:
Add taks to speed development testing
## Code After:
module.exports = function(grunt) {
grunt.initConfig({
pkg: grunt.file.readJSON('package.json'),
// contrib-watch
watch: {
all: {
files: ['src/**/*.js', 'lib/**/*.js', 'test/**/*.js', 'Gruntfile.js'],
tasks: ['test', 'build']
}
},
// contrib-jshint
jshint: {
all: ['Gruntfile.js', 'src/jquery.countdown.js', 'test/**/*.js']
},
// contrib-qunit
qunit: {
all: 'test/**/*.html',
dev: 'test/scenario-jquery-1.9.1.html'
},
// contrib-uglify
uglify: {
options: {
preserveComments: function(node, comment) {
// Preserve the license banner
return comment.col === 0 && comment.pos === 0;
}
},
all: {
files: {
'src/jquery.countdown.min.js': ['src/jquery.countdown.js']
}
}
}
});
grunt.loadNpmTasks('grunt-contrib-watch');
grunt.loadNpmTasks('grunt-contrib-qunit');
grunt.loadNpmTasks('grunt-contrib-jshint');
grunt.loadNpmTasks('grunt-contrib-uglify');
// Project tasks
grunt.registerTask('license', function() {
});
grunt.registerTask('test', ['jshint', 'qunit:all']);
grunt.registerTask('build', ['qunit:all', 'uglify']);
grunt.registerTask('build:dev', ['qunit:dev', 'uglify']);
grunt.registerTask('default', ['build:dev', 'watch']);
}; | module.exports = function(grunt) {
grunt.initConfig({
pkg: grunt.file.readJSON('package.json'),
// contrib-watch
watch: {
all: {
files: ['src/**/*.js', 'lib/**/*.js', 'test/**/*.js', 'Gruntfile.js'],
tasks: ['test', 'build']
}
},
// contrib-jshint
jshint: {
all: ['Gruntfile.js', 'src/jquery.countdown.js', 'test/**/*.js']
},
// contrib-qunit
qunit: {
- all: 'test/**/*.html'
+ all: 'test/**/*.html',
? +
+ dev: 'test/scenario-jquery-1.9.1.html'
},
// contrib-uglify
uglify: {
options: {
preserveComments: function(node, comment) {
// Preserve the license banner
return comment.col === 0 && comment.pos === 0;
}
},
all: {
files: {
'src/jquery.countdown.min.js': ['src/jquery.countdown.js']
}
}
}
});
grunt.loadNpmTasks('grunt-contrib-watch');
grunt.loadNpmTasks('grunt-contrib-qunit');
grunt.loadNpmTasks('grunt-contrib-jshint');
grunt.loadNpmTasks('grunt-contrib-uglify');
// Project tasks
grunt.registerTask('license', function() {
});
- grunt.registerTask('test', ['jshint', 'qunit']);
+ grunt.registerTask('test', ['jshint', 'qunit:all']);
? ++++
- grunt.registerTask('build', ['license', 'uglify']);
? ^^^^^^
+ grunt.registerTask('build', ['qunit:all', 'uglify']);
? +++++++ ^
+ grunt.registerTask('build:dev', ['qunit:dev', 'uglify']);
- grunt.registerTask('default', ['test', 'build', 'watch']);
? --------
+ grunt.registerTask('default', ['build:dev', 'watch']);
? ++++
}; | 10 | 0.208333 | 6 | 4 |
0610fc00badf60a5113d3706819bcfd02231b79f | metadata/jp.redmine.redmineclient.txt | metadata/jp.redmine.redmineclient.txt | Categories:Internet
License:GPLv2+
Web Site:http://indication.github.io/OpenRedmine
Source Code:https://github.com/indication/OpenRedmine
Issue Tracker:https://github.com/indication/OpenRedmine/issues
Auto Name:OpenRedmine
Summary:Redmine client
Description:
Connect to Redmine servers (running 1.2 or later) and access the Wiki, Issue
Tracker, Source code, etc. easily.
Features:
* View issues offline
* Get API key from web site
* Filter (by Status/Tracker/Category/Priority)
* Download files
* Wiki
* Connect to unsafe SSL sites powered by transdroid
[https://raw.github.com/indication/OpenRedmine/development/doc/VERSION.en.md Release history]
.
Repo Type:git
Repo:https://github.com/indication/OpenRedmine.git
Build:3.6,38
commit=v3.6
subdir=OpenRedmine
gradle=yes
Build:3.7,39
commit=v3.7
subdir=OpenRedmine
gradle=yes
Build:3.8,40
commit=v3.8
subdir=OpenRedmine
gradle=yes
Build:3.9,41
commit=v3.9
subdir=OpenRedmine
gradle=yes
Auto Update Mode:Version v%v
Update Check Mode:Tags
Current Version:3.9
Current Version Code:41
| Categories:Internet
License:GPLv2+
Web Site:http://indication.github.io/OpenRedmine
Source Code:https://github.com/indication/OpenRedmine
Issue Tracker:https://github.com/indication/OpenRedmine/issues
Auto Name:OpenRedmine
Summary:Redmine client
Description:
Connect to Redmine servers (running 1.2 or later) and access the Wiki, Issue
Tracker, Source code, etc. easily.
Features:
* View issues offline
* Get API key from web site
* Filter (by Status/Tracker/Category/Priority)
* Download files
* Wiki
* Connect to unsafe SSL sites powered by transdroid
[https://raw.github.com/indication/OpenRedmine/development/doc/VERSION.en.md Release history]
.
Repo Type:git
Repo:https://github.com/indication/OpenRedmine.git
Build:3.6,38
commit=v3.6
subdir=OpenRedmine
gradle=yes
Build:3.7,39
commit=v3.7
subdir=OpenRedmine
gradle=yes
Build:3.8,40
commit=v3.8
subdir=OpenRedmine
gradle=yes
Build:3.9,41
commit=v3.9
subdir=OpenRedmine
gradle=yes
Build:3.10,42
commit=v3.10
subdir=OpenRedmine
gradle=yes
Auto Update Mode:Version v%v
Update Check Mode:Tags
Current Version:3.10
Current Version Code:42
| Update OpenRedmine to 3.10 (42) | Update OpenRedmine to 3.10 (42)
| Text | agpl-3.0 | f-droid/fdroiddata,f-droid/fdroiddata,f-droid/fdroid-data | text | ## Code Before:
Categories:Internet
License:GPLv2+
Web Site:http://indication.github.io/OpenRedmine
Source Code:https://github.com/indication/OpenRedmine
Issue Tracker:https://github.com/indication/OpenRedmine/issues
Auto Name:OpenRedmine
Summary:Redmine client
Description:
Connect to Redmine servers (running 1.2 or later) and access the Wiki, Issue
Tracker, Source code, etc. easily.
Features:
* View issues offline
* Get API key from web site
* Filter (by Status/Tracker/Category/Priority)
* Download files
* Wiki
* Connect to unsafe SSL sites powered by transdroid
[https://raw.github.com/indication/OpenRedmine/development/doc/VERSION.en.md Release history]
.
Repo Type:git
Repo:https://github.com/indication/OpenRedmine.git
Build:3.6,38
commit=v3.6
subdir=OpenRedmine
gradle=yes
Build:3.7,39
commit=v3.7
subdir=OpenRedmine
gradle=yes
Build:3.8,40
commit=v3.8
subdir=OpenRedmine
gradle=yes
Build:3.9,41
commit=v3.9
subdir=OpenRedmine
gradle=yes
Auto Update Mode:Version v%v
Update Check Mode:Tags
Current Version:3.9
Current Version Code:41
## Instruction:
Update OpenRedmine to 3.10 (42)
## Code After:
Categories:Internet
License:GPLv2+
Web Site:http://indication.github.io/OpenRedmine
Source Code:https://github.com/indication/OpenRedmine
Issue Tracker:https://github.com/indication/OpenRedmine/issues
Auto Name:OpenRedmine
Summary:Redmine client
Description:
Connect to Redmine servers (running 1.2 or later) and access the Wiki, Issue
Tracker, Source code, etc. easily.
Features:
* View issues offline
* Get API key from web site
* Filter (by Status/Tracker/Category/Priority)
* Download files
* Wiki
* Connect to unsafe SSL sites powered by transdroid
[https://raw.github.com/indication/OpenRedmine/development/doc/VERSION.en.md Release history]
.
Repo Type:git
Repo:https://github.com/indication/OpenRedmine.git
Build:3.6,38
commit=v3.6
subdir=OpenRedmine
gradle=yes
Build:3.7,39
commit=v3.7
subdir=OpenRedmine
gradle=yes
Build:3.8,40
commit=v3.8
subdir=OpenRedmine
gradle=yes
Build:3.9,41
commit=v3.9
subdir=OpenRedmine
gradle=yes
Build:3.10,42
commit=v3.10
subdir=OpenRedmine
gradle=yes
Auto Update Mode:Version v%v
Update Check Mode:Tags
Current Version:3.10
Current Version Code:42
| Categories:Internet
License:GPLv2+
Web Site:http://indication.github.io/OpenRedmine
Source Code:https://github.com/indication/OpenRedmine
Issue Tracker:https://github.com/indication/OpenRedmine/issues
Auto Name:OpenRedmine
Summary:Redmine client
Description:
Connect to Redmine servers (running 1.2 or later) and access the Wiki, Issue
Tracker, Source code, etc. easily.
Features:
* View issues offline
* Get API key from web site
* Filter (by Status/Tracker/Category/Priority)
* Download files
* Wiki
* Connect to unsafe SSL sites powered by transdroid
[https://raw.github.com/indication/OpenRedmine/development/doc/VERSION.en.md Release history]
.
Repo Type:git
Repo:https://github.com/indication/OpenRedmine.git
Build:3.6,38
commit=v3.6
subdir=OpenRedmine
gradle=yes
Build:3.7,39
commit=v3.7
subdir=OpenRedmine
gradle=yes
Build:3.8,40
commit=v3.8
subdir=OpenRedmine
gradle=yes
Build:3.9,41
commit=v3.9
subdir=OpenRedmine
gradle=yes
+ Build:3.10,42
+ commit=v3.10
+ subdir=OpenRedmine
+ gradle=yes
+
Auto Update Mode:Version v%v
Update Check Mode:Tags
- Current Version:3.9
? ^
+ Current Version:3.10
? ^^
- Current Version Code:41
? ^
+ Current Version Code:42
? ^
| 9 | 0.176471 | 7 | 2 |
382affc198c59122f549ef1e9b1e4d4206323481 | core/src/main/kotlin/com/github/ktoolz/model/namespace.kt | core/src/main/kotlin/com/github/ktoolz/model/namespace.kt | package com.github.ktoolz.model
import javaslang.collection.List
import java.io.File
/**
* Store a search result
*/
data class SearchResult(val score: Int, val file: File) {
val filename: String by lazy { file.name }
}
data class SearchQuery(val term: String, val contexts: List<String>, val directories: List<String>) | package com.github.ktoolz.model
import javaslang.collection.List
import java.io.File
/**
* Store a search result
*/
data class SearchResult(val score: Int, val file: File) {
val filename: String by lazy { file.name }
}
data class FilterQuery(val name: String, val negated: Boolean = false)
data class SearchQuery(val term: String, val contexts: List<FilterQuery>, val directories: List<String>) | Convert filter query from String to dedicated object (FilterQuery). A filter query can be negated or not | Convert filter query from String to dedicated object (FilterQuery). A filter query can be negated or not
| Kotlin | mit | ktoolz/file-finder | kotlin | ## Code Before:
package com.github.ktoolz.model
import javaslang.collection.List
import java.io.File
/**
* Store a search result
*/
data class SearchResult(val score: Int, val file: File) {
val filename: String by lazy { file.name }
}
data class SearchQuery(val term: String, val contexts: List<String>, val directories: List<String>)
## Instruction:
Convert filter query from String to dedicated object (FilterQuery). A filter query can be negated or not
## Code After:
package com.github.ktoolz.model
import javaslang.collection.List
import java.io.File
/**
* Store a search result
*/
data class SearchResult(val score: Int, val file: File) {
val filename: String by lazy { file.name }
}
data class FilterQuery(val name: String, val negated: Boolean = false)
data class SearchQuery(val term: String, val contexts: List<FilterQuery>, val directories: List<String>) | package com.github.ktoolz.model
import javaslang.collection.List
import java.io.File
/**
* Store a search result
*/
data class SearchResult(val score: Int, val file: File) {
val filename: String by lazy { file.name }
}
+ data class FilterQuery(val name: String, val negated: Boolean = false)
+
- data class SearchQuery(val term: String, val contexts: List<String>, val directories: List<String>)
? ^ ^^^
+ data class SearchQuery(val term: String, val contexts: List<FilterQuery>, val directories: List<String>)
? ^^^ + ^^^^^
| 4 | 0.307692 | 3 | 1 |
427558fbdc6d6c7e61dd9efacd68df9f26e72a1d | src/javascript/binary/components/trackjs_onerror.js | src/javascript/binary/components/trackjs_onerror.js | window._trackJs = {
onError: function(payload, error) {
// ignore an error caused by DealPly (http://www.dealply.com/) chrome extension
if (payload.message.indexOf("DealPly") > 0) {
return false;
}
if (payload.message.indexOf("NSA") > 0) {
payload.message = "Redacted!";
}
// remove 401 network events, they happen all the time for us!
payload.network = payload.network.filter(function(item) {
if (item.statusCode !== 401) {
return true;
} else {
return false;
}
});
return true;
}
};
| window._trackJs = {
onError: function(payload, error) {
// ignore an error caused by DealPly (http://www.dealply.com/) chrome extension
if (payload.message.indexOf("DealPly") > 0) {
return false;
}
payload.network = payload.network.filter(function(item) {
// ignore random errors from Intercom
if (item.statusCode === 403 && payload.message.indexOf("intercom") > 0) {
return false;
}
return true;
});
return true;
}
};
| Add ignore for 303 Intercom error not related to us | Add ignore for 303 Intercom error not related to us
| JavaScript | apache-2.0 | massihx/binary-static,brodiecapel16/binary-static,borisyankov/binary-static,borisyankov/binary-static,einhverfr/binary-static,einhverfr/binary-static,animeshsaxena/binary-static,brodiecapel16/binary-static,animeshsaxena/binary-static,brodiecapel16/binary-static,tfoertsch/binary-static,brodiecapel16/binary-static,massihx/binary-static,animeshsaxena/binary-static,massihx/binary-static,tfoertsch/binary-static,einhverfr/binary-static,junbon/binary-static-www2,massihx/binary-static,borisyankov/binary-static,animeshsaxena/binary-static,einhverfr/binary-static,junbon/binary-static-www2,borisyankov/binary-static | javascript | ## Code Before:
window._trackJs = {
onError: function(payload, error) {
// ignore an error caused by DealPly (http://www.dealply.com/) chrome extension
if (payload.message.indexOf("DealPly") > 0) {
return false;
}
if (payload.message.indexOf("NSA") > 0) {
payload.message = "Redacted!";
}
// remove 401 network events, they happen all the time for us!
payload.network = payload.network.filter(function(item) {
if (item.statusCode !== 401) {
return true;
} else {
return false;
}
});
return true;
}
};
## Instruction:
Add ignore for 303 Intercom error not related to us
## Code After:
window._trackJs = {
onError: function(payload, error) {
// ignore an error caused by DealPly (http://www.dealply.com/) chrome extension
if (payload.message.indexOf("DealPly") > 0) {
return false;
}
payload.network = payload.network.filter(function(item) {
// ignore random errors from Intercom
if (item.statusCode === 403 && payload.message.indexOf("intercom") > 0) {
return false;
}
return true;
});
return true;
}
};
| window._trackJs = {
onError: function(payload, error) {
// ignore an error caused by DealPly (http://www.dealply.com/) chrome extension
if (payload.message.indexOf("DealPly") > 0) {
return false;
}
+ payload.network = payload.network.filter(function(item) {
- if (payload.message.indexOf("NSA") > 0) {
- payload.message = "Redacted!";
- }
+ // ignore random errors from Intercom
+ if (item.statusCode === 403 && payload.message.indexOf("intercom") > 0) {
- // remove 401 network events, they happen all the time for us!
- payload.network = payload.network.filter(function(item) {
- if (item.statusCode !== 401) {
- return true;
- } else {
return false;
}
+
+ return true;
});
return true;
}
}; | 13 | 0.541667 | 5 | 8 |
650c8cfd8a09b109a8324956a4df1756ce4db566 | src/ggrc/assets/mustache/base_objects/test_plan.mustache | src/ggrc/assets/mustache/base_objects/test_plan.mustache | {{!
Copyright (C) 2015 Google Inc., authors, and contributors <see AUTHORS file>
Licensed under http://www.apache.org/licenses/LICENSE-2.0 <see LICENSE file>
Created By: brad@reciprocitylabs.com
Maintained By: brad@reciprocitylabs.com
}}
{{#instance}}
<div class="row-fluid wrap-row">
<div class="span12">
<h6>Test Plan</h6>
<div class="rtf-block">
{{{firstnonempty test_plan '<span class="empty-message">None</span>'}}}
</div>
</div>
</div>
{{/instance}}
| {{!
Copyright (C) 2015 Google Inc., authors, and contributors <see AUTHORS file>
Licensed under http://www.apache.org/licenses/LICENSE-2.0 <see LICENSE file>
Created By: brad@reciprocitylabs.com
Maintained By: brad@reciprocitylabs.com
}}
{{#instance}}
{{#if test_plan}}
<div class="row-fluid wrap-row">
<div class="span12">
<h6>Test Plan</h6>
<div class="rtf-block">
{{{firstnonempty test_plan '<span class="empty-message">None</span>'}}}
</div>
</div>
</div>
{{/if}}
{{/instance}}
| Hide test plan if no value | Hide test plan if no value
| HTML+Django | apache-2.0 | AleksNeStu/ggrc-core,NejcZupec/ggrc-core,andrei-karalionak/ggrc-core,kr41/ggrc-core,VinnieJohns/ggrc-core,plamut/ggrc-core,kr41/ggrc-core,AleksNeStu/ggrc-core,selahssea/ggrc-core,j0gurt/ggrc-core,j0gurt/ggrc-core,AleksNeStu/ggrc-core,edofic/ggrc-core,josthkko/ggrc-core,andrei-karalionak/ggrc-core,NejcZupec/ggrc-core,kr41/ggrc-core,plamut/ggrc-core,selahssea/ggrc-core,VinnieJohns/ggrc-core,selahssea/ggrc-core,josthkko/ggrc-core,j0gurt/ggrc-core,selahssea/ggrc-core,prasannav7/ggrc-core,edofic/ggrc-core,prasannav7/ggrc-core,edofic/ggrc-core,VinnieJohns/ggrc-core,josthkko/ggrc-core,NejcZupec/ggrc-core,VinnieJohns/ggrc-core,andrei-karalionak/ggrc-core,edofic/ggrc-core,prasannav7/ggrc-core,AleksNeStu/ggrc-core,plamut/ggrc-core,prasannav7/ggrc-core,NejcZupec/ggrc-core,plamut/ggrc-core,josthkko/ggrc-core,j0gurt/ggrc-core,kr41/ggrc-core,andrei-karalionak/ggrc-core | html+django | ## Code Before:
{{!
Copyright (C) 2015 Google Inc., authors, and contributors <see AUTHORS file>
Licensed under http://www.apache.org/licenses/LICENSE-2.0 <see LICENSE file>
Created By: brad@reciprocitylabs.com
Maintained By: brad@reciprocitylabs.com
}}
{{#instance}}
<div class="row-fluid wrap-row">
<div class="span12">
<h6>Test Plan</h6>
<div class="rtf-block">
{{{firstnonempty test_plan '<span class="empty-message">None</span>'}}}
</div>
</div>
</div>
{{/instance}}
## Instruction:
Hide test plan if no value
## Code After:
{{!
Copyright (C) 2015 Google Inc., authors, and contributors <see AUTHORS file>
Licensed under http://www.apache.org/licenses/LICENSE-2.0 <see LICENSE file>
Created By: brad@reciprocitylabs.com
Maintained By: brad@reciprocitylabs.com
}}
{{#instance}}
{{#if test_plan}}
<div class="row-fluid wrap-row">
<div class="span12">
<h6>Test Plan</h6>
<div class="rtf-block">
{{{firstnonempty test_plan '<span class="empty-message">None</span>'}}}
</div>
</div>
</div>
{{/if}}
{{/instance}}
| {{!
Copyright (C) 2015 Google Inc., authors, and contributors <see AUTHORS file>
Licensed under http://www.apache.org/licenses/LICENSE-2.0 <see LICENSE file>
Created By: brad@reciprocitylabs.com
Maintained By: brad@reciprocitylabs.com
}}
{{#instance}}
-
+ {{#if test_plan}}
<div class="row-fluid wrap-row">
<div class="span12">
<h6>Test Plan</h6>
<div class="rtf-block">
{{{firstnonempty test_plan '<span class="empty-message">None</span>'}}}
</div>
</div>
</div>
-
+ {{/if}}
{{/instance}} | 4 | 0.210526 | 2 | 2 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.