commit stringlengths 40 40 | old_file stringlengths 4 184 | new_file stringlengths 4 184 | old_contents stringlengths 1 3.6k | new_contents stringlengths 5 3.38k | subject stringlengths 15 778 | message stringlengths 16 6.74k | lang stringclasses 201 values | license stringclasses 13 values | repos stringlengths 6 116k | config stringclasses 201 values | content stringlengths 137 7.24k | diff stringlengths 26 5.55k | diff_length int64 1 123 | relative_diff_length float64 0.01 89 | n_lines_added int64 0 108 | n_lines_deleted int64 0 106 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
879d43c7f6f78f66d9223e8126b0d502c207500d | CONTRIBUTING.md | CONTRIBUTING.md |
Before we can accept pull requests, you'll need you to sign the Polis Contributor Agreement (link coming soon)
|
Before we can accept pull requests, you'll need you to sign the Polis Contributor Agreement. https://pol.is/contrib
| Add link to contrib form | Add link to contrib form
| Markdown | bsd-3-clause | filiperibeiro77/polisClientAdmin,filiperibeiro77/polisClientAdmin | markdown | ## Code Before:
Before we can accept pull requests, you'll need you to sign the Polis Contributor Agreement (link coming soon)
## Instruction:
Add link to contrib form
## Code After:
Before we can accept pull requests, you'll need you to sign the Polis Contributor Agreement. https://pol.is/contrib
|
- Before we can accept pull requests, you'll need you to sign the Polis Contributor Agreement (link coming soon)
? ^ ^^^ ^ ^^^^^^^^
+ Before we can accept pull requests, you'll need you to sign the Polis Contributor Agreement. https://pol.is/contrib
? + ^^^^^^^^^^ + ^^ ^^^ ^
| 2 | 1 | 1 | 1 |
39244ae733b749e480b3e3813fb25c0289690d9a | Etc/versioninfo.json | Etc/versioninfo.json | {
"FixedFileInfo":
{
"FileFlagsMask": "3f",
"FileFlags ": "00",
"FileOS": "040004",
"FileType": "01",
"FileSubType": "00"
},
"StringFileInfo":
{
"Comments": "",
"CompanyName": "NYAOS.ORG",
"FileDescription": "Unixlike Commandline Shell",
"InternalName": "",
"LegalCopyright": "Copyright (C) 2014-2018 HAYAMA_Kaoru",
"LegalTrademarks": "",
"OriginalFilename": "NYAGOS.EXE",
"PrivateBuild": "",
"ProductName": "Nihongo Yet Another GOing Shell",
"SpecialBuild": ""
},
"VarFileInfo":
{
"Translation": {
"LangID": "0411",
"CharsetID": "04E4"
}
}
}
| {
"FixedFileInfo":
{
"FileFlagsMask": "3f",
"FileFlags ": "00",
"FileOS": "040004",
"FileType": "01",
"FileSubType": "00"
},
"StringFileInfo":
{
"Comments": "",
"CompanyName": "NYAOS.ORG",
"FileDescription": "Nihongo Yet Another GOing Shell",
"InternalName": "",
"LegalCopyright": "Copyright (C) 2014-2018 HAYAMA_Kaoru",
"LegalTrademarks": "",
"OriginalFilename": "NYAGOS.EXE",
"PrivateBuild": "",
"ProductName": "Nihongo Yet Another GOing Shell",
"SpecialBuild": ""
},
"VarFileInfo":
{
"Translation": {
"LangID": "0411",
"CharsetID": "04E4"
}
}
}
| Change FileDescription to `Nihongo Yet Another GOing Shell` | Change FileDescription to `Nihongo Yet Another GOing Shell`
from `Unixlike Commandline Shell` because it is shown in Task Manager
| JSON | bsd-3-clause | nocd5/nyagos,zetamatta/nyagos,tsuyoshicho/nyagos | json | ## Code Before:
{
"FixedFileInfo":
{
"FileFlagsMask": "3f",
"FileFlags ": "00",
"FileOS": "040004",
"FileType": "01",
"FileSubType": "00"
},
"StringFileInfo":
{
"Comments": "",
"CompanyName": "NYAOS.ORG",
"FileDescription": "Unixlike Commandline Shell",
"InternalName": "",
"LegalCopyright": "Copyright (C) 2014-2018 HAYAMA_Kaoru",
"LegalTrademarks": "",
"OriginalFilename": "NYAGOS.EXE",
"PrivateBuild": "",
"ProductName": "Nihongo Yet Another GOing Shell",
"SpecialBuild": ""
},
"VarFileInfo":
{
"Translation": {
"LangID": "0411",
"CharsetID": "04E4"
}
}
}
## Instruction:
Change FileDescription to `Nihongo Yet Another GOing Shell`
from `Unixlike Commandline Shell` because it is shown in Task Manager
## Code After:
{
"FixedFileInfo":
{
"FileFlagsMask": "3f",
"FileFlags ": "00",
"FileOS": "040004",
"FileType": "01",
"FileSubType": "00"
},
"StringFileInfo":
{
"Comments": "",
"CompanyName": "NYAOS.ORG",
"FileDescription": "Nihongo Yet Another GOing Shell",
"InternalName": "",
"LegalCopyright": "Copyright (C) 2014-2018 HAYAMA_Kaoru",
"LegalTrademarks": "",
"OriginalFilename": "NYAGOS.EXE",
"PrivateBuild": "",
"ProductName": "Nihongo Yet Another GOing Shell",
"SpecialBuild": ""
},
"VarFileInfo":
{
"Translation": {
"LangID": "0411",
"CharsetID": "04E4"
}
}
}
| {
"FixedFileInfo":
{
"FileFlagsMask": "3f",
"FileFlags ": "00",
"FileOS": "040004",
"FileType": "01",
"FileSubType": "00"
},
"StringFileInfo":
{
"Comments": "",
"CompanyName": "NYAOS.ORG",
- "FileDescription": "Unixlike Commandline Shell",
+ "FileDescription": "Nihongo Yet Another GOing Shell",
"InternalName": "",
"LegalCopyright": "Copyright (C) 2014-2018 HAYAMA_Kaoru",
"LegalTrademarks": "",
"OriginalFilename": "NYAGOS.EXE",
"PrivateBuild": "",
"ProductName": "Nihongo Yet Another GOing Shell",
"SpecialBuild": ""
},
"VarFileInfo":
{
"Translation": {
"LangID": "0411",
"CharsetID": "04E4"
}
}
} | 2 | 0.066667 | 1 | 1 |
f6d44e3a256031b41650f1cf3e889485f2e44eeb | lib/metamorpher/refactorer.rb | lib/metamorpher/refactorer.rb | require "metamorpher/refactorer/merger"
require "metamorpher/refactorer/replacement"
require "metamorpher/builder"
require "metamorpher/rewriter/rule"
require "metamorpher/drivers/ruby"
module Metamorpher
module Refactorer
def refactor(src, &block)
literal = driver.parse(src)
replacements = reduce_to_replacements(literal)
Merger.new(src).merge(*replacements, &block)
end
def builder
@builder ||= Builder.new
end
def driver
@driver ||= Metamorpher::Drivers::Ruby.new
end
private
def reduce_to_replacements(literal)
[].tap do |replacements|
rule.reduce(literal) do |original, rewritten|
position = driver.source_location_for(original)
new_code = driver.unparse(rewritten)
replacements << Replacement.new(position, new_code)
end
end
end
def rule
@rule ||= Rewriter::Rule.new(pattern: pattern, replacement: replacement)
end
end
end
| require "metamorpher/refactorer/merger"
require "metamorpher/refactorer/replacement"
require "metamorpher/builder"
require "metamorpher/rewriter/rule"
require "metamorpher/drivers/ruby"
module Metamorpher
module Refactorer
def refactor(src, &block)
literal = driver.parse(src)
replacements = reduce_to_replacements(literal)
Merger.new(src).merge(*replacements, &block)
end
def refactor_file(path, &block)
changes = []
refactored = refactor(File.read(path)) { |change| changes << change }
block.call(path, refactored, changes) if block
refactored
end
def refactor_files(paths, &block)
paths.reduce({}) do |result, path|
result[path] = refactor_file(path, &block)
result
end
end
def builder
@builder ||= Builder.new
end
def driver
@driver ||= Metamorpher::Drivers::Ruby.new
end
private
def reduce_to_replacements(literal)
[].tap do |replacements|
rule.reduce(literal) do |original, rewritten|
position = driver.source_location_for(original)
new_code = driver.unparse(rewritten)
replacements << Replacement.new(position, new_code)
end
end
end
def rule
@rule ||= Rewriter::Rule.new(pattern: pattern, replacement: replacement)
end
end
end
| Add support for refactoring code contained in files. | Add support for refactoring code contained in files. | Ruby | mit | mutiny/metamorpher | ruby | ## Code Before:
require "metamorpher/refactorer/merger"
require "metamorpher/refactorer/replacement"
require "metamorpher/builder"
require "metamorpher/rewriter/rule"
require "metamorpher/drivers/ruby"
module Metamorpher
module Refactorer
def refactor(src, &block)
literal = driver.parse(src)
replacements = reduce_to_replacements(literal)
Merger.new(src).merge(*replacements, &block)
end
def builder
@builder ||= Builder.new
end
def driver
@driver ||= Metamorpher::Drivers::Ruby.new
end
private
def reduce_to_replacements(literal)
[].tap do |replacements|
rule.reduce(literal) do |original, rewritten|
position = driver.source_location_for(original)
new_code = driver.unparse(rewritten)
replacements << Replacement.new(position, new_code)
end
end
end
def rule
@rule ||= Rewriter::Rule.new(pattern: pattern, replacement: replacement)
end
end
end
## Instruction:
Add support for refactoring code contained in files.
## Code After:
require "metamorpher/refactorer/merger"
require "metamorpher/refactorer/replacement"
require "metamorpher/builder"
require "metamorpher/rewriter/rule"
require "metamorpher/drivers/ruby"
module Metamorpher
module Refactorer
def refactor(src, &block)
literal = driver.parse(src)
replacements = reduce_to_replacements(literal)
Merger.new(src).merge(*replacements, &block)
end
def refactor_file(path, &block)
changes = []
refactored = refactor(File.read(path)) { |change| changes << change }
block.call(path, refactored, changes) if block
refactored
end
def refactor_files(paths, &block)
paths.reduce({}) do |result, path|
result[path] = refactor_file(path, &block)
result
end
end
def builder
@builder ||= Builder.new
end
def driver
@driver ||= Metamorpher::Drivers::Ruby.new
end
private
def reduce_to_replacements(literal)
[].tap do |replacements|
rule.reduce(literal) do |original, rewritten|
position = driver.source_location_for(original)
new_code = driver.unparse(rewritten)
replacements << Replacement.new(position, new_code)
end
end
end
def rule
@rule ||= Rewriter::Rule.new(pattern: pattern, replacement: replacement)
end
end
end
| require "metamorpher/refactorer/merger"
require "metamorpher/refactorer/replacement"
require "metamorpher/builder"
require "metamorpher/rewriter/rule"
require "metamorpher/drivers/ruby"
module Metamorpher
module Refactorer
def refactor(src, &block)
literal = driver.parse(src)
replacements = reduce_to_replacements(literal)
Merger.new(src).merge(*replacements, &block)
+ end
+
+ def refactor_file(path, &block)
+ changes = []
+ refactored = refactor(File.read(path)) { |change| changes << change }
+ block.call(path, refactored, changes) if block
+ refactored
+ end
+
+ def refactor_files(paths, &block)
+ paths.reduce({}) do |result, path|
+ result[path] = refactor_file(path, &block)
+ result
+ end
end
def builder
@builder ||= Builder.new
end
def driver
@driver ||= Metamorpher::Drivers::Ruby.new
end
private
def reduce_to_replacements(literal)
[].tap do |replacements|
rule.reduce(literal) do |original, rewritten|
position = driver.source_location_for(original)
new_code = driver.unparse(rewritten)
replacements << Replacement.new(position, new_code)
end
end
end
def rule
@rule ||= Rewriter::Rule.new(pattern: pattern, replacement: replacement)
end
end
end | 14 | 0.358974 | 14 | 0 |
54691f9be052e5564ca0e5c6a503e641ea3142e1 | keras/layers/normalization.py | keras/layers/normalization.py | from ..layers.core import Layer
from ..utils.theano_utils import shared_zeros
from .. import initializations
class BatchNormalization(Layer):
'''
Reference:
Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift
http://arxiv.org/pdf/1502.03167v3.pdf
'''
def __init__(self, input_shape, epsilon=1e-6, weights=None):
self.init = initializations.get("uniform")
self.input_shape = input_shape
self.epsilon = epsilon
self.gamma = self.init((self.input_shape))
self.beta = shared_zeros(self.input_shape)
self.params = [self.gamma, self.beta]
if weights is not None:
self.set_weights(weights)
def output(self, train):
X = self.get_input(train)
X_normed = (X - X.mean(keepdims=True)) / (X.std(keepdims=True) + self.epsilon)
out = self.gamma * X_normed + self.beta
return out
def get_config(self):
return {"name":self.__class__.__name__,
"input_shape":self.input_shape,
"epsilon":self.epsilon} | from ..layers.core import Layer
from ..utils.theano_utils import shared_zeros
from .. import initializations
import theano.tensor as T
class BatchNormalization(Layer):
'''
Reference:
Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift
http://arxiv.org/pdf/1502.03167v3.pdf
mode: 0 -> featurewise normalization
1 -> samplewise normalization (may sometimes outperform featurewise mode)
'''
def __init__(self, input_shape, epsilon=1e-6, mode=0, weights=None):
self.init = initializations.get("uniform")
self.input_shape = input_shape
self.epsilon = epsilon
self.mode = mode
self.gamma = self.init((self.input_shape))
self.beta = shared_zeros(self.input_shape)
self.params = [self.gamma, self.beta]
if weights is not None:
self.set_weights(weights)
def output(self, train):
X = self.get_input(train)
if self.mode == 0:
m = X.mean(axis=0)
# manual computation of std to prevent NaNs
std = T.mean((X-m)**2 + self.epsilon, axis=0) ** 0.5
X_normed = (X - m) / (std + self.epsilon)
elif self.mode == 1:
m = X.mean(axis=-1, keepdims=True)
std = X.std(axis=-1, keepdims=True)
X_normed = (X - m) / (std + self.epsilon)
out = self.gamma * X_normed + self.beta
return out
def get_config(self):
return {"name":self.__class__.__name__,
"input_shape":self.input_shape,
"epsilon":self.epsilon,
"mode":self.mode}
| Add modes to BatchNormalization, fix BN issues | Add modes to BatchNormalization, fix BN issues | Python | mit | yingzha/keras,imcomking/Convolutional-GRU-keras-extension-,jayhetee/keras,why11002526/keras,marchick209/keras,relh/keras,mikekestemont/keras,florentchandelier/keras,jbolinge/keras,tencrance/keras,meanmee/keras,kuza55/keras,nehz/keras,keras-team/keras,EderSantana/keras,kemaswill/keras,abayowbo/keras,dhruvparamhans/keras,cheng6076/keras,jasonyaw/keras,3dconv/keras,ogrisel/keras,marcelo-amancio/keras,jalexvig/keras,johmathe/keras,ekamioka/keras,stonebig/keras,Smerity/keras,dribnet/keras,Aureliu/keras,nt/keras,nebw/keras,ashhher3/keras,ypkang/keras,fmacias64/keras,gamer13/keras,harshhemani/keras,LIBOTAO/keras,rudaoshi/keras,danielforsyth/keras,hhaoyan/keras,jiumem/keras,Yingmin-Li/keras,iScienceLuvr/keras,dxj19831029/keras,aleju/keras,nzer0/keras,bottler/keras,printedheart/keras,wxs/keras,jfsantos/keras,ml-lab/keras,navyjeff/keras,wubr2000/keras,rlkelly/keras,zxytim/keras,gavinmh/keras,keras-team/keras,brainwater/keras,pthaike/keras,DeepGnosis/keras,pjadzinsky/keras,jonberliner/keras,rodrigob/keras,zxsted/keras,vseledkin/keras,asampat3090/keras,OlafLee/keras,daviddiazvico/keras,Cadene/keras,llcao/keras,sjuvekar/keras,zhmz90/keras,stephenbalaban/keras,dolaameng/keras,kfoss/keras,jimgoo/keras,iamtrask/keras,xiaoda99/keras,keskarnitish/keras,untom/keras,happyboy310/keras,bboalimoe/keras,DLlearn/keras,JasonTam/keras,kod3r/keras,MagicSen/keras,chenych11/keras,jmportilla/keras,jslhs/keras,xurantju/keras,saurav111/keras,zhangxujinsh/keras,amy12xx/keras,cmyr/keras,ledbetdr/keras,eulerreich/keras,cvfish/keras,jhauswald/keras | python | ## Code Before:
from ..layers.core import Layer
from ..utils.theano_utils import shared_zeros
from .. import initializations
class BatchNormalization(Layer):
'''
Reference:
Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift
http://arxiv.org/pdf/1502.03167v3.pdf
'''
def __init__(self, input_shape, epsilon=1e-6, weights=None):
self.init = initializations.get("uniform")
self.input_shape = input_shape
self.epsilon = epsilon
self.gamma = self.init((self.input_shape))
self.beta = shared_zeros(self.input_shape)
self.params = [self.gamma, self.beta]
if weights is not None:
self.set_weights(weights)
def output(self, train):
X = self.get_input(train)
X_normed = (X - X.mean(keepdims=True)) / (X.std(keepdims=True) + self.epsilon)
out = self.gamma * X_normed + self.beta
return out
def get_config(self):
return {"name":self.__class__.__name__,
"input_shape":self.input_shape,
"epsilon":self.epsilon}
## Instruction:
Add modes to BatchNormalization, fix BN issues
## Code After:
from ..layers.core import Layer
from ..utils.theano_utils import shared_zeros
from .. import initializations
import theano.tensor as T
class BatchNormalization(Layer):
'''
Reference:
Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift
http://arxiv.org/pdf/1502.03167v3.pdf
mode: 0 -> featurewise normalization
1 -> samplewise normalization (may sometimes outperform featurewise mode)
'''
def __init__(self, input_shape, epsilon=1e-6, mode=0, weights=None):
self.init = initializations.get("uniform")
self.input_shape = input_shape
self.epsilon = epsilon
self.mode = mode
self.gamma = self.init((self.input_shape))
self.beta = shared_zeros(self.input_shape)
self.params = [self.gamma, self.beta]
if weights is not None:
self.set_weights(weights)
def output(self, train):
X = self.get_input(train)
if self.mode == 0:
m = X.mean(axis=0)
# manual computation of std to prevent NaNs
std = T.mean((X-m)**2 + self.epsilon, axis=0) ** 0.5
X_normed = (X - m) / (std + self.epsilon)
elif self.mode == 1:
m = X.mean(axis=-1, keepdims=True)
std = X.std(axis=-1, keepdims=True)
X_normed = (X - m) / (std + self.epsilon)
out = self.gamma * X_normed + self.beta
return out
def get_config(self):
return {"name":self.__class__.__name__,
"input_shape":self.input_shape,
"epsilon":self.epsilon,
"mode":self.mode}
| from ..layers.core import Layer
from ..utils.theano_utils import shared_zeros
from .. import initializations
+
+ import theano.tensor as T
class BatchNormalization(Layer):
'''
Reference:
Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift
http://arxiv.org/pdf/1502.03167v3.pdf
+
+ mode: 0 -> featurewise normalization
+ 1 -> samplewise normalization (may sometimes outperform featurewise mode)
'''
- def __init__(self, input_shape, epsilon=1e-6, weights=None):
+ def __init__(self, input_shape, epsilon=1e-6, mode=0, weights=None):
? ++++++++
self.init = initializations.get("uniform")
self.input_shape = input_shape
self.epsilon = epsilon
+ self.mode = mode
self.gamma = self.init((self.input_shape))
self.beta = shared_zeros(self.input_shape)
self.params = [self.gamma, self.beta]
if weights is not None:
self.set_weights(weights)
def output(self, train):
X = self.get_input(train)
- X_normed = (X - X.mean(keepdims=True)) / (X.std(keepdims=True) + self.epsilon)
+
+ if self.mode == 0:
+ m = X.mean(axis=0)
+ # manual computation of std to prevent NaNs
+ std = T.mean((X-m)**2 + self.epsilon, axis=0) ** 0.5
+ X_normed = (X - m) / (std + self.epsilon)
+
+ elif self.mode == 1:
+ m = X.mean(axis=-1, keepdims=True)
+ std = X.std(axis=-1, keepdims=True)
+ X_normed = (X - m) / (std + self.epsilon)
+
out = self.gamma * X_normed + self.beta
return out
def get_config(self):
return {"name":self.__class__.__name__,
"input_shape":self.input_shape,
- "epsilon":self.epsilon}
? ^
+ "epsilon":self.epsilon,
? ^
+ "mode":self.mode} | 24 | 0.75 | 21 | 3 |
be078ecf5994ddd4a2c66b7d05998d6a9953a88b | packages/most-nth/src/most-nth.d.ts | packages/most-nth/src/most-nth.d.ts | import { Stream } from 'most';
export function nth<T>(index: number): (stream: Stream<T>) => Promise<T>;
export function first<T>(stream: Stream<T>): Promise<T>;
export function last<T>(stream: Stream<T>): Promise<T>;
| import { Stream } from 'most';
export function nth<T>(index: number): (stream: Stream<T>) => Promise<T | undefined>;
export function first<T>(stream: Stream<T>): Promise<T | undefined>;
export function last<T>(stream: Stream<T>): Promise<T | undefined>;
| Fix the resolve type of most-nth functions | Fix the resolve type of most-nth functions | TypeScript | bsd-3-clause | craft-ai/most-utils,craft-ai/most-utils | typescript | ## Code Before:
import { Stream } from 'most';
export function nth<T>(index: number): (stream: Stream<T>) => Promise<T>;
export function first<T>(stream: Stream<T>): Promise<T>;
export function last<T>(stream: Stream<T>): Promise<T>;
## Instruction:
Fix the resolve type of most-nth functions
## Code After:
import { Stream } from 'most';
export function nth<T>(index: number): (stream: Stream<T>) => Promise<T | undefined>;
export function first<T>(stream: Stream<T>): Promise<T | undefined>;
export function last<T>(stream: Stream<T>): Promise<T | undefined>;
| import { Stream } from 'most';
- export function nth<T>(index: number): (stream: Stream<T>) => Promise<T>;
+ export function nth<T>(index: number): (stream: Stream<T>) => Promise<T | undefined>;
? ++++++++++++
- export function first<T>(stream: Stream<T>): Promise<T>;
+ export function first<T>(stream: Stream<T>): Promise<T | undefined>;
? ++++++++++++
- export function last<T>(stream: Stream<T>): Promise<T>;
+ export function last<T>(stream: Stream<T>): Promise<T | undefined>;
? ++++++++++++
| 6 | 1.2 | 3 | 3 |
7b2a5d48333d02081b1ebb37fc2820a215f0d2a6 | resources/views/url/store.blade.php | resources/views/url/store.blade.php | $('.form-group-url').removeClass('has-error')
.find('.alert')
.fadeOut();
$('#input-url').val('{{ $shortUrl }}').select();
{{--$('#btn-copy-url').focus();--}}
{{--$('#input-original-url').val('').focus();--}} | $('.form-group-url').removeClass('has-error')
.find('.alert')
.fadeOut();
$('#input-original-url').val('');
$('#input-url').val('{{ $shortUrl }}').select();
{{--$('#btn-copy-url').focus();--}}
{{--$('#input-original-url').val('').focus();--}} | Clean input field after url generation | Clean input field after url generation
| PHP | mit | jgrossi/petty,jgrossi/petty | php | ## Code Before:
$('.form-group-url').removeClass('has-error')
.find('.alert')
.fadeOut();
$('#input-url').val('{{ $shortUrl }}').select();
{{--$('#btn-copy-url').focus();--}}
{{--$('#input-original-url').val('').focus();--}}
## Instruction:
Clean input field after url generation
## Code After:
$('.form-group-url').removeClass('has-error')
.find('.alert')
.fadeOut();
$('#input-original-url').val('');
$('#input-url').val('{{ $shortUrl }}').select();
{{--$('#btn-copy-url').focus();--}}
{{--$('#input-original-url').val('').focus();--}} | $('.form-group-url').removeClass('has-error')
.find('.alert')
.fadeOut();
+ $('#input-original-url').val('');
$('#input-url').val('{{ $shortUrl }}').select();
{{--$('#btn-copy-url').focus();--}}
{{--$('#input-original-url').val('').focus();--}} | 1 | 0.142857 | 1 | 0 |
f5ef426751594eeb4a6839c6883090b8ebcda544 | src/lib/components/emitter-mixin.js | src/lib/components/emitter-mixin.js | 'use strict';
var glimpse = require('glimpse');
module.exports = {
addListener: function (event, listener) {
if (!this.events) {
this.events = {};
}
this.events[event] = glimpse.on(event, listener);
},
componentWillUnmount: function () {
if (this.events) {
for (var key in this.events) {
glimpse.off(this.events[key]);
}
this.events = null;
}
}
};
| 'use strict';
var glimpse = require('glimpse');
module.exports = {
addListener: function (event, listener) {
if (!this.events) {
this.events = {};
}
this.events[event] = glimpse.on(event, listener);
},
componentWillUnmount: function () {
if (this.events) {
for (var key in this.events) {
glimpse.off(key);
}
this.events = null;
}
}
};
| Fix bug in how events are unloaded | Fix bug in how events are unloaded
| JavaScript | unknown | Glimpse/Glimpse.Client.Prototype,Glimpse/Glimpse.Client.Prototype,avanderhoorn/Glimpse.Client.Prototype,Glimpse/Glimpse.Client.Prototype,avanderhoorn/Glimpse.Client.Prototype | javascript | ## Code Before:
'use strict';
var glimpse = require('glimpse');
module.exports = {
addListener: function (event, listener) {
if (!this.events) {
this.events = {};
}
this.events[event] = glimpse.on(event, listener);
},
componentWillUnmount: function () {
if (this.events) {
for (var key in this.events) {
glimpse.off(this.events[key]);
}
this.events = null;
}
}
};
## Instruction:
Fix bug in how events are unloaded
## Code After:
'use strict';
var glimpse = require('glimpse');
module.exports = {
addListener: function (event, listener) {
if (!this.events) {
this.events = {};
}
this.events[event] = glimpse.on(event, listener);
},
componentWillUnmount: function () {
if (this.events) {
for (var key in this.events) {
glimpse.off(key);
}
this.events = null;
}
}
};
| 'use strict';
var glimpse = require('glimpse');
module.exports = {
addListener: function (event, listener) {
if (!this.events) {
this.events = {};
}
this.events[event] = glimpse.on(event, listener);
},
componentWillUnmount: function () {
if (this.events) {
for (var key in this.events) {
- glimpse.off(this.events[key]);
? ------------ -
+ glimpse.off(key);
}
this.events = null;
}
}
}; | 2 | 0.090909 | 1 | 1 |
94db48f187d45dd674fc2d744a9822ec0979daca | recipes/decasu/meta.yaml | recipes/decasu/meta.yaml | {% set name = "decasu" %}
{% set version = "0.7.1" %}
{% set sha256 = "d5558cd419c8d46bdc958064cb97f963d1ea793866414c025906ec15033512ed" %}
package:
name: {{ name|lower }}
version: {{ version }}
source:
url: https://github.com/erykoff/{{ name }}/archive/{{ version }}.tar.gz
sha256: {{ sha256 }}
build:
number: 0
noarch: python
script: {{ PYTHON }} -m pip install . -vv
requirements:
host:
- python
- pip
- setuptools_scm
- setuptools_scm_git_archive
run:
- python
- numpy
- healpy
- astropy
- healsparse
- healsparse
- fitsio
- esutil
- LSSTDESC.Coord
- pyyaml
test:
imports:
- decasu
requires:
- pip
commands:
- pip check
about:
home: https://github.com/erykoff/decasu
summary: DECam (and HSC and LSSTCam) Survey Property Maps with HealSparse
license: BSD-3-Clause
license_family: BSD
license_file: LICENSE
dev_url: https://github.com/erykoff/decasu
extra:
recipe-maintainers:
- erykoff
| {% set name = "decasu" %}
{% set version = "0.7.1" %}
{% set sha256 = "d5558cd419c8d46bdc958064cb97f963d1ea793866414c025906ec15033512ed" %}
package:
name: {{ name|lower }}
version: {{ version }}
source:
url: https://github.com/erykoff/{{ name }}/archive/{{ version }}.tar.gz
sha256: {{ sha256 }}
build:
number: 0
noarch: python
script: {{ PYTHON }} -m pip install . -vv
requirements:
host:
- python >=3.7
- pip
- setuptools_scm
- setuptools_scm_git_archive
run:
- python >=3.7
- numpy
- healpy
- astropy
- healsparse
- healsparse
- fitsio
- esutil
- LSSTDESC.Coord
- pyyaml
test:
imports:
- decasu
requires:
- pip
commands:
- pip check
about:
home: https://github.com/erykoff/decasu
summary: DECam (and HSC and LSSTCam) Survey Property Maps with HealSparse
license: BSD-3-Clause
license_family: BSD
license_file: LICENSE
dev_url: https://github.com/erykoff/decasu
extra:
recipe-maintainers:
- erykoff
| Add lower bound on python support. | Add lower bound on python support.
| YAML | bsd-3-clause | goanpeca/staged-recipes,conda-forge/staged-recipes,hadim/staged-recipes,jakirkham/staged-recipes,ocefpaf/staged-recipes,johanneskoester/staged-recipes,conda-forge/staged-recipes,kwilcox/staged-recipes,goanpeca/staged-recipes,mariusvniekerk/staged-recipes,johanneskoester/staged-recipes,stuertz/staged-recipes,hadim/staged-recipes,kwilcox/staged-recipes,mariusvniekerk/staged-recipes,ocefpaf/staged-recipes,jakirkham/staged-recipes,stuertz/staged-recipes | yaml | ## Code Before:
{% set name = "decasu" %}
{% set version = "0.7.1" %}
{% set sha256 = "d5558cd419c8d46bdc958064cb97f963d1ea793866414c025906ec15033512ed" %}
package:
name: {{ name|lower }}
version: {{ version }}
source:
url: https://github.com/erykoff/{{ name }}/archive/{{ version }}.tar.gz
sha256: {{ sha256 }}
build:
number: 0
noarch: python
script: {{ PYTHON }} -m pip install . -vv
requirements:
host:
- python
- pip
- setuptools_scm
- setuptools_scm_git_archive
run:
- python
- numpy
- healpy
- astropy
- healsparse
- healsparse
- fitsio
- esutil
- LSSTDESC.Coord
- pyyaml
test:
imports:
- decasu
requires:
- pip
commands:
- pip check
about:
home: https://github.com/erykoff/decasu
summary: DECam (and HSC and LSSTCam) Survey Property Maps with HealSparse
license: BSD-3-Clause
license_family: BSD
license_file: LICENSE
dev_url: https://github.com/erykoff/decasu
extra:
recipe-maintainers:
- erykoff
## Instruction:
Add lower bound on python support.
## Code After:
{% set name = "decasu" %}
{% set version = "0.7.1" %}
{% set sha256 = "d5558cd419c8d46bdc958064cb97f963d1ea793866414c025906ec15033512ed" %}
package:
name: {{ name|lower }}
version: {{ version }}
source:
url: https://github.com/erykoff/{{ name }}/archive/{{ version }}.tar.gz
sha256: {{ sha256 }}
build:
number: 0
noarch: python
script: {{ PYTHON }} -m pip install . -vv
requirements:
host:
- python >=3.7
- pip
- setuptools_scm
- setuptools_scm_git_archive
run:
- python >=3.7
- numpy
- healpy
- astropy
- healsparse
- healsparse
- fitsio
- esutil
- LSSTDESC.Coord
- pyyaml
test:
imports:
- decasu
requires:
- pip
commands:
- pip check
about:
home: https://github.com/erykoff/decasu
summary: DECam (and HSC and LSSTCam) Survey Property Maps with HealSparse
license: BSD-3-Clause
license_family: BSD
license_file: LICENSE
dev_url: https://github.com/erykoff/decasu
extra:
recipe-maintainers:
- erykoff
| {% set name = "decasu" %}
{% set version = "0.7.1" %}
{% set sha256 = "d5558cd419c8d46bdc958064cb97f963d1ea793866414c025906ec15033512ed" %}
package:
name: {{ name|lower }}
version: {{ version }}
source:
url: https://github.com/erykoff/{{ name }}/archive/{{ version }}.tar.gz
sha256: {{ sha256 }}
build:
number: 0
noarch: python
script: {{ PYTHON }} -m pip install . -vv
requirements:
host:
- - python
+ - python >=3.7
? ++++++
- pip
- setuptools_scm
- setuptools_scm_git_archive
run:
- - python
+ - python >=3.7
? ++++++
- numpy
- healpy
- astropy
- healsparse
- healsparse
- fitsio
- esutil
- LSSTDESC.Coord
- pyyaml
test:
imports:
- decasu
requires:
- pip
commands:
- pip check
about:
home: https://github.com/erykoff/decasu
summary: DECam (and HSC and LSSTCam) Survey Property Maps with HealSparse
license: BSD-3-Clause
license_family: BSD
license_file: LICENSE
dev_url: https://github.com/erykoff/decasu
extra:
recipe-maintainers:
- erykoff | 4 | 0.074074 | 2 | 2 |
6d144b0b9ed2f1cfd38a0d58fb21dd8d74146282 | frameworks/Elixir/phoenix/setup.sh | frameworks/Elixir/phoenix/setup.sh |
export PATH=$PATH:$ERL_BIN:$ELX_BIN
# sed -i 's|db_host: "localhost",|db_host: "${DBHOST}",|g' config/config.exs
mix local.hex --force
mix local.rebar --force
mix deps.get --force
MIX_ENV=prod mix compile.protocols --force
MIX_ENV=prod elixir --detached -pa _build/$MIX_ENV/consolidated -S mix phoenix.server
|
export PATH=$PATH:$ERL_BIN:$ELX_BIN
# sed -i 's|db_host: "localhost",|db_host: "${DBHOST}",|g' config/config.exs
rm -rf _build deps
mix local.hex --force
mix local.rebar --force
mix deps.get --force
MIX_ENV=prod mix compile.protocols --force
MIX_ENV=prod elixir --detached -pa _build/$MIX_ENV/consolidated -S mix phoenix.server
| Remove _build and deps each build | Remove _build and deps each build
| Shell | bsd-3-clause | zdanek/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,valyala/FrameworkBenchmarks,doom369/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,sgml/FrameworkBenchmarks,sxend/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,valyala/FrameworkBenchmarks,zloster/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,sgml/FrameworkBenchmarks,actframework/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,testn/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,Verber/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,grob/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,hperadin/FrameworkBenchmarks,valyala/FrameworkBenchmarks,jamming/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,Verber/FrameworkBenchmarks,testn/FrameworkBenchmarks,doom369/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,zapov/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,grob/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,testn/FrameworkBenchmarks,sgml/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,denkab/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,herloct/FrameworkBenchmarks,testn/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,methane/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,khellang/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,kbrock/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,khellang/FrameworkBenchmarks,denkab/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,actframework/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,doom369/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,valyala/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,denkab/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,methane/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,jamming/FrameworkBenchmarks,zloster/FrameworkBenchmarks,jamming/FrameworkBenchmarks,khellang/FrameworkBenchmarks,zapov/FrameworkBenchmarks,sgml/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,actframework/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,zapov/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,Verber/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,methane/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,actframework/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,sxend/FrameworkBenchmarks,sgml/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,sgml/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,joshk/FrameworkBenchmarks,joshk/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,doom369/FrameworkBenchmarks,testn/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,valyala/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,kbrock/FrameworkBenchmarks,hperadin/FrameworkBenchmarks,actframework/FrameworkBenchmarks,zapov/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,herloct/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,sgml/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,herloct/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,valyala/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,sxend/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,kbrock/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,testn/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,sgml/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,zloster/FrameworkBenchmarks,hperadin/FrameworkBenchmarks,khellang/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,doom369/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,denkab/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,kbrock/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,sxend/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,actframework/FrameworkBenchmarks,doom369/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,denkab/FrameworkBenchmarks,Verber/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,methane/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,grob/FrameworkBenchmarks,valyala/FrameworkBenchmarks,jamming/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,jamming/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,zloster/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,valyala/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,zapov/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,denkab/FrameworkBenchmarks,methane/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,denkab/FrameworkBenchmarks,zapov/FrameworkBenchmarks,grob/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,joshk/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,hperadin/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,joshk/FrameworkBenchmarks,doom369/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,joshk/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,methane/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,valyala/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,sxend/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,hperadin/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,grob/FrameworkBenchmarks,sxend/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,sxend/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,khellang/FrameworkBenchmarks,valyala/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,zapov/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,joshk/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,denkab/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,zloster/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,sxend/FrameworkBenchmarks,grob/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,zloster/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,herloct/FrameworkBenchmarks,hperadin/FrameworkBenchmarks,hperadin/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,grob/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,khellang/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,zapov/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,sxend/FrameworkBenchmarks,khellang/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,zloster/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,jamming/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,sxend/FrameworkBenchmarks,herloct/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,Verber/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,Verber/FrameworkBenchmarks,methane/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,joshk/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,sxend/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,kbrock/FrameworkBenchmarks,joshk/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,sxend/FrameworkBenchmarks,kbrock/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,valyala/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,doom369/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,zloster/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,kbrock/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,zloster/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,grob/FrameworkBenchmarks,zapov/FrameworkBenchmarks,kbrock/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,herloct/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,joshk/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,jamming/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,Verber/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,zloster/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,actframework/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,zapov/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,sxend/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,grob/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,khellang/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,grob/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,zloster/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,grob/FrameworkBenchmarks,kbrock/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,zloster/FrameworkBenchmarks,herloct/FrameworkBenchmarks,actframework/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,doom369/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,sxend/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,denkab/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,Verber/FrameworkBenchmarks,sgml/FrameworkBenchmarks,actframework/FrameworkBenchmarks,hperadin/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,sgml/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,zapov/FrameworkBenchmarks,khellang/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,herloct/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,zloster/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,jamming/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,Verber/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,valyala/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,actframework/FrameworkBenchmarks,valyala/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,jamming/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,sgml/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,jamming/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,sgml/FrameworkBenchmarks,sxend/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,khellang/FrameworkBenchmarks,sxend/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,grob/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,methane/FrameworkBenchmarks,actframework/FrameworkBenchmarks,doom369/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,testn/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,doom369/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,khellang/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,zapov/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,Verber/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,herloct/FrameworkBenchmarks,zloster/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,Verber/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,kbrock/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,doom369/FrameworkBenchmarks,testn/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,doom369/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,doom369/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,methane/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,zapov/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,hperadin/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,joshk/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,khellang/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,joshk/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,grob/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,testn/FrameworkBenchmarks,denkab/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,sgml/FrameworkBenchmarks,zapov/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,methane/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,zapov/FrameworkBenchmarks,doom369/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,actframework/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,sxend/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,herloct/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,herloct/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,zapov/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,zloster/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,Verber/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,sgml/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,denkab/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,kbrock/FrameworkBenchmarks,doom369/FrameworkBenchmarks,joshk/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,methane/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,herloct/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,denkab/FrameworkBenchmarks,methane/FrameworkBenchmarks,methane/FrameworkBenchmarks,actframework/FrameworkBenchmarks,herloct/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,denkab/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,sxend/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,jamming/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,methane/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,joshk/FrameworkBenchmarks,testn/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,doom369/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,jamming/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,hperadin/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,Verber/FrameworkBenchmarks,actframework/FrameworkBenchmarks,herloct/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,joshk/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,jamming/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,testn/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,herloct/FrameworkBenchmarks,denkab/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,zloster/FrameworkBenchmarks,testn/FrameworkBenchmarks,jamming/FrameworkBenchmarks,khellang/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,testn/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,khellang/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,testn/FrameworkBenchmarks,zloster/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,hperadin/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,actframework/FrameworkBenchmarks,hperadin/FrameworkBenchmarks,kbrock/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,valyala/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,hperadin/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,Verber/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,grob/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,zloster/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,actframework/FrameworkBenchmarks | shell | ## Code Before:
export PATH=$PATH:$ERL_BIN:$ELX_BIN
# sed -i 's|db_host: "localhost",|db_host: "${DBHOST}",|g' config/config.exs
mix local.hex --force
mix local.rebar --force
mix deps.get --force
MIX_ENV=prod mix compile.protocols --force
MIX_ENV=prod elixir --detached -pa _build/$MIX_ENV/consolidated -S mix phoenix.server
## Instruction:
Remove _build and deps each build
## Code After:
export PATH=$PATH:$ERL_BIN:$ELX_BIN
# sed -i 's|db_host: "localhost",|db_host: "${DBHOST}",|g' config/config.exs
rm -rf _build deps
mix local.hex --force
mix local.rebar --force
mix deps.get --force
MIX_ENV=prod mix compile.protocols --force
MIX_ENV=prod elixir --detached -pa _build/$MIX_ENV/consolidated -S mix phoenix.server
|
export PATH=$PATH:$ERL_BIN:$ELX_BIN
# sed -i 's|db_host: "localhost",|db_host: "${DBHOST}",|g' config/config.exs
+
+ rm -rf _build deps
mix local.hex --force
mix local.rebar --force
mix deps.get --force
MIX_ENV=prod mix compile.protocols --force
MIX_ENV=prod elixir --detached -pa _build/$MIX_ENV/consolidated -S mix phoenix.server | 2 | 0.181818 | 2 | 0 |
9d0e657401d7e7ab179cb66ac68851009deb2a7e | src/edu/northwestern/bioinformatics/studycalendar/dao/delta/AmendmentDao.java | src/edu/northwestern/bioinformatics/studycalendar/dao/delta/AmendmentDao.java | package edu.northwestern.bioinformatics.studycalendar.dao.delta;
import org.springframework.transaction.annotation.Transactional;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
import edu.northwestern.bioinformatics.studycalendar.domain.delta.Amendment;
import edu.northwestern.bioinformatics.studycalendar.dao.StudyCalendarDao;
import java.util.List;
/**
* Created by IntelliJ IDEA.
* User: nshurupova
* Date: Aug 29, 2007
* Time: 3:41:38 PM
* To change this template use File | Settings | File Templates.
*/
public class AmendmentDao extends StudyCalendarDao<Amendment> {
private static final Logger log = LoggerFactory.getLogger(AmendmentDao.class.getName());
public Class<Amendment> domainClass() {
return Amendment.class;
}
public List<Amendment> getAll() {
return getHibernateTemplate().find("from Amendment");
}
@Transactional(readOnly = false)
public void save(Amendment amendment) {
getHibernateTemplate().saveOrUpdate(amendment);
}
public List<Amendment> getByName (String name) {
return getHibernateTemplate().find("from Amendment a where a.name= ?", name);
}
public List<Amendment> getByDate(String date) {
return getHibernateTemplate().find("from Amendment a where a.date= ?", date);
}
public Amendment getByPreviousAmendmentId(Integer previousAmendmentId) {
List<Amendment> results = getHibernateTemplate().find("from Amendment a where a.previousAmendment= ?", previousAmendmentId);
return results.get(0);
}
public Amendment getByStudyId(Integer studyId) {
List<Amendment> results = getHibernateTemplate().find("from Amendment a where a.studyId= ?", studyId);
return results.get(0);
}
}
| package edu.northwestern.bioinformatics.studycalendar.dao.delta;
import edu.northwestern.bioinformatics.studycalendar.dao.StudyCalendarDao;
import edu.northwestern.bioinformatics.studycalendar.domain.delta.Amendment;
import org.springframework.transaction.annotation.Transactional;
import java.util.List;
/**
* @author Nataliya Shurupova
*/
public class AmendmentDao extends StudyCalendarDao<Amendment> {
@Override
public Class<Amendment> domainClass() {
return Amendment.class;
}
public List<Amendment> getAll() {
return getHibernateTemplate().find("from Amendment");
}
@Transactional(readOnly = false)
public void save(Amendment amendment) {
getHibernateTemplate().saveOrUpdate(amendment);
}
@Deprecated
public Amendment getByStudyId(Integer studyId) {
throw new UnsupportedOperationException("Deprecated");
}
}
| Remove unused/untested methods; deprecate & de-support getStudyById because that field's about to disappear | Remove unused/untested methods; deprecate & de-support getStudyById because that field's about to disappear
git-svn-id: 4b387fe5ada7764508e2ca96c335714e4c1692c6@1279 0d517254-b314-0410-acde-c619094fa49f
| Java | bsd-3-clause | NUBIC/psc-mirror,NUBIC/psc-mirror,NUBIC/psc-mirror,NUBIC/psc-mirror | java | ## Code Before:
package edu.northwestern.bioinformatics.studycalendar.dao.delta;
import org.springframework.transaction.annotation.Transactional;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
import edu.northwestern.bioinformatics.studycalendar.domain.delta.Amendment;
import edu.northwestern.bioinformatics.studycalendar.dao.StudyCalendarDao;
import java.util.List;
/**
* Created by IntelliJ IDEA.
* User: nshurupova
* Date: Aug 29, 2007
* Time: 3:41:38 PM
* To change this template use File | Settings | File Templates.
*/
public class AmendmentDao extends StudyCalendarDao<Amendment> {
private static final Logger log = LoggerFactory.getLogger(AmendmentDao.class.getName());
public Class<Amendment> domainClass() {
return Amendment.class;
}
public List<Amendment> getAll() {
return getHibernateTemplate().find("from Amendment");
}
@Transactional(readOnly = false)
public void save(Amendment amendment) {
getHibernateTemplate().saveOrUpdate(amendment);
}
public List<Amendment> getByName (String name) {
return getHibernateTemplate().find("from Amendment a where a.name= ?", name);
}
public List<Amendment> getByDate(String date) {
return getHibernateTemplate().find("from Amendment a where a.date= ?", date);
}
public Amendment getByPreviousAmendmentId(Integer previousAmendmentId) {
List<Amendment> results = getHibernateTemplate().find("from Amendment a where a.previousAmendment= ?", previousAmendmentId);
return results.get(0);
}
public Amendment getByStudyId(Integer studyId) {
List<Amendment> results = getHibernateTemplate().find("from Amendment a where a.studyId= ?", studyId);
return results.get(0);
}
}
## Instruction:
Remove unused/untested methods; deprecate & de-support getStudyById because that field's about to disappear
git-svn-id: 4b387fe5ada7764508e2ca96c335714e4c1692c6@1279 0d517254-b314-0410-acde-c619094fa49f
## Code After:
package edu.northwestern.bioinformatics.studycalendar.dao.delta;
import edu.northwestern.bioinformatics.studycalendar.dao.StudyCalendarDao;
import edu.northwestern.bioinformatics.studycalendar.domain.delta.Amendment;
import org.springframework.transaction.annotation.Transactional;
import java.util.List;
/**
* @author Nataliya Shurupova
*/
public class AmendmentDao extends StudyCalendarDao<Amendment> {
@Override
public Class<Amendment> domainClass() {
return Amendment.class;
}
public List<Amendment> getAll() {
return getHibernateTemplate().find("from Amendment");
}
@Transactional(readOnly = false)
public void save(Amendment amendment) {
getHibernateTemplate().saveOrUpdate(amendment);
}
@Deprecated
public Amendment getByStudyId(Integer studyId) {
throw new UnsupportedOperationException("Deprecated");
}
}
| package edu.northwestern.bioinformatics.studycalendar.dao.delta;
+ import edu.northwestern.bioinformatics.studycalendar.dao.StudyCalendarDao;
+ import edu.northwestern.bioinformatics.studycalendar.domain.delta.Amendment;
import org.springframework.transaction.annotation.Transactional;
- import org.slf4j.Logger;
- import org.slf4j.LoggerFactory;
- import edu.northwestern.bioinformatics.studycalendar.domain.delta.Amendment;
- import edu.northwestern.bioinformatics.studycalendar.dao.StudyCalendarDao;
import java.util.List;
/**
+ * @author Nataliya Shurupova
- * Created by IntelliJ IDEA.
- * User: nshurupova
- * Date: Aug 29, 2007
- * Time: 3:41:38 PM
- * To change this template use File | Settings | File Templates.
*/
public class AmendmentDao extends StudyCalendarDao<Amendment> {
+ @Override
- private static final Logger log = LoggerFactory.getLogger(AmendmentDao.class.getName());
-
public Class<Amendment> domainClass() {
return Amendment.class;
}
public List<Amendment> getAll() {
return getHibernateTemplate().find("from Amendment");
}
@Transactional(readOnly = false)
public void save(Amendment amendment) {
getHibernateTemplate().saveOrUpdate(amendment);
}
+ @Deprecated
- public List<Amendment> getByName (String name) {
- return getHibernateTemplate().find("from Amendment a where a.name= ?", name);
- }
-
- public List<Amendment> getByDate(String date) {
- return getHibernateTemplate().find("from Amendment a where a.date= ?", date);
- }
-
- public Amendment getByPreviousAmendmentId(Integer previousAmendmentId) {
- List<Amendment> results = getHibernateTemplate().find("from Amendment a where a.previousAmendment= ?", previousAmendmentId);
- return results.get(0);
- }
-
public Amendment getByStudyId(Integer studyId) {
+ throw new UnsupportedOperationException("Deprecated");
- List<Amendment> results = getHibernateTemplate().find("from Amendment a where a.studyId= ?", studyId);
- return results.get(0);
}
} | 32 | 0.627451 | 6 | 26 |
280ed973c549aec8892606f7be1d5824aca80b90 | src/app/welcome.ts | src/app/welcome.ts | module welcome {
export class Welcome {
heading = 'Welcome to the Aurelia Navigation App!';
firstName = 'John';
lastName = 'Doe';
get fullName() {
return `${this.firstName} ${this.lastName}`;
}
welcome() {
alert(`Welcome, ${this.fullName}!`);
}
}
export class UpperValueConverter {
toView(value) {
return value && value.toUpperCase();
}
}
}
| module welcome {
export class Welcome {
heading = 'Welcome to the Aurelia Navigation App!';
firstName = 'John';
lastName = 'Doe';
get fullName() {
return `${this.firstName} ${this.lastName}`;
}
welcome() {
alert(`Welcome, ${this.fullName}!`);
}
}
export class UpperValueConverter {
toView(value) {
return value && value.toUpperCase();
}
}
}
export = welcome;
| Add export statement to Welcome module | Add export statement to Welcome module
| TypeScript | mit | bohnen/aurelia-skeleton-navigation-gulp-typescript,bohnen/aurelia-skeleton-navigation-gulp-typescript,Enrapt/aurelia-skeleton-navigation-gulp-typescript,bohnen/aurelia-skeleton-navigation-gulp-typescript,Enrapt/aurelia-skeleton-navigation-gulp-typescript,Enrapt/aurelia-skeleton-navigation-gulp-typescript | typescript | ## Code Before:
module welcome {
export class Welcome {
heading = 'Welcome to the Aurelia Navigation App!';
firstName = 'John';
lastName = 'Doe';
get fullName() {
return `${this.firstName} ${this.lastName}`;
}
welcome() {
alert(`Welcome, ${this.fullName}!`);
}
}
export class UpperValueConverter {
toView(value) {
return value && value.toUpperCase();
}
}
}
## Instruction:
Add export statement to Welcome module
## Code After:
module welcome {
export class Welcome {
heading = 'Welcome to the Aurelia Navigation App!';
firstName = 'John';
lastName = 'Doe';
get fullName() {
return `${this.firstName} ${this.lastName}`;
}
welcome() {
alert(`Welcome, ${this.fullName}!`);
}
}
export class UpperValueConverter {
toView(value) {
return value && value.toUpperCase();
}
}
}
export = welcome;
| module welcome {
export class Welcome {
heading = 'Welcome to the Aurelia Navigation App!';
firstName = 'John';
lastName = 'Doe';
get fullName() {
return `${this.firstName} ${this.lastName}`;
}
welcome() {
alert(`Welcome, ${this.fullName}!`);
}
}
export class UpperValueConverter {
toView(value) {
return value && value.toUpperCase();
}
}
}
+
+ export = welcome; | 2 | 0.083333 | 2 | 0 |
3fa0dd1c2183fd00f8ec7162936dfdf63a34f858 | data-service/src/main/java/org/ihtsdo/buildcloud/service/build/transform/RepeatableRelationshipUUIDTransform.java | data-service/src/main/java/org/ihtsdo/buildcloud/service/build/transform/RepeatableRelationshipUUIDTransform.java | package org.ihtsdo.buildcloud.service.build.transform;
import org.ihtsdo.buildcloud.service.build.RF2Constants;
import org.ihtsdo.buildcloud.service.helper.Type5UuidFactory;
import java.io.UnsupportedEncodingException;
import java.security.NoSuchAlgorithmException;
public class RepeatableRelationshipUUIDTransform implements LineTransformation {
private Type5UuidFactory type5UuidFactory;
public RepeatableRelationshipUUIDTransform() throws NoSuchAlgorithmException {
type5UuidFactory = new Type5UuidFactory();
}
@Override
public void transformLine(String[] columnValues) throws TransformationException {
// Create repeatable UUID to ensure SCTIDs are reused.
// (Technique lifted from workbench release process.)
// sourceId + destinationId + typeId + relationshipGroup
if (columnValues[0].equals(RF2Constants.NULL_STRING) || columnValues[0].isEmpty()) {
try {
columnValues[0] = getCalculatedUuidFromRelationshipValues(columnValues);
} catch (UnsupportedEncodingException e) {
throw new TransformationException("Failed to create UUID.", e);
}
}
}
public String getCalculatedUuidFromRelationshipValues(String[] columnValues) throws UnsupportedEncodingException {
return type5UuidFactory.get(columnValues[4] + columnValues[5] + columnValues[7] + columnValues[6]).toString();
}
@Override
public int getColumnIndex() {
return -1;
}
}
| package org.ihtsdo.buildcloud.service.build.transform;
import org.ihtsdo.buildcloud.service.build.RF2Constants;
import org.ihtsdo.buildcloud.service.helper.Type5UuidFactory;
import java.io.UnsupportedEncodingException;
import java.security.NoSuchAlgorithmException;
public class RepeatableRelationshipUUIDTransform implements LineTransformation {
private Type5UuidFactory type5UuidFactory;
public RepeatableRelationshipUUIDTransform() throws NoSuchAlgorithmException {
type5UuidFactory = new Type5UuidFactory();
}
@Override
public void transformLine(String[] columnValues) throws TransformationException {
// Create repeatable UUID to ensure SCTIDs are reused.
// (Technique lifted from workbench release process.)
// sourceId + destinationId + typeId + relationshipGroup
if (columnValues[0] == null || columnValues[0].equals(RF2Constants.NULL_STRING) || columnValues[0].isEmpty()) {
try {
columnValues[0] = getCalculatedUuidFromRelationshipValues(columnValues);
} catch (UnsupportedEncodingException e) {
throw new TransformationException("Failed to create UUID.", e);
}
}
}
public String getCalculatedUuidFromRelationshipValues(String[] columnValues) throws UnsupportedEncodingException {
return type5UuidFactory.get(columnValues[4] + columnValues[5] + columnValues[7] + columnValues[6]).toString();
}
@Override
public int getColumnIndex() {
return -1;
}
}
| Allow SCTID in relationship file to be actually null, not just "null" | Allow SCTID in relationship file to be actually null, not just "null" | Java | apache-2.0 | IHTSDO/snomed-release-service,IHTSDO/snomed-release-service | java | ## Code Before:
package org.ihtsdo.buildcloud.service.build.transform;
import org.ihtsdo.buildcloud.service.build.RF2Constants;
import org.ihtsdo.buildcloud.service.helper.Type5UuidFactory;
import java.io.UnsupportedEncodingException;
import java.security.NoSuchAlgorithmException;
public class RepeatableRelationshipUUIDTransform implements LineTransformation {
private Type5UuidFactory type5UuidFactory;
public RepeatableRelationshipUUIDTransform() throws NoSuchAlgorithmException {
type5UuidFactory = new Type5UuidFactory();
}
@Override
public void transformLine(String[] columnValues) throws TransformationException {
// Create repeatable UUID to ensure SCTIDs are reused.
// (Technique lifted from workbench release process.)
// sourceId + destinationId + typeId + relationshipGroup
if (columnValues[0].equals(RF2Constants.NULL_STRING) || columnValues[0].isEmpty()) {
try {
columnValues[0] = getCalculatedUuidFromRelationshipValues(columnValues);
} catch (UnsupportedEncodingException e) {
throw new TransformationException("Failed to create UUID.", e);
}
}
}
public String getCalculatedUuidFromRelationshipValues(String[] columnValues) throws UnsupportedEncodingException {
return type5UuidFactory.get(columnValues[4] + columnValues[5] + columnValues[7] + columnValues[6]).toString();
}
@Override
public int getColumnIndex() {
return -1;
}
}
## Instruction:
Allow SCTID in relationship file to be actually null, not just "null"
## Code After:
package org.ihtsdo.buildcloud.service.build.transform;
import org.ihtsdo.buildcloud.service.build.RF2Constants;
import org.ihtsdo.buildcloud.service.helper.Type5UuidFactory;
import java.io.UnsupportedEncodingException;
import java.security.NoSuchAlgorithmException;
public class RepeatableRelationshipUUIDTransform implements LineTransformation {
private Type5UuidFactory type5UuidFactory;
public RepeatableRelationshipUUIDTransform() throws NoSuchAlgorithmException {
type5UuidFactory = new Type5UuidFactory();
}
@Override
public void transformLine(String[] columnValues) throws TransformationException {
// Create repeatable UUID to ensure SCTIDs are reused.
// (Technique lifted from workbench release process.)
// sourceId + destinationId + typeId + relationshipGroup
if (columnValues[0] == null || columnValues[0].equals(RF2Constants.NULL_STRING) || columnValues[0].isEmpty()) {
try {
columnValues[0] = getCalculatedUuidFromRelationshipValues(columnValues);
} catch (UnsupportedEncodingException e) {
throw new TransformationException("Failed to create UUID.", e);
}
}
}
public String getCalculatedUuidFromRelationshipValues(String[] columnValues) throws UnsupportedEncodingException {
return type5UuidFactory.get(columnValues[4] + columnValues[5] + columnValues[7] + columnValues[6]).toString();
}
@Override
public int getColumnIndex() {
return -1;
}
}
| package org.ihtsdo.buildcloud.service.build.transform;
import org.ihtsdo.buildcloud.service.build.RF2Constants;
import org.ihtsdo.buildcloud.service.helper.Type5UuidFactory;
import java.io.UnsupportedEncodingException;
import java.security.NoSuchAlgorithmException;
public class RepeatableRelationshipUUIDTransform implements LineTransformation {
private Type5UuidFactory type5UuidFactory;
public RepeatableRelationshipUUIDTransform() throws NoSuchAlgorithmException {
type5UuidFactory = new Type5UuidFactory();
}
@Override
public void transformLine(String[] columnValues) throws TransformationException {
// Create repeatable UUID to ensure SCTIDs are reused.
// (Technique lifted from workbench release process.)
// sourceId + destinationId + typeId + relationshipGroup
- if (columnValues[0].equals(RF2Constants.NULL_STRING) || columnValues[0].isEmpty()) {
+ if (columnValues[0] == null || columnValues[0].equals(RF2Constants.NULL_STRING) || columnValues[0].isEmpty()) {
? +++++++++++++++++++++++++++
try {
columnValues[0] = getCalculatedUuidFromRelationshipValues(columnValues);
} catch (UnsupportedEncodingException e) {
throw new TransformationException("Failed to create UUID.", e);
}
}
}
public String getCalculatedUuidFromRelationshipValues(String[] columnValues) throws UnsupportedEncodingException {
return type5UuidFactory.get(columnValues[4] + columnValues[5] + columnValues[7] + columnValues[6]).toString();
}
@Override
public int getColumnIndex() {
return -1;
}
} | 2 | 0.05 | 1 | 1 |
1c4db4907283095afd15e8114dfb297b62844a5f | roles/nginx/tasks/main.yml | roles/nginx/tasks/main.yml | ---
- name: Adding NGINX Repository, Hold on...
apt_repository: repo=ppa:miteshshah/nginx-pagespeed
register: nginx_repo
- include: roles/libs/tasks/apt/update.yml
when: nginx_repo.changed == True or nginx_key.changed == True
- name: Installing NGINX, Hold on...
apt: name=nginx-pagespeed state=present
register: package_install
# The notify will call the ../handlers/main.yml
notify: service nginx restart
| ---
- name: Adding NGINX Repository, Hold on...
apt_repository: repo=ppa:miteshshah/nginx-pagespeed
register: nginx_repo
- include: roles/libs/tasks/apt/update.yml
when: nginx_repo.changed == True
- name: Installing NGINX, Hold on...
apt: name=nginx-pagespeed state=present
register: package_install
# The notify will call the ../handlers/main.yml
notify: service nginx restart
| Remove NGINX key not needed | Remove NGINX key not needed
| YAML | mit | AnsiPress/AnsiPress,AnsiPress/AnsiPress | yaml | ## Code Before:
---
- name: Adding NGINX Repository, Hold on...
apt_repository: repo=ppa:miteshshah/nginx-pagespeed
register: nginx_repo
- include: roles/libs/tasks/apt/update.yml
when: nginx_repo.changed == True or nginx_key.changed == True
- name: Installing NGINX, Hold on...
apt: name=nginx-pagespeed state=present
register: package_install
# The notify will call the ../handlers/main.yml
notify: service nginx restart
## Instruction:
Remove NGINX key not needed
## Code After:
---
- name: Adding NGINX Repository, Hold on...
apt_repository: repo=ppa:miteshshah/nginx-pagespeed
register: nginx_repo
- include: roles/libs/tasks/apt/update.yml
when: nginx_repo.changed == True
- name: Installing NGINX, Hold on...
apt: name=nginx-pagespeed state=present
register: package_install
# The notify will call the ../handlers/main.yml
notify: service nginx restart
| ---
- name: Adding NGINX Repository, Hold on...
apt_repository: repo=ppa:miteshshah/nginx-pagespeed
register: nginx_repo
- include: roles/libs/tasks/apt/update.yml
- when: nginx_repo.changed == True or nginx_key.changed == True
+ when: nginx_repo.changed == True
- name: Installing NGINX, Hold on...
apt: name=nginx-pagespeed state=present
register: package_install
# The notify will call the ../handlers/main.yml
notify: service nginx restart | 2 | 0.153846 | 1 | 1 |
4047581980cf46fa86d372c525f7b0ebd0636d2c | app/controllers/index.rb | app/controllers/index.rb | get '/' do
erb :index
end
get '/upload' do
haml :"/partials/_upload"
end
post '/images' do
p "*" * 100
p params
p "*" * 100
return params[:file]
end
# Handle POST-request (Receive and save the uploaded file)
post "/" do
p "*" * 100
p params
p "*" * 100
File.open(File.dirname(__FILE__) + '/../../public/uploads/' + params['myfile'][:filename], "w") do |f|
f.write(File.open(params['myfile'][:tempfile], "r").read)
end
new_image = Imagefile.create( filename: params['myfile'][:filename],
url: '/uploads/' + params['myfile'][:filename])
if request.xhr?
return File.dirname(__FILE__) + '/../../public/uploads/' + params['myfile'][:filename]
else
redirect "/review/#{new_image.id}"
end
end
get '/review/:id' do
@image = Imagefile.find_by(id: params[:id])
if request.xhr?
# haml :"/partials/_review"
"HAML"
else
erb :"/partials/_review"
end
end | get '/' do
erb :index
end
get '/upload' do
haml :"/partials/_upload"
end
post '/images' do
p "*" * 100
p params
p "*" * 100
return params[:file]
end
# Handle POST-request (Receive and save the uploaded file)
post "/" do
p "*" * 100
p APP_ROOT.join(Dir.pwd + '/public/uploads/' + params['myfile'][:filename]).to_s
p "*" * 100
File.open(APP_ROOT.join(Dir.pwd + '/public/uploads/' + params['myfile'][:filename]).to_s, "w") do |f|
f.write(File.open(params['myfile'][:tempfile], "r").read)
end
new_image = Imagefile.create( filename: params['myfile'][:filename],
url: '/uploads/' + params['myfile'][:filename])
if request.xhr?
return File.dirname(__FILE__) + '/../../public/uploads/' + params['myfile'][:filename]
else
redirect "/review/#{new_image.id}"
end
end
get '/review/:id' do
@image = Imagefile.find_by(id: params[:id])
if request.xhr?
# haml :"/partials/_review"
"HAML"
else
erb :"/partials/_review"
end
end | Fix pathing issue with image upload | Fix pathing issue with image upload
| Ruby | mit | ShadyLogic/WillThisKillMe,ShadyLogic/WillThisKillMe,ShadyLogic/WillThisKillMe | ruby | ## Code Before:
get '/' do
erb :index
end
get '/upload' do
haml :"/partials/_upload"
end
post '/images' do
p "*" * 100
p params
p "*" * 100
return params[:file]
end
# Handle POST-request (Receive and save the uploaded file)
post "/" do
p "*" * 100
p params
p "*" * 100
File.open(File.dirname(__FILE__) + '/../../public/uploads/' + params['myfile'][:filename], "w") do |f|
f.write(File.open(params['myfile'][:tempfile], "r").read)
end
new_image = Imagefile.create( filename: params['myfile'][:filename],
url: '/uploads/' + params['myfile'][:filename])
if request.xhr?
return File.dirname(__FILE__) + '/../../public/uploads/' + params['myfile'][:filename]
else
redirect "/review/#{new_image.id}"
end
end
get '/review/:id' do
@image = Imagefile.find_by(id: params[:id])
if request.xhr?
# haml :"/partials/_review"
"HAML"
else
erb :"/partials/_review"
end
end
## Instruction:
Fix pathing issue with image upload
## Code After:
get '/' do
erb :index
end
get '/upload' do
haml :"/partials/_upload"
end
post '/images' do
p "*" * 100
p params
p "*" * 100
return params[:file]
end
# Handle POST-request (Receive and save the uploaded file)
post "/" do
p "*" * 100
p APP_ROOT.join(Dir.pwd + '/public/uploads/' + params['myfile'][:filename]).to_s
p "*" * 100
File.open(APP_ROOT.join(Dir.pwd + '/public/uploads/' + params['myfile'][:filename]).to_s, "w") do |f|
f.write(File.open(params['myfile'][:tempfile], "r").read)
end
new_image = Imagefile.create( filename: params['myfile'][:filename],
url: '/uploads/' + params['myfile'][:filename])
if request.xhr?
return File.dirname(__FILE__) + '/../../public/uploads/' + params['myfile'][:filename]
else
redirect "/review/#{new_image.id}"
end
end
get '/review/:id' do
@image = Imagefile.find_by(id: params[:id])
if request.xhr?
# haml :"/partials/_review"
"HAML"
else
erb :"/partials/_review"
end
end | get '/' do
erb :index
end
get '/upload' do
haml :"/partials/_upload"
end
post '/images' do
p "*" * 100
p params
p "*" * 100
return params[:file]
end
# Handle POST-request (Receive and save the uploaded file)
post "/" do
p "*" * 100
- p params
+ p APP_ROOT.join(Dir.pwd + '/public/uploads/' + params['myfile'][:filename]).to_s
p "*" * 100
- File.open(File.dirname(__FILE__) + '/../../public/uploads/' + params['myfile'][:filename], "w") do |f|
? ^ ^^^^ ^^^^^^^^^^^^^^ ------
+ File.open(APP_ROOT.join(Dir.pwd + '/public/uploads/' + params['myfile'][:filename]).to_s, "w") do |f|
? ^^^^^^^^^^^ ^^^ ^^^^ ++++++
f.write(File.open(params['myfile'][:tempfile], "r").read)
end
new_image = Imagefile.create( filename: params['myfile'][:filename],
url: '/uploads/' + params['myfile'][:filename])
if request.xhr?
return File.dirname(__FILE__) + '/../../public/uploads/' + params['myfile'][:filename]
else
redirect "/review/#{new_image.id}"
end
end
get '/review/:id' do
@image = Imagefile.find_by(id: params[:id])
if request.xhr?
# haml :"/partials/_review"
"HAML"
else
erb :"/partials/_review"
end
end | 4 | 0.090909 | 2 | 2 |
75e85a99c2b092e55076081c435c93873a49f8e3 | .travis.yml | .travis.yml | language: objective-c
xcode_workspace: DFDiff.xcworkspace
xcode_scheme: DFDiff
| language: objective-c
osx_image: xcode7.2
xcode_workspace: DFDiff.xcworkspace
xcode_scheme: DFDiff
| Select xcode7.2 for travi builds. | Select xcode7.2 for travi builds.
| YAML | mit | marciniwanicki/DFDiff,marciniwanicki/DFDiff | yaml | ## Code Before:
language: objective-c
xcode_workspace: DFDiff.xcworkspace
xcode_scheme: DFDiff
## Instruction:
Select xcode7.2 for travi builds.
## Code After:
language: objective-c
osx_image: xcode7.2
xcode_workspace: DFDiff.xcworkspace
xcode_scheme: DFDiff
| language: objective-c
+ osx_image: xcode7.2
xcode_workspace: DFDiff.xcworkspace
xcode_scheme: DFDiff | 1 | 0.333333 | 1 | 0 |
53b06419934de739df2c1e681703c9db249a9334 | .travis.yml | .travis.yml | language: "node_js"
node_js:
- "0.10"
before_script:
- sudo apt-get install -qq graphicsmagick
- cp ./conf/envDefault.js ./conf/envLocal.js
| language: "node_js"
node_js:
- "0.10"
services:
- redis-server
before_script:
- sudo apt-get install -qq graphicsmagick
- cp ./conf/envDefault.js ./conf/envLocal.js
| Add Redis to Travis CI config | Add Redis to Travis CI config
Fixes "error event - 127.0.0.1:6379 - Error: Redis connection to 127.0.0.1:6379 failed - connect ECONNREFUSED" in Travis CI | YAML | mit | clbn/pepyatka-server,berkus/pepyatka-server | yaml | ## Code Before:
language: "node_js"
node_js:
- "0.10"
before_script:
- sudo apt-get install -qq graphicsmagick
- cp ./conf/envDefault.js ./conf/envLocal.js
## Instruction:
Add Redis to Travis CI config
Fixes "error event - 127.0.0.1:6379 - Error: Redis connection to 127.0.0.1:6379 failed - connect ECONNREFUSED" in Travis CI
## Code After:
language: "node_js"
node_js:
- "0.10"
services:
- redis-server
before_script:
- sudo apt-get install -qq graphicsmagick
- cp ./conf/envDefault.js ./conf/envLocal.js
| language: "node_js"
node_js:
- "0.10"
+ services:
+ - redis-server
before_script:
- sudo apt-get install -qq graphicsmagick
- cp ./conf/envDefault.js ./conf/envLocal.js | 2 | 0.333333 | 2 | 0 |
9309b70caa84bab66cee111a8e2ab556656ca6f9 | lib/schemeIO.js | lib/schemeIO.js | (function (modFactory) {
if (typeof module === 'object' && module.exports) {
module.exports = modFactory(require('./scheme'), require('./parseScheme'));
} else if (typeof define === "function" && define.amd) {
define(['./parser', './parseScheme'], modFactory);
} else {
throw new Error('Use a module loader!');
}
}(function (s, parseScheme) {
function printMixin (env, fn) {
env.print = new s.SchemePrimativeFunction(function (val) {
fn(val.toString());
return val;
});
}
function loadMixin (env, fn) {
env.load = new s.SchemePrimativeFunction(function (filename) {
if (!(filename instanceof s.SchemeString)) {
throw new s.SchemeError('Expecting ' + s.SchemeString.typeName() + '. ' +
'Found ' + filename.toString());
}
var content = fn(filename.val),
expressions = parseScheme.many(content);
return expressions
.get('matched')
.each(function (expr) {
expr.eval(env);
})
.then(function () {
return new s.SchemeBool(true);
});
});
}
return {
printMixin: printMixin,
loadMixin: loadMixin
};
}));
| (function (modFactory) {
if (typeof module === 'object' && module.exports) {
module.exports = modFactory(require('./scheme'), require('./parseScheme'));
} else if (typeof define === "function" && define.amd) {
define(['./parser', './parseScheme'], modFactory);
} else {
throw new Error('Use a module loader!');
}
}(function (s, parseScheme) {
function printMixin (env, fn) {
env.print = new s.SchemePrimativeFunction(function (val) {
fn(val.toString());
return val;
});
}
function loadMixin (env, fn) {
env.load = new s.SchemePrimativeFunction(function (filename) {
var content, expressions;
if (fn.length > 0) {
if (!(filename instanceof s.SchemeString)) {
throw new s.SchemeError('Expecting ' + s.SchemeString.typeName() + '. ' +
'Found ' + (filename ? filename.toString() : 'Nothing'));
}
content = fn(filename.val);
} else {
content = fn();
}
expressions = parseScheme.many(content);
return expressions
.get('matched')
.each(function (expr) {
expr.eval(env);
})
.then(function () {
return new s.SchemeBool(true);
});
});
}
return {
printMixin: printMixin,
loadMixin: loadMixin
};
}));
| Allow load mixin function to have no args | Allow load mixin function to have no args
| JavaScript | mit | ljwall/tinyJSScheme | javascript | ## Code Before:
(function (modFactory) {
if (typeof module === 'object' && module.exports) {
module.exports = modFactory(require('./scheme'), require('./parseScheme'));
} else if (typeof define === "function" && define.amd) {
define(['./parser', './parseScheme'], modFactory);
} else {
throw new Error('Use a module loader!');
}
}(function (s, parseScheme) {
function printMixin (env, fn) {
env.print = new s.SchemePrimativeFunction(function (val) {
fn(val.toString());
return val;
});
}
function loadMixin (env, fn) {
env.load = new s.SchemePrimativeFunction(function (filename) {
if (!(filename instanceof s.SchemeString)) {
throw new s.SchemeError('Expecting ' + s.SchemeString.typeName() + '. ' +
'Found ' + filename.toString());
}
var content = fn(filename.val),
expressions = parseScheme.many(content);
return expressions
.get('matched')
.each(function (expr) {
expr.eval(env);
})
.then(function () {
return new s.SchemeBool(true);
});
});
}
return {
printMixin: printMixin,
loadMixin: loadMixin
};
}));
## Instruction:
Allow load mixin function to have no args
## Code After:
(function (modFactory) {
if (typeof module === 'object' && module.exports) {
module.exports = modFactory(require('./scheme'), require('./parseScheme'));
} else if (typeof define === "function" && define.amd) {
define(['./parser', './parseScheme'], modFactory);
} else {
throw new Error('Use a module loader!');
}
}(function (s, parseScheme) {
function printMixin (env, fn) {
env.print = new s.SchemePrimativeFunction(function (val) {
fn(val.toString());
return val;
});
}
function loadMixin (env, fn) {
env.load = new s.SchemePrimativeFunction(function (filename) {
var content, expressions;
if (fn.length > 0) {
if (!(filename instanceof s.SchemeString)) {
throw new s.SchemeError('Expecting ' + s.SchemeString.typeName() + '. ' +
'Found ' + (filename ? filename.toString() : 'Nothing'));
}
content = fn(filename.val);
} else {
content = fn();
}
expressions = parseScheme.many(content);
return expressions
.get('matched')
.each(function (expr) {
expr.eval(env);
})
.then(function () {
return new s.SchemeBool(true);
});
});
}
return {
printMixin: printMixin,
loadMixin: loadMixin
};
}));
| (function (modFactory) {
if (typeof module === 'object' && module.exports) {
module.exports = modFactory(require('./scheme'), require('./parseScheme'));
} else if (typeof define === "function" && define.amd) {
define(['./parser', './parseScheme'], modFactory);
} else {
throw new Error('Use a module loader!');
}
}(function (s, parseScheme) {
function printMixin (env, fn) {
env.print = new s.SchemePrimativeFunction(function (val) {
fn(val.toString());
return val;
});
}
function loadMixin (env, fn) {
env.load = new s.SchemePrimativeFunction(function (filename) {
+ var content, expressions;
+
+ if (fn.length > 0) {
- if (!(filename instanceof s.SchemeString)) {
+ if (!(filename instanceof s.SchemeString)) {
? ++
- throw new s.SchemeError('Expecting ' + s.SchemeString.typeName() + '. ' +
+ throw new s.SchemeError('Expecting ' + s.SchemeString.typeName() + '. ' +
? ++
- 'Found ' + filename.toString());
+ 'Found ' + (filename ? filename.toString() : 'Nothing'));
? ++ ++++++++++++ +++++++++++++
+ }
+ content = fn(filename.val);
+ } else {
+ content = fn();
}
- var content = fn(filename.val),
- expressions = parseScheme.many(content);
? ----
+ expressions = parseScheme.many(content);
return expressions
.get('matched')
.each(function (expr) {
expr.eval(env);
})
.then(function () {
return new s.SchemeBool(true);
});
});
}
return {
printMixin: printMixin,
loadMixin: loadMixin
};
}));
| 16 | 0.372093 | 11 | 5 |
e8f71f64c1ec2a5f386fcd94c701d6ad95c3553c | app/views/work_logs/_form.html.haml | app/views/work_logs/_form.html.haml | = simple_form_for(@work_log) do |f|
= f.error_notification
%fieldset
%legend= content_for?(:title) ? yield(:title) : "Untitled"
.form_inputs
= f.input :start_date, as: :string, input_html: { type: "date" }
= f.input :start_time, as: :string, input_html: { type: "time" }
= f.input :end_date, as: :string, input_html: { type: "date" }
= f.input :end_time, as: :string, input_html: { type: "time" }
= f.input :hours
.form-actions
= f.button :submit
| = simple_form_for(@work_log) do |f|
= f.error_notification
%fieldset
%legend= content_for?(:title) ? yield(:title) : "Untitled"
.form_inputs
= f.input :start_date, as: :string, input_html: { type: "date" }
= f.input :start_time, as: :string, input_html: { type: "time" }
= f.input :end_date, as: :string, input_html: { type: "date" }
= f.input :end_time, as: :string, input_html: { type: "time" }
.form-actions
= f.button :submit
| Remove 'Hours' Field From Work Log Form | Remove 'Hours' Field From Work Log Form
| Haml | agpl-3.0 | Moneyvate/invoices-site,Moneyvate/invoices-site,Moneyvate/invoices-site | haml | ## Code Before:
= simple_form_for(@work_log) do |f|
= f.error_notification
%fieldset
%legend= content_for?(:title) ? yield(:title) : "Untitled"
.form_inputs
= f.input :start_date, as: :string, input_html: { type: "date" }
= f.input :start_time, as: :string, input_html: { type: "time" }
= f.input :end_date, as: :string, input_html: { type: "date" }
= f.input :end_time, as: :string, input_html: { type: "time" }
= f.input :hours
.form-actions
= f.button :submit
## Instruction:
Remove 'Hours' Field From Work Log Form
## Code After:
= simple_form_for(@work_log) do |f|
= f.error_notification
%fieldset
%legend= content_for?(:title) ? yield(:title) : "Untitled"
.form_inputs
= f.input :start_date, as: :string, input_html: { type: "date" }
= f.input :start_time, as: :string, input_html: { type: "time" }
= f.input :end_date, as: :string, input_html: { type: "date" }
= f.input :end_time, as: :string, input_html: { type: "time" }
.form-actions
= f.button :submit
| = simple_form_for(@work_log) do |f|
= f.error_notification
%fieldset
%legend= content_for?(:title) ? yield(:title) : "Untitled"
.form_inputs
= f.input :start_date, as: :string, input_html: { type: "date" }
= f.input :start_time, as: :string, input_html: { type: "time" }
= f.input :end_date, as: :string, input_html: { type: "date" }
= f.input :end_time, as: :string, input_html: { type: "time" }
- = f.input :hours
.form-actions
= f.button :submit | 1 | 0.083333 | 0 | 1 |
795838ce80da218ec5f12d1a9c03c01386547d39 | bench-combine.sh | bench-combine.sh |
mlr --icsv --ocsv --rs lf put '$OutputBits = "32"' HashBenchmark32-report.csv > benchmarks.csv
mlr --icsv --ocsv --rs lf put '$OutputBits = "64"' HashBenchmark64-report.csv | \
xsv cat rows - benchmarks.csv | \
sponge benchmarks.csv
mlr --icsv --ocsv --rs lf put '$OutputBits = "32"' ../BenchmarkDotNet.Artifacts/results/HashBenchmark32-report.csv | \
xsv cat rows - benchmarks.csv | \
sponge benchmarks.csv
mlr --icsv --ocsv --rs lf put '$OutputBits = "64"' ../BenchmarkDotNet.Artifacts/results/HashBenchmark64-report.csv | \
xsv cat rows - benchmarks.csv | \
sponge benchmarks.csv
mlr --icsv --ocsv put '$OutputBits = "32"' net46-results/HashBenchmark32-report.csv | \
xsv cat rows - benchmarks.csv | \
sponge benchmarks.csv
mlr --icsv --ocsv put '$OutputBits = "64"' net46-results/HashBenchmark64-report.csv | \
xsv cat rows - benchmarks.csv | \
sponge benchmarks.csv
|
xsv cat rows core-HashBenchmark32-report.csv core-HashBenchmark64-report.csv \
mono-HashBenchmark32-report.csv mono-HashBenchmark64-report.csv \
clr-HashBenchmark32-report.csv clr-HashBenchmark64-report.csv > benchmarks.csv
| Update the combine benchmark script | Update the combine benchmark script
| Shell | mit | nickbabcock/Farmhash.Sharp,nickbabcock/Farmhash.Sharp,nickbabcock/Farmhash.Sharp,nickbabcock/Farmhash.Sharp | shell | ## Code Before:
mlr --icsv --ocsv --rs lf put '$OutputBits = "32"' HashBenchmark32-report.csv > benchmarks.csv
mlr --icsv --ocsv --rs lf put '$OutputBits = "64"' HashBenchmark64-report.csv | \
xsv cat rows - benchmarks.csv | \
sponge benchmarks.csv
mlr --icsv --ocsv --rs lf put '$OutputBits = "32"' ../BenchmarkDotNet.Artifacts/results/HashBenchmark32-report.csv | \
xsv cat rows - benchmarks.csv | \
sponge benchmarks.csv
mlr --icsv --ocsv --rs lf put '$OutputBits = "64"' ../BenchmarkDotNet.Artifacts/results/HashBenchmark64-report.csv | \
xsv cat rows - benchmarks.csv | \
sponge benchmarks.csv
mlr --icsv --ocsv put '$OutputBits = "32"' net46-results/HashBenchmark32-report.csv | \
xsv cat rows - benchmarks.csv | \
sponge benchmarks.csv
mlr --icsv --ocsv put '$OutputBits = "64"' net46-results/HashBenchmark64-report.csv | \
xsv cat rows - benchmarks.csv | \
sponge benchmarks.csv
## Instruction:
Update the combine benchmark script
## Code After:
xsv cat rows core-HashBenchmark32-report.csv core-HashBenchmark64-report.csv \
mono-HashBenchmark32-report.csv mono-HashBenchmark64-report.csv \
clr-HashBenchmark32-report.csv clr-HashBenchmark64-report.csv > benchmarks.csv
|
+ xsv cat rows core-HashBenchmark32-report.csv core-HashBenchmark64-report.csv \
+ mono-HashBenchmark32-report.csv mono-HashBenchmark64-report.csv \
+ clr-HashBenchmark32-report.csv clr-HashBenchmark64-report.csv > benchmarks.csv
- mlr --icsv --ocsv --rs lf put '$OutputBits = "32"' HashBenchmark32-report.csv > benchmarks.csv
- mlr --icsv --ocsv --rs lf put '$OutputBits = "64"' HashBenchmark64-report.csv | \
- xsv cat rows - benchmarks.csv | \
- sponge benchmarks.csv
-
- mlr --icsv --ocsv --rs lf put '$OutputBits = "32"' ../BenchmarkDotNet.Artifacts/results/HashBenchmark32-report.csv | \
- xsv cat rows - benchmarks.csv | \
- sponge benchmarks.csv
- mlr --icsv --ocsv --rs lf put '$OutputBits = "64"' ../BenchmarkDotNet.Artifacts/results/HashBenchmark64-report.csv | \
- xsv cat rows - benchmarks.csv | \
- sponge benchmarks.csv
-
- mlr --icsv --ocsv put '$OutputBits = "32"' net46-results/HashBenchmark32-report.csv | \
- xsv cat rows - benchmarks.csv | \
- sponge benchmarks.csv
- mlr --icsv --ocsv put '$OutputBits = "64"' net46-results/HashBenchmark64-report.csv | \
- xsv cat rows - benchmarks.csv | \
- sponge benchmarks.csv | 21 | 1.105263 | 3 | 18 |
3df1a8cc5ed5d1e80c7da308d9d66a94702019ec | sipXconfig/web/src/org/sipfoundry/sipxconfig/site/phone/EditPhoneForm.java | sipXconfig/web/src/org/sipfoundry/sipxconfig/site/phone/EditPhoneForm.java | /*
*
*
* Copyright (C) 2004 SIPfoundry Inc.
* Licensed by SIPfoundry under the LGPL license.
*
* Copyright (C) 2004 Pingtel Corp.
* Licensed to SIPfoundry under a Contributor Agreement.
*
* $
*/
package org.sipfoundry.sipxconfig.site.phone;
import org.apache.tapestry.BaseComponent;
import org.apache.tapestry.event.PageEvent;
import org.apache.tapestry.event.PageRenderListener;
import org.apache.tapestry.form.IPropertySelectionModel;
import org.sipfoundry.sipxconfig.components.MapSelectionModel;
import org.sipfoundry.sipxconfig.phone.PhoneContext;
/**
* Editing a phone is reused for creating new phones and editing existing ones
*/
public abstract class EditPhoneForm extends BaseComponent implements PageRenderListener {
/**
* all the phoneModels available to the system ready for UI selection
*/
public abstract void setPhoneSelectionModel(IPropertySelectionModel phoneModels);
public abstract PhoneContext getPhoneContext();
public void pageBeginRender(PageEvent event_) {
PhoneContext phoneContext = getPhoneContext();
setPhoneSelectionModel(new MapSelectionModel(phoneContext.getPhoneFactoryIds()));
}
}
| /*
*
*
* Copyright (C) 2004 SIPfoundry Inc.
* Licensed by SIPfoundry under the LGPL license.
*
* Copyright (C) 2004 Pingtel Corp.
* Licensed to SIPfoundry under a Contributor Agreement.
*
* $
*/
package org.sipfoundry.sipxconfig.site.phone;
import org.apache.tapestry.BaseComponent;
public abstract class EditPhoneForm extends BaseComponent {
}
| FIX BUILD: Resolve subversion evil twin issue, File was not shown as modified however it had been. This file should have been commited as part of r4886 merge | FIX BUILD: Resolve subversion evil twin issue, File was not shown as modified however it had been. This file should have been commited as part of r4886 merge
git-svn-id: f26ccc5efe72c2bd8e1c40f599fe313f2692e4de@4888 a612230a-c5fa-0310-af8b-88eea846685b
| Java | lgpl-2.1 | sipXtapi/sipXtapi-svn-mirror,sipXtapi/sipXtapi-svn-mirror,sipXtapi/sipXtapi-svn-mirror,sipXtapi/sipXtapi-svn-mirror,sipXtapi/sipXtapi-svn-mirror,sipXtapi/sipXtapi-svn-mirror,sipXtapi/sipXtapi-svn-mirror | java | ## Code Before:
/*
*
*
* Copyright (C) 2004 SIPfoundry Inc.
* Licensed by SIPfoundry under the LGPL license.
*
* Copyright (C) 2004 Pingtel Corp.
* Licensed to SIPfoundry under a Contributor Agreement.
*
* $
*/
package org.sipfoundry.sipxconfig.site.phone;
import org.apache.tapestry.BaseComponent;
import org.apache.tapestry.event.PageEvent;
import org.apache.tapestry.event.PageRenderListener;
import org.apache.tapestry.form.IPropertySelectionModel;
import org.sipfoundry.sipxconfig.components.MapSelectionModel;
import org.sipfoundry.sipxconfig.phone.PhoneContext;
/**
* Editing a phone is reused for creating new phones and editing existing ones
*/
public abstract class EditPhoneForm extends BaseComponent implements PageRenderListener {
/**
* all the phoneModels available to the system ready for UI selection
*/
public abstract void setPhoneSelectionModel(IPropertySelectionModel phoneModels);
public abstract PhoneContext getPhoneContext();
public void pageBeginRender(PageEvent event_) {
PhoneContext phoneContext = getPhoneContext();
setPhoneSelectionModel(new MapSelectionModel(phoneContext.getPhoneFactoryIds()));
}
}
## Instruction:
FIX BUILD: Resolve subversion evil twin issue, File was not shown as modified however it had been. This file should have been commited as part of r4886 merge
git-svn-id: f26ccc5efe72c2bd8e1c40f599fe313f2692e4de@4888 a612230a-c5fa-0310-af8b-88eea846685b
## Code After:
/*
*
*
* Copyright (C) 2004 SIPfoundry Inc.
* Licensed by SIPfoundry under the LGPL license.
*
* Copyright (C) 2004 Pingtel Corp.
* Licensed to SIPfoundry under a Contributor Agreement.
*
* $
*/
package org.sipfoundry.sipxconfig.site.phone;
import org.apache.tapestry.BaseComponent;
public abstract class EditPhoneForm extends BaseComponent {
}
| /*
*
*
* Copyright (C) 2004 SIPfoundry Inc.
* Licensed by SIPfoundry under the LGPL license.
*
* Copyright (C) 2004 Pingtel Corp.
* Licensed to SIPfoundry under a Contributor Agreement.
*
* $
*/
package org.sipfoundry.sipxconfig.site.phone;
import org.apache.tapestry.BaseComponent;
- import org.apache.tapestry.event.PageEvent;
- import org.apache.tapestry.event.PageRenderListener;
- import org.apache.tapestry.form.IPropertySelectionModel;
- import org.sipfoundry.sipxconfig.components.MapSelectionModel;
- import org.sipfoundry.sipxconfig.phone.PhoneContext;
- /**
- * Editing a phone is reused for creating new phones and editing existing ones
- */
- public abstract class EditPhoneForm extends BaseComponent implements PageRenderListener {
? ------------------------------
+ public abstract class EditPhoneForm extends BaseComponent {
-
- /**
- * all the phoneModels available to the system ready for UI selection
- */
- public abstract void setPhoneSelectionModel(IPropertySelectionModel phoneModels);
-
? -
+
- public abstract PhoneContext getPhoneContext();
-
- public void pageBeginRender(PageEvent event_) {
- PhoneContext phoneContext = getPhoneContext();
- setPhoneSelectionModel(new MapSelectionModel(phoneContext.getPhoneFactoryIds()));
- }
} | 23 | 0.621622 | 2 | 21 |
5ceed9a936ce7d2a79289a9c400a3292c28137d9 | pcl/meta.yaml | pcl/meta.yaml | package:
name: pcl
version: 1.7.1
source:
fn: pcl-1.7.1.tar.gz
url: https://github.com/PointCloudLibrary/pcl/archive/pcl-1.7.1.tar.gz
patches:
- supervoxel.patch [win]
- gp3_surface.patch [win]
- mls_smoothing.patch [win]
requirements:
build:
- cmake
- eigen3
- flann
- boost <1.56.0
- libpng [unix]
- zlib [unix]
run:
- eigen3
- flann
- boost <1.56.0
- libpng [unix]
- zlib [unix]
about:
home: http://pointclouds.org/
license: BSD
| package:
name: pcl
version: 1.7.2
source:
fn: pcl-1.7.2.tar.gz
url: https://github.com/PointCloudLibrary/pcl/archive/pcl-1.7.2.tar.gz
patches:
- supervoxel.patch [win]
- gp3_surface.patch [win]
- mls_smoothing.patch [win]
requirements:
build:
- cmake
- eigen3
- flann
- boost <1.56.0
- libpng [unix]
- zlib [unix]
- pkg-config [unix]
run:
- eigen3
- flann
- boost <1.56.0
- libpng [unix]
- zlib [unix]
about:
home: http://pointclouds.org/
license: BSD
| Update it to version 1.7.2 to fix compilation error on Mac | PCL: Update it to version 1.7.2 to fix compilation error on Mac
- Also add pkg-config as a build dep after reviewing the Homebrew
formula
| YAML | bsd-2-clause | ccordoba12/pcl-conda-recipes,ccordoba12/pcl-conda-recipes | yaml | ## Code Before:
package:
name: pcl
version: 1.7.1
source:
fn: pcl-1.7.1.tar.gz
url: https://github.com/PointCloudLibrary/pcl/archive/pcl-1.7.1.tar.gz
patches:
- supervoxel.patch [win]
- gp3_surface.patch [win]
- mls_smoothing.patch [win]
requirements:
build:
- cmake
- eigen3
- flann
- boost <1.56.0
- libpng [unix]
- zlib [unix]
run:
- eigen3
- flann
- boost <1.56.0
- libpng [unix]
- zlib [unix]
about:
home: http://pointclouds.org/
license: BSD
## Instruction:
PCL: Update it to version 1.7.2 to fix compilation error on Mac
- Also add pkg-config as a build dep after reviewing the Homebrew
formula
## Code After:
package:
name: pcl
version: 1.7.2
source:
fn: pcl-1.7.2.tar.gz
url: https://github.com/PointCloudLibrary/pcl/archive/pcl-1.7.2.tar.gz
patches:
- supervoxel.patch [win]
- gp3_surface.patch [win]
- mls_smoothing.patch [win]
requirements:
build:
- cmake
- eigen3
- flann
- boost <1.56.0
- libpng [unix]
- zlib [unix]
- pkg-config [unix]
run:
- eigen3
- flann
- boost <1.56.0
- libpng [unix]
- zlib [unix]
about:
home: http://pointclouds.org/
license: BSD
| package:
name: pcl
- version: 1.7.1
? ^
+ version: 1.7.2
? ^
source:
- fn: pcl-1.7.1.tar.gz
? ^
+ fn: pcl-1.7.2.tar.gz
? ^
- url: https://github.com/PointCloudLibrary/pcl/archive/pcl-1.7.1.tar.gz
? ^
+ url: https://github.com/PointCloudLibrary/pcl/archive/pcl-1.7.2.tar.gz
? ^
patches:
- supervoxel.patch [win]
- gp3_surface.patch [win]
- mls_smoothing.patch [win]
requirements:
build:
- cmake
- eigen3
- flann
- boost <1.56.0
- libpng [unix]
- zlib [unix]
+ - pkg-config [unix]
run:
- eigen3
- flann
- boost <1.56.0
- libpng [unix]
- zlib [unix]
about:
home: http://pointclouds.org/
license: BSD | 7 | 0.233333 | 4 | 3 |
b4e78339970f151f04a8167e97c64bc6c8ae5eb7 | Framework/Lumberjack/CocoaLumberjack.h | Framework/Lumberjack/CocoaLumberjack.h | //
// CocoaLumberjack.h
// CocoaLumberjack
//
// Created by Andrew Mackenzie-Ross on 3/02/2015.
//
//
#import <Foundation/Foundation.h>
//! Project version number for CocoaLumberjack.
FOUNDATION_EXPORT double CocoaLumberjackVersionNumber;
//! Project version string for CocoaLumberjack.
FOUNDATION_EXPORT const unsigned char CocoaLumberjackVersionString[];
// In this header, you should import all the public headers of your framework using statements like #import <CocoaLumberjack/PublicHeader.h>
// Disable legacy macros
#ifndef DD_LEGACY_MACROS
#define DD_LEGACY_MACROS 0
#endif
// Core
#import <CocoaLumberjack/DDLog.h>
// Main macros
#import <CocoaLumberjack/DDLogMacros.h>
#import <CocoaLumberjack/DDAssertMacros.h>
// Capture ASL
#import <CocoaLumberjack/DDASLLogCapture.h>
// Loggers
#import <CocoaLumberjack/DDTTYLogger.h>
#import <CocoaLumberjack/DDASLLogger.h>
#import <CocoaLumberjack/DDFileLogger.h>
// CLI
#if defined(DD_CLI) || !__has_include(<AppKit/NSColor.h>)
#import <CocoaLumberjack/CLIColor.h>
#endif
| //
// CocoaLumberjack.h
// CocoaLumberjack
//
// Created by Andrew Mackenzie-Ross on 3/02/2015.
//
//
#import <Foundation/Foundation.h>
//! Project version number for CocoaLumberjack.
FOUNDATION_EXPORT double CocoaLumberjackVersionNumber;
//! Project version string for CocoaLumberjack.
FOUNDATION_EXPORT const unsigned char CocoaLumberjackVersionString[];
// In this header, you should import all the public headers of your framework using statements like #import <CocoaLumberjack/PublicHeader.h>
// Disable legacy macros
#ifndef DD_LEGACY_MACROS
#define DD_LEGACY_MACROS 0
#endif
// Core
#import <CocoaLumberjack/DDLog.h>
// Main macros
#import <CocoaLumberjack/DDLogMacros.h>
#import <CocoaLumberjack/DDAssertMacros.h>
// Capture ASL
#import <CocoaLumberjack/DDASLLogCapture.h>
// Loggers
#import <CocoaLumberjack/DDTTYLogger.h>
#import <CocoaLumberjack/DDASLLogger.h>
#import <CocoaLumberjack/DDFileLogger.h>
// CLI
#import <CocoaLumberjack/CLIColor.h>
| Revert "Conditionally include CLIColor in umbrella header" | Revert "Conditionally include CLIColor in umbrella header"
| C | bsd-3-clause | CocoaLumberjack/CocoaLumberjack,jum/CocoaLumberjack,sushichop/CocoaLumberjack,sushichop/CocoaLumberjack,jum/CocoaLumberjack,CocoaLumberjack/CocoaLumberjack,sushichop/CocoaLumberjack,jum/CocoaLumberjack,DD-P/CocoaLumberjack,CocoaLumberjack/CocoaLumberjack,DD-P/CocoaLumberjack,sushichop/CocoaLumberjack,jum/CocoaLumberjack,DD-P/CocoaLumberjack,CocoaLumberjack/CocoaLumberjack | c | ## Code Before:
//
// CocoaLumberjack.h
// CocoaLumberjack
//
// Created by Andrew Mackenzie-Ross on 3/02/2015.
//
//
#import <Foundation/Foundation.h>
//! Project version number for CocoaLumberjack.
FOUNDATION_EXPORT double CocoaLumberjackVersionNumber;
//! Project version string for CocoaLumberjack.
FOUNDATION_EXPORT const unsigned char CocoaLumberjackVersionString[];
// In this header, you should import all the public headers of your framework using statements like #import <CocoaLumberjack/PublicHeader.h>
// Disable legacy macros
#ifndef DD_LEGACY_MACROS
#define DD_LEGACY_MACROS 0
#endif
// Core
#import <CocoaLumberjack/DDLog.h>
// Main macros
#import <CocoaLumberjack/DDLogMacros.h>
#import <CocoaLumberjack/DDAssertMacros.h>
// Capture ASL
#import <CocoaLumberjack/DDASLLogCapture.h>
// Loggers
#import <CocoaLumberjack/DDTTYLogger.h>
#import <CocoaLumberjack/DDASLLogger.h>
#import <CocoaLumberjack/DDFileLogger.h>
// CLI
#if defined(DD_CLI) || !__has_include(<AppKit/NSColor.h>)
#import <CocoaLumberjack/CLIColor.h>
#endif
## Instruction:
Revert "Conditionally include CLIColor in umbrella header"
## Code After:
//
// CocoaLumberjack.h
// CocoaLumberjack
//
// Created by Andrew Mackenzie-Ross on 3/02/2015.
//
//
#import <Foundation/Foundation.h>
//! Project version number for CocoaLumberjack.
FOUNDATION_EXPORT double CocoaLumberjackVersionNumber;
//! Project version string for CocoaLumberjack.
FOUNDATION_EXPORT const unsigned char CocoaLumberjackVersionString[];
// In this header, you should import all the public headers of your framework using statements like #import <CocoaLumberjack/PublicHeader.h>
// Disable legacy macros
#ifndef DD_LEGACY_MACROS
#define DD_LEGACY_MACROS 0
#endif
// Core
#import <CocoaLumberjack/DDLog.h>
// Main macros
#import <CocoaLumberjack/DDLogMacros.h>
#import <CocoaLumberjack/DDAssertMacros.h>
// Capture ASL
#import <CocoaLumberjack/DDASLLogCapture.h>
// Loggers
#import <CocoaLumberjack/DDTTYLogger.h>
#import <CocoaLumberjack/DDASLLogger.h>
#import <CocoaLumberjack/DDFileLogger.h>
// CLI
#import <CocoaLumberjack/CLIColor.h>
| //
// CocoaLumberjack.h
// CocoaLumberjack
//
// Created by Andrew Mackenzie-Ross on 3/02/2015.
//
//
#import <Foundation/Foundation.h>
//! Project version number for CocoaLumberjack.
FOUNDATION_EXPORT double CocoaLumberjackVersionNumber;
//! Project version string for CocoaLumberjack.
FOUNDATION_EXPORT const unsigned char CocoaLumberjackVersionString[];
// In this header, you should import all the public headers of your framework using statements like #import <CocoaLumberjack/PublicHeader.h>
// Disable legacy macros
#ifndef DD_LEGACY_MACROS
#define DD_LEGACY_MACROS 0
#endif
// Core
#import <CocoaLumberjack/DDLog.h>
// Main macros
#import <CocoaLumberjack/DDLogMacros.h>
#import <CocoaLumberjack/DDAssertMacros.h>
// Capture ASL
#import <CocoaLumberjack/DDASLLogCapture.h>
// Loggers
#import <CocoaLumberjack/DDTTYLogger.h>
#import <CocoaLumberjack/DDASLLogger.h>
#import <CocoaLumberjack/DDFileLogger.h>
// CLI
- #if defined(DD_CLI) || !__has_include(<AppKit/NSColor.h>)
#import <CocoaLumberjack/CLIColor.h>
- #endif | 2 | 0.045455 | 0 | 2 |
513ec2f179e698ff95b74cbea3687a85182d870e | src/get-started/gulp.md | src/get-started/gulp.md | ---
layout: page
title: Get Started > Gulp
---
# Gulp
Use gulp and `workbox-build` to build a precaching service worker. Simply
install the module then cut and paste the code sample.
1. [Install Node.js](https://nodejs.org/en/).
1. Install the module with NPM.
```
npm install workbox-build --save-dev
```
1. Require `workbox-build` in your gulp `gulpfile.js`.
```
const wbBuild = require('workbox-build');
```
1. Also in `gulpfile.js` add a task to build a service worker.
```
gulp.task('generate-sw', () => {
return wbBuild.generateSW({
globDirectory: './app/',
swDest: './app/sw.js',
staticFileGlobs: ['**\/*.{html,js,css}'],
globIgnores: ['admin.html'],
templatedUrls: {
'/shell': ['shell.hbs', 'main.css', 'shell.css'],
},
})
.then(() => {
console.log('Service worker generated.');
})
.catch((err) => {
console.log('[ERROR] This happened: ' + err);
});
})
```
| ---
layout: page
title: Get Started > Gulp
---
# Gulp
Use gulp and `workbox-build` to build a precaching service worker. Simply
install the module then cut and paste the code sample.
1. [Install Node.js](https://nodejs.org/en/).
1. Install the module with NPM.
```
npm install workbox-build --save-dev
```
1. Require `workbox-build` in your gulp `gulpfile.js`.
```
const wbBuild = require('workbox-build');
```
1. Also in `gulpfile.js` add a task to build a service worker.
```
gulp.task('bundle-sw', () => {
return wbBuild.generateSW({
globDirectory: './app/',
swDest: './app/sw.js',
staticFileGlobs: ['**\/*.{html,js,css}'],
globIgnores: ['admin.html'],
templatedUrls: {
'/shell': ['shell.hbs', 'main.css', 'shell.css'],
},
})
.then(() => {
console.log('Service worker generated.');
})
.catch((err) => {
console.log('[ERROR] This happened: ' + err);
});
})
```
**Note:** The gulp task for generating the service worker should always be
run as the last step in each build. This ensures that your service worker
contains any changes made during development. An example is shown below.
```
gulp.task('build:dist', ['clean'], callback => {
process.env.NODE_ENV = 'production';
sequence(
['bundle-app', 'bundle-third-party', 'copy-static', 'sass', 'lint'],
'uglify-js',
'version-assets',
'bundle-sw',
callback
);
});
```
| Add final note about build order. | Add final note about build order.
| Markdown | apache-2.0 | googlearchive/workbox-microsite,googlearchive/workbox-microsite,googlearchive/workbox-microsite | markdown | ## Code Before:
---
layout: page
title: Get Started > Gulp
---
# Gulp
Use gulp and `workbox-build` to build a precaching service worker. Simply
install the module then cut and paste the code sample.
1. [Install Node.js](https://nodejs.org/en/).
1. Install the module with NPM.
```
npm install workbox-build --save-dev
```
1. Require `workbox-build` in your gulp `gulpfile.js`.
```
const wbBuild = require('workbox-build');
```
1. Also in `gulpfile.js` add a task to build a service worker.
```
gulp.task('generate-sw', () => {
return wbBuild.generateSW({
globDirectory: './app/',
swDest: './app/sw.js',
staticFileGlobs: ['**\/*.{html,js,css}'],
globIgnores: ['admin.html'],
templatedUrls: {
'/shell': ['shell.hbs', 'main.css', 'shell.css'],
},
})
.then(() => {
console.log('Service worker generated.');
})
.catch((err) => {
console.log('[ERROR] This happened: ' + err);
});
})
```
## Instruction:
Add final note about build order.
## Code After:
---
layout: page
title: Get Started > Gulp
---
# Gulp
Use gulp and `workbox-build` to build a precaching service worker. Simply
install the module then cut and paste the code sample.
1. [Install Node.js](https://nodejs.org/en/).
1. Install the module with NPM.
```
npm install workbox-build --save-dev
```
1. Require `workbox-build` in your gulp `gulpfile.js`.
```
const wbBuild = require('workbox-build');
```
1. Also in `gulpfile.js` add a task to build a service worker.
```
gulp.task('bundle-sw', () => {
return wbBuild.generateSW({
globDirectory: './app/',
swDest: './app/sw.js',
staticFileGlobs: ['**\/*.{html,js,css}'],
globIgnores: ['admin.html'],
templatedUrls: {
'/shell': ['shell.hbs', 'main.css', 'shell.css'],
},
})
.then(() => {
console.log('Service worker generated.');
})
.catch((err) => {
console.log('[ERROR] This happened: ' + err);
});
})
```
**Note:** The gulp task for generating the service worker should always be
run as the last step in each build. This ensures that your service worker
contains any changes made during development. An example is shown below.
```
gulp.task('build:dist', ['clean'], callback => {
process.env.NODE_ENV = 'production';
sequence(
['bundle-app', 'bundle-third-party', 'copy-static', 'sass', 'lint'],
'uglify-js',
'version-assets',
'bundle-sw',
callback
);
});
```
| ---
layout: page
title: Get Started > Gulp
---
# Gulp
Use gulp and `workbox-build` to build a precaching service worker. Simply
install the module then cut and paste the code sample.
1. [Install Node.js](https://nodejs.org/en/).
1. Install the module with NPM.
```
npm install workbox-build --save-dev
```
1. Require `workbox-build` in your gulp `gulpfile.js`.
```
const wbBuild = require('workbox-build');
```
1. Also in `gulpfile.js` add a task to build a service worker.
```
- gulp.task('generate-sw', () => {
? ^^ ^^^^
+ gulp.task('bundle-sw', () => {
? ^^ ^^
return wbBuild.generateSW({
globDirectory: './app/',
swDest: './app/sw.js',
staticFileGlobs: ['**\/*.{html,js,css}'],
globIgnores: ['admin.html'],
templatedUrls: {
'/shell': ['shell.hbs', 'main.css', 'shell.css'],
},
})
.then(() => {
console.log('Service worker generated.');
})
.catch((err) => {
console.log('[ERROR] This happened: ' + err);
});
})
```
+
+ **Note:** The gulp task for generating the service worker should always be
+ run as the last step in each build. This ensures that your service worker
+ contains any changes made during development. An example is shown below.
+
+
+ ```
+ gulp.task('build:dist', ['clean'], callback => {
+ process.env.NODE_ENV = 'production';
+ sequence(
+ ['bundle-app', 'bundle-third-party', 'copy-static', 'sass', 'lint'],
+ 'uglify-js',
+ 'version-assets',
+ 'bundle-sw',
+ callback
+ );
+ });
+ ``` | 20 | 0.454545 | 19 | 1 |
9658325b454f07c0254824ee649a1f5d9db768ba | bin/crontab/crontab.tpl | bin/crontab/crontab.tpl |
MAILTO=input-tracebacks@mozilla.com
HOME=/tmp
# Once a day at 2:00am run the l10n completion script.
# Note: This runs as root so it has access to the locale/ directory
# to do an svn up.
10 2 * * * root cd {{ source }} && (./bin/run_l10n_completion.sh {{ source }} {{ python }} > /dev/null)
# Once a day at 3:00am run the translation daily activites.
0 3 * * * {{ user }} cd {{ source }} && {{ python }} manage.py translation_daily -v 0 --traceback
# Every hour, sync translations. This pulls and pushes to the various
# translation systems.
0 * * * * {{ user }} cd {{ source }} && {{ python }} manage.py translation_sync -v 0 --traceback
# On Sundays at 3:10am, purge data per our data retention policies.
10 3 * * 0 {{ user }} cd {{ source }} && {{ python }} manage.py purge_data -v 3 --traceback
|
MAILTO=input-tracebacks@mozilla.com
HOME=/tmp
# Once a day at 2:00am run the l10n completion script.
# Note: This runs as root so it has access to the locale/ directory
# to do an svn up.
10 2 * * * root cd {{ source }} && (./bin/run_l10n_completion.sh {{ source }} {{ python }})
# Once a day at 3:00am run the translation daily activites.
0 3 * * * {{ user }} cd {{ source }} && {{ python }} manage.py translation_daily -v 0 --traceback
# Every hour, sync translations. This pulls and pushes to the various
# translation systems.
0 * * * * {{ user }} cd {{ source }} && {{ python }} manage.py translation_sync -v 0 --traceback
# On Sundays at 3:10am, purge data per our data retention policies.
10 3 * * 0 {{ user }} cd {{ source }} && {{ python }} manage.py purge_data -v 3 --traceback
| Stop sending output to /dev/null | Stop sending output to /dev/null
Need to debug the l10n_completion script. While doing that, stop
sending the output to /dev/null.
| Smarty | bsd-3-clause | lgp171188/fjord,hoosteeno/fjord,lgp171188/fjord,mozilla/fjord,mozilla/fjord,mozilla/fjord,hoosteeno/fjord,mozilla/fjord,lgp171188/fjord,hoosteeno/fjord,hoosteeno/fjord,lgp171188/fjord | smarty | ## Code Before:
MAILTO=input-tracebacks@mozilla.com
HOME=/tmp
# Once a day at 2:00am run the l10n completion script.
# Note: This runs as root so it has access to the locale/ directory
# to do an svn up.
10 2 * * * root cd {{ source }} && (./bin/run_l10n_completion.sh {{ source }} {{ python }} > /dev/null)
# Once a day at 3:00am run the translation daily activites.
0 3 * * * {{ user }} cd {{ source }} && {{ python }} manage.py translation_daily -v 0 --traceback
# Every hour, sync translations. This pulls and pushes to the various
# translation systems.
0 * * * * {{ user }} cd {{ source }} && {{ python }} manage.py translation_sync -v 0 --traceback
# On Sundays at 3:10am, purge data per our data retention policies.
10 3 * * 0 {{ user }} cd {{ source }} && {{ python }} manage.py purge_data -v 3 --traceback
## Instruction:
Stop sending output to /dev/null
Need to debug the l10n_completion script. While doing that, stop
sending the output to /dev/null.
## Code After:
MAILTO=input-tracebacks@mozilla.com
HOME=/tmp
# Once a day at 2:00am run the l10n completion script.
# Note: This runs as root so it has access to the locale/ directory
# to do an svn up.
10 2 * * * root cd {{ source }} && (./bin/run_l10n_completion.sh {{ source }} {{ python }})
# Once a day at 3:00am run the translation daily activites.
0 3 * * * {{ user }} cd {{ source }} && {{ python }} manage.py translation_daily -v 0 --traceback
# Every hour, sync translations. This pulls and pushes to the various
# translation systems.
0 * * * * {{ user }} cd {{ source }} && {{ python }} manage.py translation_sync -v 0 --traceback
# On Sundays at 3:10am, purge data per our data retention policies.
10 3 * * 0 {{ user }} cd {{ source }} && {{ python }} manage.py purge_data -v 3 --traceback
|
MAILTO=input-tracebacks@mozilla.com
HOME=/tmp
# Once a day at 2:00am run the l10n completion script.
# Note: This runs as root so it has access to the locale/ directory
# to do an svn up.
- 10 2 * * * root cd {{ source }} && (./bin/run_l10n_completion.sh {{ source }} {{ python }} > /dev/null)
? ------------
+ 10 2 * * * root cd {{ source }} && (./bin/run_l10n_completion.sh {{ source }} {{ python }})
# Once a day at 3:00am run the translation daily activites.
0 3 * * * {{ user }} cd {{ source }} && {{ python }} manage.py translation_daily -v 0 --traceback
# Every hour, sync translations. This pulls and pushes to the various
# translation systems.
0 * * * * {{ user }} cd {{ source }} && {{ python }} manage.py translation_sync -v 0 --traceback
# On Sundays at 3:10am, purge data per our data retention policies.
10 3 * * 0 {{ user }} cd {{ source }} && {{ python }} manage.py purge_data -v 3 --traceback | 2 | 0.105263 | 1 | 1 |
49843ce9db94647d1cfe9ed71c0c07c7e45495c1 | app/controllers/executions_controller.rb | app/controllers/executions_controller.rb | class ExecutionsController < ApplicationController
before_action :verify_login
# POST /executions
def create
unless GalleryConfig.instrumentation.enabled
render json: { message: 'instrumentation disabled' }, status: :forbidden
return
end
nb = Notebook.find_by!(uuid: params[:uuid])
cell = nb.code_cells.find_by!(md5: params[:md5])
success = params[:success].to_bool
@execution = Execution.new(
user: @user,
code_cell: cell,
success: success,
runtime: params[:runtime].to_f
)
@execution.save!
# Not perfect, but try to log a click for each execution of the whole notebook
clickstream('executed notebook') if success && cell.cell_number.zero?
render json: { message: 'execution log accepted' }, status: :ok
end
end
| class ExecutionsController < ApplicationController
before_action :verify_login
# POST /executions
def create
unless GalleryConfig.instrumentation.enabled
render json: { message: 'instrumentation disabled' }, status: :forbidden
return
end
@notebook = Notebook.find_by!(uuid: params[:uuid])
cell = @notebook.code_cells.find_by!(md5: params[:md5])
success = params[:success].to_bool
@execution = Execution.new(
user: @user,
code_cell: cell,
success: success,
runtime: params[:runtime].to_f
)
@execution.save!
# Not perfect, but try to log a click for each execution of the whole notebook
origin = ENV['HTTP_ORIGIN'] || request.headers['HTTP_ORIGIN']
origin.sub!(%r{https?://}, '')
clickstream('executed notebook', tracking: origin) if success && cell.cell_number.zero?
render json: { message: 'execution log accepted' }, status: :ok
end
end
| Fix up execution click log | Fix up execution click log
| Ruby | mit | nbgallery/nbgallery,nbgallery/nbgallery,nbgallery/nbgallery,nbgallery/nbgallery | ruby | ## Code Before:
class ExecutionsController < ApplicationController
before_action :verify_login
# POST /executions
def create
unless GalleryConfig.instrumentation.enabled
render json: { message: 'instrumentation disabled' }, status: :forbidden
return
end
nb = Notebook.find_by!(uuid: params[:uuid])
cell = nb.code_cells.find_by!(md5: params[:md5])
success = params[:success].to_bool
@execution = Execution.new(
user: @user,
code_cell: cell,
success: success,
runtime: params[:runtime].to_f
)
@execution.save!
# Not perfect, but try to log a click for each execution of the whole notebook
clickstream('executed notebook') if success && cell.cell_number.zero?
render json: { message: 'execution log accepted' }, status: :ok
end
end
## Instruction:
Fix up execution click log
## Code After:
class ExecutionsController < ApplicationController
before_action :verify_login
# POST /executions
def create
unless GalleryConfig.instrumentation.enabled
render json: { message: 'instrumentation disabled' }, status: :forbidden
return
end
@notebook = Notebook.find_by!(uuid: params[:uuid])
cell = @notebook.code_cells.find_by!(md5: params[:md5])
success = params[:success].to_bool
@execution = Execution.new(
user: @user,
code_cell: cell,
success: success,
runtime: params[:runtime].to_f
)
@execution.save!
# Not perfect, but try to log a click for each execution of the whole notebook
origin = ENV['HTTP_ORIGIN'] || request.headers['HTTP_ORIGIN']
origin.sub!(%r{https?://}, '')
clickstream('executed notebook', tracking: origin) if success && cell.cell_number.zero?
render json: { message: 'execution log accepted' }, status: :ok
end
end
| class ExecutionsController < ApplicationController
before_action :verify_login
# POST /executions
def create
unless GalleryConfig.instrumentation.enabled
render json: { message: 'instrumentation disabled' }, status: :forbidden
return
end
- nb = Notebook.find_by!(uuid: params[:uuid])
+ @notebook = Notebook.find_by!(uuid: params[:uuid])
? + +++ +++
- cell = nb.code_cells.find_by!(md5: params[:md5])
+ cell = @notebook.code_cells.find_by!(md5: params[:md5])
? + +++ +++
success = params[:success].to_bool
@execution = Execution.new(
user: @user,
code_cell: cell,
success: success,
runtime: params[:runtime].to_f
)
@execution.save!
# Not perfect, but try to log a click for each execution of the whole notebook
+ origin = ENV['HTTP_ORIGIN'] || request.headers['HTTP_ORIGIN']
+ origin.sub!(%r{https?://}, '')
- clickstream('executed notebook') if success && cell.cell_number.zero?
+ clickstream('executed notebook', tracking: origin) if success && cell.cell_number.zero?
? ++++++++++++++++++
render json: { message: 'execution log accepted' }, status: :ok
end
end | 8 | 0.32 | 5 | 3 |
39672853786e9484cae5a8a9878cf003a973b76c | app/services/additional-data.js | app/services/additional-data.js | import Service from '@ember/service';
import Evented from '@ember/object/evented';
export default Service.extend(Evented, {
shownComponents: null,
data: null,
showWindow: false,
addComponent(path) {
if(!this.get('shownComponents'))
this.set('shownComponents', []);
if (!this.get('shownComponents').includes(path)) {
this.get('shownComponents').push(path);
}
},
removeComponent(path) {
if(!this.get('shownComponents'))
return;
var index = this.get('shownComponents').indexOf(path);
if (index !== -1)
this.get('shownComponents').splice(index, 1);
// close everything when no components are left
if (this.get('shownComponents.length') == 0)
this.emptyAdditionalData()
},
emptyAdditionalData() {
this.set('shownComponents', []);
this.set('data', null);
},
closeAdditionalData() {
this.set('showWindow', false);
this.trigger('showWindow');
},
openAdditionalData() {
this.set('showWindow', true);
this.trigger('showWindow');
}
});
| import Service from '@ember/service';
import Evented from '@ember/object/evented';
export default Service.extend(Evented, {
shownComponents: null,
data: null,
showWindow: false,
addComponent(path) {
if(!this.get('shownComponents'))
this.set('shownComponents', []);
if (!this.get('shownComponents').includes(path)) {
this.get('shownComponents').push(path);
}
},
removeComponent(path) {
if(!this.get('shownComponents'))
return;
var index = this.get('shownComponents').indexOf(path);
if (index !== -1)
this.get('shownComponents').splice(index, 1);
// close everything when no components are left
if (this.get('shownComponents.length') == 0)
this.emptyAdditionalData()
},
emptyAdditionalData() {
this.closeAdditionalData();
this.set('shownComponents.length', 0);
this.set('data', null);
},
closeAdditionalData() {
this.set('showWindow', false);
this.trigger('showWindow');
},
openAdditionalData() {
this.set('showWindow', true);
this.trigger('showWindow');
}
});
| Fix bug causing components to not show up in data-selection | Fix bug causing components to not show up in data-selection
| JavaScript | apache-2.0 | ExplorViz/explorviz-ui-frontend,ExplorViz/explorviz-ui-frontend,ExplorViz/explorviz-ui-frontend | javascript | ## Code Before:
import Service from '@ember/service';
import Evented from '@ember/object/evented';
export default Service.extend(Evented, {
shownComponents: null,
data: null,
showWindow: false,
addComponent(path) {
if(!this.get('shownComponents'))
this.set('shownComponents', []);
if (!this.get('shownComponents').includes(path)) {
this.get('shownComponents').push(path);
}
},
removeComponent(path) {
if(!this.get('shownComponents'))
return;
var index = this.get('shownComponents').indexOf(path);
if (index !== -1)
this.get('shownComponents').splice(index, 1);
// close everything when no components are left
if (this.get('shownComponents.length') == 0)
this.emptyAdditionalData()
},
emptyAdditionalData() {
this.set('shownComponents', []);
this.set('data', null);
},
closeAdditionalData() {
this.set('showWindow', false);
this.trigger('showWindow');
},
openAdditionalData() {
this.set('showWindow', true);
this.trigger('showWindow');
}
});
## Instruction:
Fix bug causing components to not show up in data-selection
## Code After:
import Service from '@ember/service';
import Evented from '@ember/object/evented';
export default Service.extend(Evented, {
shownComponents: null,
data: null,
showWindow: false,
addComponent(path) {
if(!this.get('shownComponents'))
this.set('shownComponents', []);
if (!this.get('shownComponents').includes(path)) {
this.get('shownComponents').push(path);
}
},
removeComponent(path) {
if(!this.get('shownComponents'))
return;
var index = this.get('shownComponents').indexOf(path);
if (index !== -1)
this.get('shownComponents').splice(index, 1);
// close everything when no components are left
if (this.get('shownComponents.length') == 0)
this.emptyAdditionalData()
},
emptyAdditionalData() {
this.closeAdditionalData();
this.set('shownComponents.length', 0);
this.set('data', null);
},
closeAdditionalData() {
this.set('showWindow', false);
this.trigger('showWindow');
},
openAdditionalData() {
this.set('showWindow', true);
this.trigger('showWindow');
}
});
| import Service from '@ember/service';
import Evented from '@ember/object/evented';
export default Service.extend(Evented, {
shownComponents: null,
data: null,
showWindow: false,
addComponent(path) {
if(!this.get('shownComponents'))
this.set('shownComponents', []);
if (!this.get('shownComponents').includes(path)) {
this.get('shownComponents').push(path);
}
},
removeComponent(path) {
if(!this.get('shownComponents'))
return;
var index = this.get('shownComponents').indexOf(path);
if (index !== -1)
this.get('shownComponents').splice(index, 1);
// close everything when no components are left
if (this.get('shownComponents.length') == 0)
this.emptyAdditionalData()
},
emptyAdditionalData() {
+ this.closeAdditionalData();
- this.set('shownComponents', []);
? ^^
+ this.set('shownComponents.length', 0);
? +++++++ ^
this.set('data', null);
},
closeAdditionalData() {
this.set('showWindow', false);
this.trigger('showWindow');
},
openAdditionalData() {
this.set('showWindow', true);
this.trigger('showWindow');
}
}); | 3 | 0.06383 | 2 | 1 |
57a774c7dc30f2d0e3f104a6ab09fa1716ce1dd9 | doc/release-notes.md | doc/release-notes.md | Bitcoin ABC version 0.19.9 is now available from:
<https://download.bitcoinabc.org/0.19.9/>
This release includes the following features and fixes:
- Return amounts from `decoderawtransaction` are padded to 8 decimal places.
- Deprecated 'softforks' information from `getblockchaininfo` RPC call, which
had only been reporting on some very old upgrades. To keep this information,
start bitcoind with the '-deprecatedrpc=getblockchaininfo' option.
| Bitcoin ABC version 0.19.9 is now available from:
<https://download.bitcoinabc.org/0.19.9/>
This release includes the following features and fixes:
- Return amounts from `decoderawtransaction` are padded to 8 decimal places.
- Deprecated 'softforks' information from `getblockchaininfo` RPC call, which
had only been reporting on some very old upgrades. To keep this information,
start bitcoind with the '-deprecatedrpc=getblockchaininfo' option.
- A new `-avoidpartialspends` flag has been added (default=false). If enabled,
the wallet will try to spend UTXO's that point at the same destination together.
This is a privacy increase, as there will no longer be cases where a wallet will
inadvertently spend only parts of the coins sent to the same address (note that
if someone were to send coins to that address after it was used, those coins
will still be included in future coin selections).
| Add release notes for -avoidpartialspends | doc: Add release notes for -avoidpartialspends
Summary:
This is a partial backport of Core PR12257 : https://github.com/bitcoin/bitcoin/pull/12257/commits/232f96f5c8a3920c09db92f4dbac2ad7d10ce8cf
Depends on D3394
Test Plan: Read the notes, maybe?
Reviewers: #bitcoin_abc, Fabien
Reviewed By: #bitcoin_abc, Fabien
Differential Revision: https://reviews.bitcoinabc.org/D3398
| Markdown | mit | ftrader-bitcoinabc/bitcoin-abc,ftrader-bitcoinabc/bitcoin-abc,Bitcoin-ABC/bitcoin-abc,cculianu/bitcoin-abc,Bitcoin-ABC/bitcoin-abc,Bitcoin-ABC/bitcoin-abc,ftrader-bitcoinabc/bitcoin-abc,Bitcoin-ABC/bitcoin-abc,Bitcoin-ABC/bitcoin-abc,cculianu/bitcoin-abc,Bitcoin-ABC/bitcoin-abc,cculianu/bitcoin-abc,Bitcoin-ABC/bitcoin-abc,ftrader-bitcoinabc/bitcoin-abc,cculianu/bitcoin-abc,Bitcoin-ABC/bitcoin-abc,ftrader-bitcoinabc/bitcoin-abc,cculianu/bitcoin-abc,cculianu/bitcoin-abc,cculianu/bitcoin-abc,Bitcoin-ABC/bitcoin-abc,ftrader-bitcoinabc/bitcoin-abc,ftrader-bitcoinabc/bitcoin-abc,ftrader-bitcoinabc/bitcoin-abc,Bitcoin-ABC/bitcoin-abc,cculianu/bitcoin-abc,ftrader-bitcoinabc/bitcoin-abc,Bitcoin-ABC/bitcoin-abc,Bitcoin-ABC/bitcoin-abc,ftrader-bitcoinabc/bitcoin-abc,Bitcoin-ABC/bitcoin-abc | markdown | ## Code Before:
Bitcoin ABC version 0.19.9 is now available from:
<https://download.bitcoinabc.org/0.19.9/>
This release includes the following features and fixes:
- Return amounts from `decoderawtransaction` are padded to 8 decimal places.
- Deprecated 'softforks' information from `getblockchaininfo` RPC call, which
had only been reporting on some very old upgrades. To keep this information,
start bitcoind with the '-deprecatedrpc=getblockchaininfo' option.
## Instruction:
doc: Add release notes for -avoidpartialspends
Summary:
This is a partial backport of Core PR12257 : https://github.com/bitcoin/bitcoin/pull/12257/commits/232f96f5c8a3920c09db92f4dbac2ad7d10ce8cf
Depends on D3394
Test Plan: Read the notes, maybe?
Reviewers: #bitcoin_abc, Fabien
Reviewed By: #bitcoin_abc, Fabien
Differential Revision: https://reviews.bitcoinabc.org/D3398
## Code After:
Bitcoin ABC version 0.19.9 is now available from:
<https://download.bitcoinabc.org/0.19.9/>
This release includes the following features and fixes:
- Return amounts from `decoderawtransaction` are padded to 8 decimal places.
- Deprecated 'softforks' information from `getblockchaininfo` RPC call, which
had only been reporting on some very old upgrades. To keep this information,
start bitcoind with the '-deprecatedrpc=getblockchaininfo' option.
- A new `-avoidpartialspends` flag has been added (default=false). If enabled,
the wallet will try to spend UTXO's that point at the same destination together.
This is a privacy increase, as there will no longer be cases where a wallet will
inadvertently spend only parts of the coins sent to the same address (note that
if someone were to send coins to that address after it was used, those coins
will still be included in future coin selections).
| Bitcoin ABC version 0.19.9 is now available from:
<https://download.bitcoinabc.org/0.19.9/>
This release includes the following features and fixes:
- Return amounts from `decoderawtransaction` are padded to 8 decimal places.
- Deprecated 'softforks' information from `getblockchaininfo` RPC call, which
had only been reporting on some very old upgrades. To keep this information,
start bitcoind with the '-deprecatedrpc=getblockchaininfo' option.
+ - A new `-avoidpartialspends` flag has been added (default=false). If enabled,
+ the wallet will try to spend UTXO's that point at the same destination together.
+ This is a privacy increase, as there will no longer be cases where a wallet will
+ inadvertently spend only parts of the coins sent to the same address (note that
+ if someone were to send coins to that address after it was used, those coins
+ will still be included in future coin selections). | 6 | 0.666667 | 6 | 0 |
440becfe243a982779cddcbca8d01ac630f8965b | kie-ml-dist/src/main/java/org/kie/server/swarm/ml/KieServerMain.java | kie-ml-dist/src/main/java/org/kie/server/swarm/ml/KieServerMain.java | package org.kie.server.swarm.ml;
import java.util.Arrays;
import org.kie.server.swarm.AbstractKieServerMain;
import org.wildfly.swarm.Swarm;
import org.wildfly.swarm.jaxrs.JAXRSArchive;
public class KieServerMain extends AbstractKieServerMain {
public static void main(String[] args) throws Exception {
Swarm container = new Swarm();
System.out.println("\tBuilding kie server deployable...");
JAXRSArchive deployment = createDeployment(container);
System.out.println("\tStaring Wildfly Swarm....");
container.start();
System.out.println("\tConfiguring kjars to be auto deployed to server " + Arrays.toString(args));
installKJars(args);
System.out.println("\tDeploying kie server ....");
container.deploy(deployment);
}
}
| package org.kie.server.swarm.ml;
import java.util.Arrays;
import java.util.HashMap;
import org.kie.server.swarm.AbstractKieServerMain;
import org.wildfly.swarm.Swarm;
import org.wildfly.swarm.config.logging.Level;
import org.wildfly.swarm.jaxrs.JAXRSArchive;
import org.wildfly.swarm.logging.LoggingFraction;
public class KieServerMain extends AbstractKieServerMain {
public static void main(String[] args) throws Exception {
Swarm container = new Swarm();
System.out.println("\tBuilding kie server deployable...");
JAXRSArchive deployment = createDeployment(container);
container.fraction(
new LoggingFraction()
.consoleHandler("CONSOLE", c -> {
c.level(Level.INFO);
c.formatter("%d{HH:mm:ss,SSS} %-5p [%c] (%t) %s%e%n");
})
.rootLogger(Level.INFO, "CONSOLE")
);
System.out.println("\tStaring Wildfly Swarm....");
container.start();
System.out.println("\tConfiguring kjars to be auto deployed to server " + Arrays.toString(args));
installKJars(args);
System.out.println("\tDeploying kie server ....");
container.deploy(deployment);
}
}
| Add Logging configuration into Kie-ML server | Add Logging configuration into Kie-ML server
| Java | apache-2.0 | jesuino/kie-ml,jesuino/kie-ml,jesuino/kie-ml | java | ## Code Before:
package org.kie.server.swarm.ml;
import java.util.Arrays;
import org.kie.server.swarm.AbstractKieServerMain;
import org.wildfly.swarm.Swarm;
import org.wildfly.swarm.jaxrs.JAXRSArchive;
public class KieServerMain extends AbstractKieServerMain {
public static void main(String[] args) throws Exception {
Swarm container = new Swarm();
System.out.println("\tBuilding kie server deployable...");
JAXRSArchive deployment = createDeployment(container);
System.out.println("\tStaring Wildfly Swarm....");
container.start();
System.out.println("\tConfiguring kjars to be auto deployed to server " + Arrays.toString(args));
installKJars(args);
System.out.println("\tDeploying kie server ....");
container.deploy(deployment);
}
}
## Instruction:
Add Logging configuration into Kie-ML server
## Code After:
package org.kie.server.swarm.ml;
import java.util.Arrays;
import java.util.HashMap;
import org.kie.server.swarm.AbstractKieServerMain;
import org.wildfly.swarm.Swarm;
import org.wildfly.swarm.config.logging.Level;
import org.wildfly.swarm.jaxrs.JAXRSArchive;
import org.wildfly.swarm.logging.LoggingFraction;
public class KieServerMain extends AbstractKieServerMain {
public static void main(String[] args) throws Exception {
Swarm container = new Swarm();
System.out.println("\tBuilding kie server deployable...");
JAXRSArchive deployment = createDeployment(container);
container.fraction(
new LoggingFraction()
.consoleHandler("CONSOLE", c -> {
c.level(Level.INFO);
c.formatter("%d{HH:mm:ss,SSS} %-5p [%c] (%t) %s%e%n");
})
.rootLogger(Level.INFO, "CONSOLE")
);
System.out.println("\tStaring Wildfly Swarm....");
container.start();
System.out.println("\tConfiguring kjars to be auto deployed to server " + Arrays.toString(args));
installKJars(args);
System.out.println("\tDeploying kie server ....");
container.deploy(deployment);
}
}
| package org.kie.server.swarm.ml;
import java.util.Arrays;
+ import java.util.HashMap;
import org.kie.server.swarm.AbstractKieServerMain;
import org.wildfly.swarm.Swarm;
+ import org.wildfly.swarm.config.logging.Level;
import org.wildfly.swarm.jaxrs.JAXRSArchive;
+ import org.wildfly.swarm.logging.LoggingFraction;
public class KieServerMain extends AbstractKieServerMain {
public static void main(String[] args) throws Exception {
Swarm container = new Swarm();
System.out.println("\tBuilding kie server deployable...");
JAXRSArchive deployment = createDeployment(container);
+
+ container.fraction(
+ new LoggingFraction()
+ .consoleHandler("CONSOLE", c -> {
+ c.level(Level.INFO);
+ c.formatter("%d{HH:mm:ss,SSS} %-5p [%c] (%t) %s%e%n");
+ })
+ .rootLogger(Level.INFO, "CONSOLE")
+ );
System.out.println("\tStaring Wildfly Swarm....");
container.start();
System.out.println("\tConfiguring kjars to be auto deployed to server " + Arrays.toString(args));
installKJars(args);
System.out.println("\tDeploying kie server ....");
container.deploy(deployment);
}
} | 12 | 0.444444 | 12 | 0 |
6749397500421970cbad92f1dddf335beccf73aa | app/assets/javascripts/school_administrator_dashboard/DashRangeButtons.js | app/assets/javascripts/school_administrator_dashboard/DashRangeButtons.js | import React from 'react';
import PropTypes from 'prop-types';
import DashButton from './DashButton';
//Custom all purpose dashboard button
class DashRangeButtons extends React.Component {
constructor(props) {
super(props);
this.state={selectedButton: 'fortyFiveDays'};
}
onClick(filterFunc, button) {
filterFunc();
this.setState({selectedButton: button});
}
render() {
return (
<div className="DashRangeButtons">
Filter:
<DashButton
onClick={() => this.onClick(this.props.fortyFiveDayFilter, 'fortyFiveDays')}
isSelected={this.state.selectedButton === 'fortyFiveDays'}
buttonText='Past 45 Days' />
<DashButton
onClick={() => this.onClick(this.props.ninetyDayFilter, 'ninetyDays')}
isSelected={this.state.selectedButton === 'ninetyDays'}
buttonText='Past 90 Days' />
<DashButton
onClick={() => this.onClick(this.props.schoolYearFilter, 'schoolYear')}
isSelected={this.state.selectedButton === 'schoolYear'}
buttonText='School Year' />
</div>
);
}
}
DashRangeButtons.propTypes = {
schoolYearFilter: PropTypes.func.isRequired,
ninetyDayFilter: PropTypes.func.isRequired,
fortyFiveDayFilter: PropTypes.func.isRequired
};
export default DashRangeButtons;
| import React from 'react';
import PropTypes from 'prop-types';
import DashButton from './DashButton';
//Custom all purpose dashboard button
class DashRangeButtons extends React.Component {
constructor(props) {
super(props);
this.state={selectedButton: 'schoolYear'};
}
onClick(filterFunc, button) {
filterFunc();
this.setState({selectedButton: button});
}
render() {
return (
<div className="DashRangeButtons">
Filter:
<DashButton
onClick={() => this.onClick(this.props.schoolYearFilter, 'schoolYear')}
isSelected={this.state.selectedButton === 'schoolYear'}
buttonText='School Year' />
<DashButton
onClick={() => this.onClick(this.props.ninetyDayFilter, 'ninetyDays')}
isSelected={this.state.selectedButton === 'ninetyDays'}
buttonText='Past 90 Days' />
<DashButton
onClick={() => this.onClick(this.props.fortyFiveDayFilter, 'fortyFiveDays')}
isSelected={this.state.selectedButton === 'fortyFiveDays'}
buttonText='Past 45 Days' />
</div>
);
}
}
DashRangeButtons.propTypes = {
schoolYearFilter: PropTypes.func.isRequired,
ninetyDayFilter: PropTypes.func.isRequired,
fortyFiveDayFilter: PropTypes.func.isRequired
};
export default DashRangeButtons;
| Revert "Reorder dashboard time filter buttons" | Revert "Reorder dashboard time filter buttons"
| JavaScript | mit | studentinsights/studentinsights,studentinsights/studentinsights,studentinsights/studentinsights,studentinsights/studentinsights | javascript | ## Code Before:
import React from 'react';
import PropTypes from 'prop-types';
import DashButton from './DashButton';
//Custom all purpose dashboard button
class DashRangeButtons extends React.Component {
constructor(props) {
super(props);
this.state={selectedButton: 'fortyFiveDays'};
}
onClick(filterFunc, button) {
filterFunc();
this.setState({selectedButton: button});
}
render() {
return (
<div className="DashRangeButtons">
Filter:
<DashButton
onClick={() => this.onClick(this.props.fortyFiveDayFilter, 'fortyFiveDays')}
isSelected={this.state.selectedButton === 'fortyFiveDays'}
buttonText='Past 45 Days' />
<DashButton
onClick={() => this.onClick(this.props.ninetyDayFilter, 'ninetyDays')}
isSelected={this.state.selectedButton === 'ninetyDays'}
buttonText='Past 90 Days' />
<DashButton
onClick={() => this.onClick(this.props.schoolYearFilter, 'schoolYear')}
isSelected={this.state.selectedButton === 'schoolYear'}
buttonText='School Year' />
</div>
);
}
}
DashRangeButtons.propTypes = {
schoolYearFilter: PropTypes.func.isRequired,
ninetyDayFilter: PropTypes.func.isRequired,
fortyFiveDayFilter: PropTypes.func.isRequired
};
export default DashRangeButtons;
## Instruction:
Revert "Reorder dashboard time filter buttons"
## Code After:
import React from 'react';
import PropTypes from 'prop-types';
import DashButton from './DashButton';
//Custom all purpose dashboard button
class DashRangeButtons extends React.Component {
constructor(props) {
super(props);
this.state={selectedButton: 'schoolYear'};
}
onClick(filterFunc, button) {
filterFunc();
this.setState({selectedButton: button});
}
render() {
return (
<div className="DashRangeButtons">
Filter:
<DashButton
onClick={() => this.onClick(this.props.schoolYearFilter, 'schoolYear')}
isSelected={this.state.selectedButton === 'schoolYear'}
buttonText='School Year' />
<DashButton
onClick={() => this.onClick(this.props.ninetyDayFilter, 'ninetyDays')}
isSelected={this.state.selectedButton === 'ninetyDays'}
buttonText='Past 90 Days' />
<DashButton
onClick={() => this.onClick(this.props.fortyFiveDayFilter, 'fortyFiveDays')}
isSelected={this.state.selectedButton === 'fortyFiveDays'}
buttonText='Past 45 Days' />
</div>
);
}
}
DashRangeButtons.propTypes = {
schoolYearFilter: PropTypes.func.isRequired,
ninetyDayFilter: PropTypes.func.isRequired,
fortyFiveDayFilter: PropTypes.func.isRequired
};
export default DashRangeButtons;
| import React from 'react';
import PropTypes from 'prop-types';
import DashButton from './DashButton';
//Custom all purpose dashboard button
class DashRangeButtons extends React.Component {
constructor(props) {
super(props);
- this.state={selectedButton: 'fortyFiveDays'};
? ^ ----------
+ this.state={selectedButton: 'schoolYear'};
? ^^^ +++++
}
onClick(filterFunc, button) {
filterFunc();
this.setState({selectedButton: button});
}
render() {
return (
<div className="DashRangeButtons">
Filter:
<DashButton
- onClick={() => this.onClick(this.props.fortyFiveDayFilter, 'fortyFiveDays')}
- isSelected={this.state.selectedButton === 'fortyFiveDays'}
- buttonText='Past 45 Days' />
- <DashButton
- onClick={() => this.onClick(this.props.ninetyDayFilter, 'ninetyDays')}
- isSelected={this.state.selectedButton === 'ninetyDays'}
- buttonText='Past 90 Days' />
- <DashButton
onClick={() => this.onClick(this.props.schoolYearFilter, 'schoolYear')}
isSelected={this.state.selectedButton === 'schoolYear'}
buttonText='School Year' />
+ <DashButton
+ onClick={() => this.onClick(this.props.ninetyDayFilter, 'ninetyDays')}
+ isSelected={this.state.selectedButton === 'ninetyDays'}
+ buttonText='Past 90 Days' />
+ <DashButton
+ onClick={() => this.onClick(this.props.fortyFiveDayFilter, 'fortyFiveDays')}
+ isSelected={this.state.selectedButton === 'fortyFiveDays'}
+ buttonText='Past 45 Days' />
</div>
);
}
}
DashRangeButtons.propTypes = {
schoolYearFilter: PropTypes.func.isRequired,
ninetyDayFilter: PropTypes.func.isRequired,
fortyFiveDayFilter: PropTypes.func.isRequired
};
export default DashRangeButtons; | 18 | 0.409091 | 9 | 9 |
9b3ca310a0c1192635e5e101eb70c8f2332d0b4e | project.clj | project.clj | (defproject tile-game "1.0.0-SNAPSHOT"
:description "A Tile Puzzle Game and Solver"
:min-lein-version "2.0.0"
:main tile-game.core
:dependencies [[org.clojure/clojure "1.8.0"]
[org.clojure/clojurescript "1.9.494"]
[reagent "0.6.1"]]
:plugins [[lein-figwheel "0.5.9"]
[lein-cljsbuild "1.1.5"]]
:sources-paths ["src"]
:clean-targets ^{:protect false}
["resources/public/js/out"
"resources/public/js/release"
"resources/public/js/tile-game.js"
:target-path]
:cljsbuild
{ :builds [{:id "dev"
:source-paths ["src"]
:figwheel true
:compiler {:main tile-game.grid
:asset-path "js/out"
:output-to "resources/public/js/tile-game.js"
:output-dir "resources/public/js/out"
:source-map-timestamp true}}
{:id "release"
:source-paths ["src"]
:compiler {:main tile-game.grid
:output-to "resources/public/js/tile-game.js"
:output-dir "resources/public/js/release"
:optimizations :advanced
:source-map "resources/public/js/tile-game.js.map"}}]}
:figwheel { :css-dirs ["resources/public/css"]
:open-file-command "emacsclient" })
| (defproject tile-game "1.0.0-SNAPSHOT"
:description "A Tile Puzzle Game and Solver"
:min-lein-version "2.0.0"
:main tile-game.core
:dependencies [[org.clojure/clojure "1.8.0"]
[org.clojure/clojurescript "1.9.494"]
[reagent "0.6.1"]]
:plugins [[lein-figwheel "0.5.9"]
[lein-cljsbuild "1.1.5"]]
:sources-paths ["src"]
:clean-targets ^{:protect false}
["resources/public/js/out"
"resources/public/js/release"
"resources/public/js/tile-game.js"
:target-path]
:cljsbuild {:builds
{"dev"
{:source-paths ["src"]
:figwheel {}
:compiler {:main tile-game.grid
:asset-path "js/out"
:output-to "resources/public/js/tile-game.js"
:output-dir "resources/public/js/out"
:optimizations :none
:source-map-timestamp true}}
"release"
{:source-paths ["src"]
:compiler {:main tile-game.grid
:output-to "resources/public/js/tile-game.js"
:output-dir "resources/public/js/release"
:optimizations :advanced
:source-map "resources/public/js/tile-game.js.map"}}}}
:figwheel {:css-dirs ["resources/public/css"]
:open-file-command "emacsclient"})
| Reformat cljsbuild & fighwheel config and specify more | Reformat cljsbuild & fighwheel config and specify more
| Clojure | mit | dgtized/tile-game,dgtized/tile-game | clojure | ## Code Before:
(defproject tile-game "1.0.0-SNAPSHOT"
:description "A Tile Puzzle Game and Solver"
:min-lein-version "2.0.0"
:main tile-game.core
:dependencies [[org.clojure/clojure "1.8.0"]
[org.clojure/clojurescript "1.9.494"]
[reagent "0.6.1"]]
:plugins [[lein-figwheel "0.5.9"]
[lein-cljsbuild "1.1.5"]]
:sources-paths ["src"]
:clean-targets ^{:protect false}
["resources/public/js/out"
"resources/public/js/release"
"resources/public/js/tile-game.js"
:target-path]
:cljsbuild
{ :builds [{:id "dev"
:source-paths ["src"]
:figwheel true
:compiler {:main tile-game.grid
:asset-path "js/out"
:output-to "resources/public/js/tile-game.js"
:output-dir "resources/public/js/out"
:source-map-timestamp true}}
{:id "release"
:source-paths ["src"]
:compiler {:main tile-game.grid
:output-to "resources/public/js/tile-game.js"
:output-dir "resources/public/js/release"
:optimizations :advanced
:source-map "resources/public/js/tile-game.js.map"}}]}
:figwheel { :css-dirs ["resources/public/css"]
:open-file-command "emacsclient" })
## Instruction:
Reformat cljsbuild & fighwheel config and specify more
## Code After:
(defproject tile-game "1.0.0-SNAPSHOT"
:description "A Tile Puzzle Game and Solver"
:min-lein-version "2.0.0"
:main tile-game.core
:dependencies [[org.clojure/clojure "1.8.0"]
[org.clojure/clojurescript "1.9.494"]
[reagent "0.6.1"]]
:plugins [[lein-figwheel "0.5.9"]
[lein-cljsbuild "1.1.5"]]
:sources-paths ["src"]
:clean-targets ^{:protect false}
["resources/public/js/out"
"resources/public/js/release"
"resources/public/js/tile-game.js"
:target-path]
:cljsbuild {:builds
{"dev"
{:source-paths ["src"]
:figwheel {}
:compiler {:main tile-game.grid
:asset-path "js/out"
:output-to "resources/public/js/tile-game.js"
:output-dir "resources/public/js/out"
:optimizations :none
:source-map-timestamp true}}
"release"
{:source-paths ["src"]
:compiler {:main tile-game.grid
:output-to "resources/public/js/tile-game.js"
:output-dir "resources/public/js/release"
:optimizations :advanced
:source-map "resources/public/js/tile-game.js.map"}}}}
:figwheel {:css-dirs ["resources/public/css"]
:open-file-command "emacsclient"})
| (defproject tile-game "1.0.0-SNAPSHOT"
:description "A Tile Puzzle Game and Solver"
:min-lein-version "2.0.0"
:main tile-game.core
:dependencies [[org.clojure/clojure "1.8.0"]
[org.clojure/clojurescript "1.9.494"]
[reagent "0.6.1"]]
:plugins [[lein-figwheel "0.5.9"]
[lein-cljsbuild "1.1.5"]]
:sources-paths ["src"]
:clean-targets ^{:protect false}
["resources/public/js/out"
"resources/public/js/release"
"resources/public/js/tile-game.js"
:target-path]
- :cljsbuild
- { :builds [{:id "dev"
+ :cljsbuild {:builds
+ {"dev"
- :source-paths ["src"]
+ {:source-paths ["src"]
? ++
- :figwheel true
? ^^^^
+ :figwheel {}
? ++ ^^
- :compiler {:main tile-game.grid
+ :compiler {:main tile-game.grid
? ++
- :asset-path "js/out"
+ :asset-path "js/out"
? ++
- :output-to "resources/public/js/tile-game.js"
+ :output-to "resources/public/js/tile-game.js"
? ++
- :output-dir "resources/public/js/out"
+ :output-dir "resources/public/js/out"
? ++
+ :optimizations :none
- :source-map-timestamp true}}
+ :source-map-timestamp true}}
? ++
- {:id "release"
? ^^^^
+ "release"
? ^
- :source-paths ["src"]
+ {:source-paths ["src"]
? ++
- :compiler {:main tile-game.grid
+ :compiler {:main tile-game.grid
? ++
- :output-to "resources/public/js/tile-game.js"
+ :output-to "resources/public/js/tile-game.js"
? ++
- :output-dir "resources/public/js/release"
+ :output-dir "resources/public/js/release"
? ++
- :optimizations :advanced
+ :optimizations :advanced
? ++
- :source-map "resources/public/js/tile-game.js.map"}}]}
? -
+ :source-map "resources/public/js/tile-game.js.map"}}}}
? ++ +
- :figwheel { :css-dirs ["resources/public/css"]
? -
+ :figwheel {:css-dirs ["resources/public/css"]
- :open-file-command "emacsclient" })
? - -
+ :open-file-command "emacsclient"}) | 37 | 1.121212 | 19 | 18 |
8b4eb312a2588b928f862c22625f22fec86f8ae9 | _config.yml | _config.yml | title: The Virtual Lab
baseurl:
host:
# Build settings
# 'github-pages' gem should automatically pull in the right markdown processor
# As Github pages build support, our site will use 'kramdown' for markdown rendering
markdown: kramdown
# Turn on recognition of GitHub Flavored Markdown
kramdown:
input: GFM
include: ["_pages"]
kramdown:
input: GFM
| title: The Virtual Lab
baseurl:
host:
# Build settings
# 'github-pages' gem should automatically pull in the right markdown processor
# As Github pages build support, our site will use 'kramdown' for markdown rendering
markdown: kramdown
# Turn on recognition of GitHub Flavored Markdown
kramdown:
input: GFM
hard_wrap: false
include: ["_pages"]
| Fix local markdown rednering:turn off hard wrap | Fix local markdown rednering:turn off hard wrap
| YAML | mit | TurkServer/docs,TurkServer/docs | yaml | ## Code Before:
title: The Virtual Lab
baseurl:
host:
# Build settings
# 'github-pages' gem should automatically pull in the right markdown processor
# As Github pages build support, our site will use 'kramdown' for markdown rendering
markdown: kramdown
# Turn on recognition of GitHub Flavored Markdown
kramdown:
input: GFM
include: ["_pages"]
kramdown:
input: GFM
## Instruction:
Fix local markdown rednering:turn off hard wrap
## Code After:
title: The Virtual Lab
baseurl:
host:
# Build settings
# 'github-pages' gem should automatically pull in the right markdown processor
# As Github pages build support, our site will use 'kramdown' for markdown rendering
markdown: kramdown
# Turn on recognition of GitHub Flavored Markdown
kramdown:
input: GFM
hard_wrap: false
include: ["_pages"]
| title: The Virtual Lab
baseurl:
host:
# Build settings
# 'github-pages' gem should automatically pull in the right markdown processor
# As Github pages build support, our site will use 'kramdown' for markdown rendering
markdown: kramdown
# Turn on recognition of GitHub Flavored Markdown
kramdown:
- input: GFM
? -
+ input: GFM
+ hard_wrap: false
include: ["_pages"]
-
- kramdown:
- input: GFM | 6 | 0.352941 | 2 | 4 |
062a0b5e3b666d0f078a9cd0d0641ebb4a7e0594 | metadata/de.csicar.ning.yml | metadata/de.csicar.ning.yml | Categories:
- Development
- Internet
License: LGPL-3.0-only
AuthorName: csicar
SourceCode: https://github.com/csicar/Ning
IssueTracker: https://github.com/csicar/Ning/issues
AutoName: Ning
RepoType: git
Repo: https://github.com/csicar/Ning
Builds:
- versionName: '1.2'
versionCode: 12
commit: v1.2
subdir: app
gradle:
- yes
- versionName: 1.2.3
versionCode: 123
commit: v1.2.3
subdir: app
gradle:
- yes
- versionName: 1.2.4
versionCode: 124
commit: v1.2.4
subdir: app
gradle:
- yes
- versionName: 2.0.0
versionCode: 200
commit: v2.0.0
subdir: app
gradle:
- yes
AutoUpdateMode: Version
UpdateCheckMode: Tags
CurrentVersion: 2.0.0
CurrentVersionCode: 200
| Categories:
- Development
- Internet
License: LGPL-3.0-only
AuthorName: csicar
SourceCode: https://github.com/csicar/Ning
IssueTracker: https://github.com/csicar/Ning/issues
AutoName: Ning
RepoType: git
Repo: https://github.com/csicar/Ning
Builds:
- versionName: '1.2'
versionCode: 12
commit: v1.2
subdir: app
gradle:
- yes
- versionName: 1.2.3
versionCode: 123
commit: v1.2.3
subdir: app
gradle:
- yes
- versionName: 1.2.4
versionCode: 124
commit: v1.2.4
subdir: app
gradle:
- yes
- versionName: 2.0.0
versionCode: 200
commit: v2.0.0
subdir: app
gradle:
- yes
- versionName: 2.0.1
versionCode: 201
commit: fd313c301c5e7ea4e296dd896530e23c6362e2ac
subdir: app
gradle:
- yes
AutoUpdateMode: Version
UpdateCheckMode: Tags
CurrentVersion: 2.0.1
CurrentVersionCode: 201
| Update Ning to 2.0.1 (201) | Update Ning to 2.0.1 (201)
| YAML | agpl-3.0 | f-droid/fdroiddata,f-droid/fdroiddata | yaml | ## Code Before:
Categories:
- Development
- Internet
License: LGPL-3.0-only
AuthorName: csicar
SourceCode: https://github.com/csicar/Ning
IssueTracker: https://github.com/csicar/Ning/issues
AutoName: Ning
RepoType: git
Repo: https://github.com/csicar/Ning
Builds:
- versionName: '1.2'
versionCode: 12
commit: v1.2
subdir: app
gradle:
- yes
- versionName: 1.2.3
versionCode: 123
commit: v1.2.3
subdir: app
gradle:
- yes
- versionName: 1.2.4
versionCode: 124
commit: v1.2.4
subdir: app
gradle:
- yes
- versionName: 2.0.0
versionCode: 200
commit: v2.0.0
subdir: app
gradle:
- yes
AutoUpdateMode: Version
UpdateCheckMode: Tags
CurrentVersion: 2.0.0
CurrentVersionCode: 200
## Instruction:
Update Ning to 2.0.1 (201)
## Code After:
Categories:
- Development
- Internet
License: LGPL-3.0-only
AuthorName: csicar
SourceCode: https://github.com/csicar/Ning
IssueTracker: https://github.com/csicar/Ning/issues
AutoName: Ning
RepoType: git
Repo: https://github.com/csicar/Ning
Builds:
- versionName: '1.2'
versionCode: 12
commit: v1.2
subdir: app
gradle:
- yes
- versionName: 1.2.3
versionCode: 123
commit: v1.2.3
subdir: app
gradle:
- yes
- versionName: 1.2.4
versionCode: 124
commit: v1.2.4
subdir: app
gradle:
- yes
- versionName: 2.0.0
versionCode: 200
commit: v2.0.0
subdir: app
gradle:
- yes
- versionName: 2.0.1
versionCode: 201
commit: fd313c301c5e7ea4e296dd896530e23c6362e2ac
subdir: app
gradle:
- yes
AutoUpdateMode: Version
UpdateCheckMode: Tags
CurrentVersion: 2.0.1
CurrentVersionCode: 201
| Categories:
- Development
- Internet
License: LGPL-3.0-only
AuthorName: csicar
SourceCode: https://github.com/csicar/Ning
IssueTracker: https://github.com/csicar/Ning/issues
AutoName: Ning
RepoType: git
Repo: https://github.com/csicar/Ning
Builds:
- versionName: '1.2'
versionCode: 12
commit: v1.2
subdir: app
gradle:
- yes
- versionName: 1.2.3
versionCode: 123
commit: v1.2.3
subdir: app
gradle:
- yes
- versionName: 1.2.4
versionCode: 124
commit: v1.2.4
subdir: app
gradle:
- yes
- versionName: 2.0.0
versionCode: 200
commit: v2.0.0
subdir: app
gradle:
- yes
+ - versionName: 2.0.1
+ versionCode: 201
+ commit: fd313c301c5e7ea4e296dd896530e23c6362e2ac
+ subdir: app
+ gradle:
+ - yes
+
AutoUpdateMode: Version
UpdateCheckMode: Tags
- CurrentVersion: 2.0.0
? ^
+ CurrentVersion: 2.0.1
? ^
- CurrentVersionCode: 200
? ^
+ CurrentVersionCode: 201
? ^
| 11 | 0.23913 | 9 | 2 |
864631113fa8421f6c1042cee105923edc59257b | lib/com/vaadin/polymer/elemental/NodeList.java | lib/com/vaadin/polymer/elemental/NodeList.java | package com.vaadin.polymer.elemental;
import com.google.gwt.core.client.js.JsProperty;
import com.google.gwt.core.client.js.JsType;
@JsType
public interface NodeList<T> {
@JsProperty
int getLength();
T item(int index);
}
| package com.vaadin.polymer.elemental;
import com.google.gwt.core.client.js.JsProperty;
import com.google.gwt.core.client.js.JsType;
@JsType
public interface NodeList {
@JsProperty
int getLength();
<T extends Node> T item(int index);
}
| Move generic from class to method | Move generic from class to method
| Java | apache-2.0 | vaadin/gwt-api-generator,florian-f/gwt-api-generator,manolo/gwt-api-generator,manolo/gwt-api-generator,vaadin/gwt-api-generator,florian-f/gwt-api-generator | java | ## Code Before:
package com.vaadin.polymer.elemental;
import com.google.gwt.core.client.js.JsProperty;
import com.google.gwt.core.client.js.JsType;
@JsType
public interface NodeList<T> {
@JsProperty
int getLength();
T item(int index);
}
## Instruction:
Move generic from class to method
## Code After:
package com.vaadin.polymer.elemental;
import com.google.gwt.core.client.js.JsProperty;
import com.google.gwt.core.client.js.JsType;
@JsType
public interface NodeList {
@JsProperty
int getLength();
<T extends Node> T item(int index);
}
| package com.vaadin.polymer.elemental;
import com.google.gwt.core.client.js.JsProperty;
import com.google.gwt.core.client.js.JsType;
@JsType
- public interface NodeList<T> {
? ---
+ public interface NodeList {
@JsProperty
int getLength();
- T item(int index);
+ <T extends Node> T item(int index);
} | 4 | 0.307692 | 2 | 2 |
b5d3daba9195e07f7faf4ce7f9106f27b8c0a20a | tools/llvm-mc/CMakeLists.txt | tools/llvm-mc/CMakeLists.txt | set(LLVM_LINK_COMPONENTS ${LLVM_TARGETS_TO_BUILD} support MC)
add_llvm_tool(llvm-mc
llvm-mc.cpp
AsmLexer.cpp
AsmParser.cpp
)
| set(LLVM_LINK_COMPONENTS ${LLVM_TARGETS_TO_BUILD} support MC)
add_llvm_tool(llvm-mc
llvm-mc.cpp
AsmLexer.cpp
AsmParser.cpp
HexDisassembler.cpp
)
| Update CMake build to include HexDisassembler.cpp. | Update CMake build to include HexDisassembler.cpp.
git-svn-id: 0ff597fd157e6f4fc38580e8d64ab130330d2411@91589 91177308-0d34-0410-b5e6-96231b3b80d8
| Text | bsd-2-clause | dslab-epfl/asap,llvm-mirror/llvm,llvm-mirror/llvm,chubbymaggie/asap,dslab-epfl/asap,dslab-epfl/asap,dslab-epfl/asap,apple/swift-llvm,apple/swift-llvm,apple/swift-llvm,llvm-mirror/llvm,dslab-epfl/asap,apple/swift-llvm,llvm-mirror/llvm,apple/swift-llvm,GPUOpen-Drivers/llvm,chubbymaggie/asap,GPUOpen-Drivers/llvm,llvm-mirror/llvm,dslab-epfl/asap,llvm-mirror/llvm,llvm-mirror/llvm,apple/swift-llvm,chubbymaggie/asap,GPUOpen-Drivers/llvm,llvm-mirror/llvm,GPUOpen-Drivers/llvm,apple/swift-llvm,chubbymaggie/asap,chubbymaggie/asap,GPUOpen-Drivers/llvm,GPUOpen-Drivers/llvm,apple/swift-llvm,GPUOpen-Drivers/llvm,llvm-mirror/llvm,GPUOpen-Drivers/llvm,dslab-epfl/asap,chubbymaggie/asap | text | ## Code Before:
set(LLVM_LINK_COMPONENTS ${LLVM_TARGETS_TO_BUILD} support MC)
add_llvm_tool(llvm-mc
llvm-mc.cpp
AsmLexer.cpp
AsmParser.cpp
)
## Instruction:
Update CMake build to include HexDisassembler.cpp.
git-svn-id: 0ff597fd157e6f4fc38580e8d64ab130330d2411@91589 91177308-0d34-0410-b5e6-96231b3b80d8
## Code After:
set(LLVM_LINK_COMPONENTS ${LLVM_TARGETS_TO_BUILD} support MC)
add_llvm_tool(llvm-mc
llvm-mc.cpp
AsmLexer.cpp
AsmParser.cpp
HexDisassembler.cpp
)
| set(LLVM_LINK_COMPONENTS ${LLVM_TARGETS_TO_BUILD} support MC)
add_llvm_tool(llvm-mc
llvm-mc.cpp
AsmLexer.cpp
AsmParser.cpp
+ HexDisassembler.cpp
) | 1 | 0.142857 | 1 | 0 |
2f03beadb27a1193f6d74fe48b892ba2384eda4d | test/test_slugizer.rb | test/test_slugizer.rb | require 'helper'
class TestSlugizer < Test::Unit::TestCase
context "working with slugs" do
setup do
BlogPost.delete_all
@blogpost = BlogPost.create(:title => "%bLog pOSt tiTLe!",
:body => "HeRe is tHe Body of the bLog pOsT")
end
should "generate the slug" do
@blogpost.slug.should =~ /\w+-blog-post-title/
end
should "return the slug as param" do
@blogpost.to_param =~ /\w+-blog-post-title/
end
should "return the id if slug was not generated" do
@blogpost.slug = nil
@blogpost.to_param.should == @blogpost.id
end
end
context "finding objects" do
setup do
BlogPost.delete_all
@blogpost = BlogPost.create(:title => "%bLog pOSt tiTLe!",
:body => "HeRe is tHe Body of the bLog pOsT")
end
should "be able to find by slug" do
BlogPost.by_slug(@blogpost.slug).should == @blogpost
end
should "be able to find by id" do
BlogPost.by_slug(@blogpost.id).should == @blogpost
end
end
end
| require 'helper'
class TestSlugizer < Test::Unit::TestCase
context "working with slugs" do
setup do
BlogPost.delete_all
@blogpost = BlogPost.create(:title => "%bLog pOSt tiTLe!",
:body => "HeRe is tHe Body of the bLog pOsT")
end
should "generate the slug" do
@blogpost.slug.should =~ /\w+-blog-post-title/
end
should "not generate the slug if the slug key is blank" do
@empty_blogpost = BlogPost.new
@empty_blogpost.slug.should be_nil
end
should "return the slug as param" do
@blogpost.to_param =~ /\w+-blog-post-title/
end
should "return the id if slug was not generated" do
@blogpost.slug = nil
@blogpost.to_param.should == @blogpost.id
end
end
context "finding objects" do
setup do
BlogPost.delete_all
@blogpost = BlogPost.create(:title => "%bLog pOSt tiTLe!",
:body => "HeRe is tHe Body of the bLog pOsT")
end
should "be able to find by slug" do
BlogPost.by_slug(@blogpost.slug).should == @blogpost
end
should "be able to find by id" do
BlogPost.by_slug(@blogpost.id).should == @blogpost
end
end
end
| Test to ensure that slugs aren't generated when slug_key values are blank | Test to ensure that slugs aren't generated when slug_key values are blank
| Ruby | mit | dcu/mongomapper_ext,dcu/mongomapper_ext | ruby | ## Code Before:
require 'helper'
class TestSlugizer < Test::Unit::TestCase
context "working with slugs" do
setup do
BlogPost.delete_all
@blogpost = BlogPost.create(:title => "%bLog pOSt tiTLe!",
:body => "HeRe is tHe Body of the bLog pOsT")
end
should "generate the slug" do
@blogpost.slug.should =~ /\w+-blog-post-title/
end
should "return the slug as param" do
@blogpost.to_param =~ /\w+-blog-post-title/
end
should "return the id if slug was not generated" do
@blogpost.slug = nil
@blogpost.to_param.should == @blogpost.id
end
end
context "finding objects" do
setup do
BlogPost.delete_all
@blogpost = BlogPost.create(:title => "%bLog pOSt tiTLe!",
:body => "HeRe is tHe Body of the bLog pOsT")
end
should "be able to find by slug" do
BlogPost.by_slug(@blogpost.slug).should == @blogpost
end
should "be able to find by id" do
BlogPost.by_slug(@blogpost.id).should == @blogpost
end
end
end
## Instruction:
Test to ensure that slugs aren't generated when slug_key values are blank
## Code After:
require 'helper'
class TestSlugizer < Test::Unit::TestCase
context "working with slugs" do
setup do
BlogPost.delete_all
@blogpost = BlogPost.create(:title => "%bLog pOSt tiTLe!",
:body => "HeRe is tHe Body of the bLog pOsT")
end
should "generate the slug" do
@blogpost.slug.should =~ /\w+-blog-post-title/
end
should "not generate the slug if the slug key is blank" do
@empty_blogpost = BlogPost.new
@empty_blogpost.slug.should be_nil
end
should "return the slug as param" do
@blogpost.to_param =~ /\w+-blog-post-title/
end
should "return the id if slug was not generated" do
@blogpost.slug = nil
@blogpost.to_param.should == @blogpost.id
end
end
context "finding objects" do
setup do
BlogPost.delete_all
@blogpost = BlogPost.create(:title => "%bLog pOSt tiTLe!",
:body => "HeRe is tHe Body of the bLog pOsT")
end
should "be able to find by slug" do
BlogPost.by_slug(@blogpost.slug).should == @blogpost
end
should "be able to find by id" do
BlogPost.by_slug(@blogpost.id).should == @blogpost
end
end
end
| require 'helper'
class TestSlugizer < Test::Unit::TestCase
context "working with slugs" do
setup do
BlogPost.delete_all
@blogpost = BlogPost.create(:title => "%bLog pOSt tiTLe!",
:body => "HeRe is tHe Body of the bLog pOsT")
end
should "generate the slug" do
@blogpost.slug.should =~ /\w+-blog-post-title/
+ end
+
+ should "not generate the slug if the slug key is blank" do
+ @empty_blogpost = BlogPost.new
+ @empty_blogpost.slug.should be_nil
end
should "return the slug as param" do
@blogpost.to_param =~ /\w+-blog-post-title/
end
should "return the id if slug was not generated" do
@blogpost.slug = nil
@blogpost.to_param.should == @blogpost.id
end
end
context "finding objects" do
setup do
BlogPost.delete_all
@blogpost = BlogPost.create(:title => "%bLog pOSt tiTLe!",
:body => "HeRe is tHe Body of the bLog pOsT")
end
should "be able to find by slug" do
BlogPost.by_slug(@blogpost.slug).should == @blogpost
end
should "be able to find by id" do
BlogPost.by_slug(@blogpost.id).should == @blogpost
end
end
end | 5 | 0.125 | 5 | 0 |
f89beada75edaa710968426bc207ec4b6778267c | res/layout/issue_edit_fab.xml | res/layout/issue_edit_fab.xml | <?xml version="1.0" encoding="utf-8"?>
<android.support.design.widget.FloatingActionButton
xmlns:android="http://schemas.android.com/apk/res/android"
xmlns:app="http://schemas.android.com/apk/res-auto"
android:id="@+id/edit_fab"
android:layout_width="wrap_content"
android:layout_height="wrap_content"
android:layout_marginRight="16dp"
android:src="@drawable/content_edit_white"
android:visibility="gone"
app:backgroundTint="?attr/colorIssueEditFab"
app:layout_anchor="@id/header"
app:layout_anchorGravity="right|bottom"
app:fabSize="mini" />
| <?xml version="1.0" encoding="utf-8"?>
<android.support.design.widget.FloatingActionButton
xmlns:android="http://schemas.android.com/apk/res/android"
xmlns:app="http://schemas.android.com/apk/res-auto"
android:id="@+id/edit_fab"
android:layout_width="wrap_content"
android:layout_height="wrap_content"
android:layout_marginRight="16dp"
android:src="@drawable/content_edit_white"
app:backgroundTint="?attr/colorIssueEditFab"
app:layout_anchor="@id/header"
app:layout_anchorGravity="right|bottom"
app:fabSize="mini" />
| Fix issue edit FAB visibility. | Fix issue edit FAB visibility.
| XML | apache-2.0 | slapperwan/gh4a,slapperwan/gh4a,edyesed/gh4a,edyesed/gh4a | xml | ## Code Before:
<?xml version="1.0" encoding="utf-8"?>
<android.support.design.widget.FloatingActionButton
xmlns:android="http://schemas.android.com/apk/res/android"
xmlns:app="http://schemas.android.com/apk/res-auto"
android:id="@+id/edit_fab"
android:layout_width="wrap_content"
android:layout_height="wrap_content"
android:layout_marginRight="16dp"
android:src="@drawable/content_edit_white"
android:visibility="gone"
app:backgroundTint="?attr/colorIssueEditFab"
app:layout_anchor="@id/header"
app:layout_anchorGravity="right|bottom"
app:fabSize="mini" />
## Instruction:
Fix issue edit FAB visibility.
## Code After:
<?xml version="1.0" encoding="utf-8"?>
<android.support.design.widget.FloatingActionButton
xmlns:android="http://schemas.android.com/apk/res/android"
xmlns:app="http://schemas.android.com/apk/res-auto"
android:id="@+id/edit_fab"
android:layout_width="wrap_content"
android:layout_height="wrap_content"
android:layout_marginRight="16dp"
android:src="@drawable/content_edit_white"
app:backgroundTint="?attr/colorIssueEditFab"
app:layout_anchor="@id/header"
app:layout_anchorGravity="right|bottom"
app:fabSize="mini" />
| <?xml version="1.0" encoding="utf-8"?>
<android.support.design.widget.FloatingActionButton
xmlns:android="http://schemas.android.com/apk/res/android"
xmlns:app="http://schemas.android.com/apk/res-auto"
android:id="@+id/edit_fab"
android:layout_width="wrap_content"
android:layout_height="wrap_content"
android:layout_marginRight="16dp"
android:src="@drawable/content_edit_white"
- android:visibility="gone"
app:backgroundTint="?attr/colorIssueEditFab"
app:layout_anchor="@id/header"
app:layout_anchorGravity="right|bottom"
app:fabSize="mini" /> | 1 | 0.071429 | 0 | 1 |
277381393cd5ab9bc0051895529cafd61baaabdb | _config.yml | _config.yml | gems: [jekyll-paginate]
# Site settings
title: Bhavik Kumar
email: contact@bhavik.io
description: > # this means to ignore newlines until "baseurl:"
<b>Bhavik</b> is a software engineer. He started his career by
working for one of New Zealand's largest IT consultancies. Now
he works for a leading payments company building new digital
products.
baseurl: "" # the subpath of your site, e.g. /blog
url: "https://bhavik.io" # the base hostname & protocol for your site
# twitter_username: bhavikkumar
github_username: bhavikkumar
google_analytics: UA-78148514-1
# Build settings
markdown: kramdown
# Pagination settings
paginate: 10
| gems: [jekyll-paginate]
# Site settings
title: Bhavik Kumar
email: contact@bhavik.io
description: > # this means to ignore newlines until "baseurl:"
<b>Bhavik</b> is a software engineer. He started his career by
working for one of New Zealand's largest IT consultancies. Now
he works for a leading payments company building new digital
products.
baseurl: "" # the subpath of your site, e.g. /blog
url: "https://bhavik.io" # the base hostname & protocol for your site
# twitter_username: bhavikkumar
github_username: bhavikkumar
google_analytics: UA-78148514-1
# Build settings
markdown: kramdown
# Pagination settings
paginate: 10
# Headers
webrick:
headers:
Content-Security-Policy: default-src https://fonts.gstatic.com 'self' 'unsafe-inline' script-src 'self' 'unsafe-inline' style-src https://fonts.googleapis.com 'self'
X-Frame-Options: "x-frame-options: SAMEORIGIN"
X-XSS-Protection: "X-XSS-Protection: 1; mode=block"
| Add headers to WEBrick for local testing | Add headers to WEBrick for local testing
| YAML | mit | bhavikkumar/bhavik.io | yaml | ## Code Before:
gems: [jekyll-paginate]
# Site settings
title: Bhavik Kumar
email: contact@bhavik.io
description: > # this means to ignore newlines until "baseurl:"
<b>Bhavik</b> is a software engineer. He started his career by
working for one of New Zealand's largest IT consultancies. Now
he works for a leading payments company building new digital
products.
baseurl: "" # the subpath of your site, e.g. /blog
url: "https://bhavik.io" # the base hostname & protocol for your site
# twitter_username: bhavikkumar
github_username: bhavikkumar
google_analytics: UA-78148514-1
# Build settings
markdown: kramdown
# Pagination settings
paginate: 10
## Instruction:
Add headers to WEBrick for local testing
## Code After:
gems: [jekyll-paginate]
# Site settings
title: Bhavik Kumar
email: contact@bhavik.io
description: > # this means to ignore newlines until "baseurl:"
<b>Bhavik</b> is a software engineer. He started his career by
working for one of New Zealand's largest IT consultancies. Now
he works for a leading payments company building new digital
products.
baseurl: "" # the subpath of your site, e.g. /blog
url: "https://bhavik.io" # the base hostname & protocol for your site
# twitter_username: bhavikkumar
github_username: bhavikkumar
google_analytics: UA-78148514-1
# Build settings
markdown: kramdown
# Pagination settings
paginate: 10
# Headers
webrick:
headers:
Content-Security-Policy: default-src https://fonts.gstatic.com 'self' 'unsafe-inline' script-src 'self' 'unsafe-inline' style-src https://fonts.googleapis.com 'self'
X-Frame-Options: "x-frame-options: SAMEORIGIN"
X-XSS-Protection: "X-XSS-Protection: 1; mode=block"
| gems: [jekyll-paginate]
# Site settings
title: Bhavik Kumar
email: contact@bhavik.io
description: > # this means to ignore newlines until "baseurl:"
<b>Bhavik</b> is a software engineer. He started his career by
working for one of New Zealand's largest IT consultancies. Now
- he works for a leading payments company building new digital
? -
+ he works for a leading payments company building new digital
products.
baseurl: "" # the subpath of your site, e.g. /blog
url: "https://bhavik.io" # the base hostname & protocol for your site
# twitter_username: bhavikkumar
github_username: bhavikkumar
google_analytics: UA-78148514-1
# Build settings
markdown: kramdown
# Pagination settings
paginate: 10
+
+ # Headers
+ webrick:
+ headers:
+ Content-Security-Policy: default-src https://fonts.gstatic.com 'self' 'unsafe-inline' script-src 'self' 'unsafe-inline' style-src https://fonts.googleapis.com 'self'
+ X-Frame-Options: "x-frame-options: SAMEORIGIN"
+ X-XSS-Protection: "X-XSS-Protection: 1; mode=block" | 9 | 0.428571 | 8 | 1 |
7461303aa5e2072a090a2145a4d2ff3d06f3a496 | data/assets/examples/slidedeck.asset | data/assets/examples/slidedeck.asset | local helper = asset.require('util/slide_deck_helper')
local deck = nil
asset.onInitialize(function ()
deck = helper.createDeck("example", {
UseRadiusAzimuthElevation = true,
RadiusAzimuthElevation = {1.0, 0.0, 0.0}, -- use for dome
UsePerspectiveProjection = true,
FaceCamera = true,
Scale = 0.7
})
helper.addSlide(deck, "${DATA}/test2.jpg")
helper.addSlide(deck, "${DATA}/test3.jpg")
local interpolationDuration = 0.5
function nextSlide()
helper.goToNextSlide(deck, interpolationDuration)
end
function previousSlide()
helper.goToPreviousSlide(deck, interpolationDuration)
end
function toggleSlides()
helper.toggleSlides(deck, interpolationDuration)
end
helper.setCurrentSlide(deck, 1)
openspace.bindKey("KP_6", "nextSlide()", "Next slide", "Next slide", "/Slides")
openspace.bindKey("KP_4", "previousSlide()", "Previous slide", "Previous slide", "/Slides")
openspace.bindKey("KP_0", "toggleSlides()", "Toggle slides", "Toggle slides", "/Slides")
end)
asset.onDeinitialize(function()
openspace.clearKey("KP_6")
openspace.clearKey("KP_4")
openspace.clearKey("KP_0")
helper.removeDeck(deck)
end)
| local helper = asset.require('util/slide_deck_helper')
local deck = nil
asset.onInitialize(function ()
deck = helper.createDeck("example", {
UseRadiusAzimuthElevation = true,
RadiusAzimuthElevation = {1.0, 0.0, 0.0}, -- use for dome
UsePerspectiveProjection = true,
FaceCamera = true,
Scale = 0.7
})
helper.addSlide(deck, "${DATA}/test2.jpg")
helper.addSlide(deck, "${DATA}/test3.jpg")
local interpolationDuration = 0.5
-- Add global functions for controlling slide deck and bind to keys
rawset(_G, "nextSlide", function()
helper.goToNextSlide(deck, interpolationDuration)
end)
rawset(_G, "previousSlide", function()
helper.goToPreviousSlide(deck, interpolationDuration)
end)
rawset(_G, "toggleSlides", function()
helper.toggleSlides(deck, interpolationDuration)
end)
helper.setCurrentSlide(deck, 1)
openspace.bindKey("KP_6", "nextSlide()", "Next slide", "Next slide", "/Slides")
openspace.bindKey("KP_4", "previousSlide()", "Previous slide", "Previous slide", "/Slides")
openspace.bindKey("KP_0", "toggleSlides()", "Toggle slides", "Toggle slides", "/Slides")
end)
asset.onDeinitialize(function()
openspace.clearKey("KP_6")
openspace.clearKey("KP_4")
openspace.clearKey("KP_0")
helper.removeDeck(deck)
end)
| Fix to make slide deck work with new strict Lua | Fix to make slide deck work with new strict Lua
| Unity3D Asset | mit | OpenSpace/OpenSpace,OpenSpace/OpenSpace,OpenSpace/OpenSpace,OpenSpace/OpenSpace | unity3d-asset | ## Code Before:
local helper = asset.require('util/slide_deck_helper')
local deck = nil
asset.onInitialize(function ()
deck = helper.createDeck("example", {
UseRadiusAzimuthElevation = true,
RadiusAzimuthElevation = {1.0, 0.0, 0.0}, -- use for dome
UsePerspectiveProjection = true,
FaceCamera = true,
Scale = 0.7
})
helper.addSlide(deck, "${DATA}/test2.jpg")
helper.addSlide(deck, "${DATA}/test3.jpg")
local interpolationDuration = 0.5
function nextSlide()
helper.goToNextSlide(deck, interpolationDuration)
end
function previousSlide()
helper.goToPreviousSlide(deck, interpolationDuration)
end
function toggleSlides()
helper.toggleSlides(deck, interpolationDuration)
end
helper.setCurrentSlide(deck, 1)
openspace.bindKey("KP_6", "nextSlide()", "Next slide", "Next slide", "/Slides")
openspace.bindKey("KP_4", "previousSlide()", "Previous slide", "Previous slide", "/Slides")
openspace.bindKey("KP_0", "toggleSlides()", "Toggle slides", "Toggle slides", "/Slides")
end)
asset.onDeinitialize(function()
openspace.clearKey("KP_6")
openspace.clearKey("KP_4")
openspace.clearKey("KP_0")
helper.removeDeck(deck)
end)
## Instruction:
Fix to make slide deck work with new strict Lua
## Code After:
local helper = asset.require('util/slide_deck_helper')
local deck = nil
asset.onInitialize(function ()
deck = helper.createDeck("example", {
UseRadiusAzimuthElevation = true,
RadiusAzimuthElevation = {1.0, 0.0, 0.0}, -- use for dome
UsePerspectiveProjection = true,
FaceCamera = true,
Scale = 0.7
})
helper.addSlide(deck, "${DATA}/test2.jpg")
helper.addSlide(deck, "${DATA}/test3.jpg")
local interpolationDuration = 0.5
-- Add global functions for controlling slide deck and bind to keys
rawset(_G, "nextSlide", function()
helper.goToNextSlide(deck, interpolationDuration)
end)
rawset(_G, "previousSlide", function()
helper.goToPreviousSlide(deck, interpolationDuration)
end)
rawset(_G, "toggleSlides", function()
helper.toggleSlides(deck, interpolationDuration)
end)
helper.setCurrentSlide(deck, 1)
openspace.bindKey("KP_6", "nextSlide()", "Next slide", "Next slide", "/Slides")
openspace.bindKey("KP_4", "previousSlide()", "Previous slide", "Previous slide", "/Slides")
openspace.bindKey("KP_0", "toggleSlides()", "Toggle slides", "Toggle slides", "/Slides")
end)
asset.onDeinitialize(function()
openspace.clearKey("KP_6")
openspace.clearKey("KP_4")
openspace.clearKey("KP_0")
helper.removeDeck(deck)
end)
| local helper = asset.require('util/slide_deck_helper')
local deck = nil
asset.onInitialize(function ()
deck = helper.createDeck("example", {
UseRadiusAzimuthElevation = true,
RadiusAzimuthElevation = {1.0, 0.0, 0.0}, -- use for dome
UsePerspectiveProjection = true,
FaceCamera = true,
Scale = 0.7
})
helper.addSlide(deck, "${DATA}/test2.jpg")
helper.addSlide(deck, "${DATA}/test3.jpg")
local interpolationDuration = 0.5
- function nextSlide()
+ -- Add global functions for controlling slide deck and bind to keys
+ rawset(_G, "nextSlide", function()
- helper.goToNextSlide(deck, interpolationDuration)
+ helper.goToNextSlide(deck, interpolationDuration)
? +
- end
+ end)
? +
- function previousSlide()
+ rawset(_G, "previousSlide", function()
helper.goToPreviousSlide(deck, interpolationDuration)
- end
+ end)
? +
- function toggleSlides()
+ rawset(_G, "toggleSlides", function()
helper.toggleSlides(deck, interpolationDuration)
- end
+ end)
? +
helper.setCurrentSlide(deck, 1)
openspace.bindKey("KP_6", "nextSlide()", "Next slide", "Next slide", "/Slides")
openspace.bindKey("KP_4", "previousSlide()", "Previous slide", "Previous slide", "/Slides")
openspace.bindKey("KP_0", "toggleSlides()", "Toggle slides", "Toggle slides", "/Slides")
end)
asset.onDeinitialize(function()
openspace.clearKey("KP_6")
openspace.clearKey("KP_4")
openspace.clearKey("KP_0")
helper.removeDeck(deck)
end) | 15 | 0.357143 | 8 | 7 |
309cbc3fc23134aa987dd1c62ca617d9852685a8 | app/assets/stylesheets/govuk-component/_document-footer.scss | app/assets/stylesheets/govuk-component/_document-footer.scss | .govuk-document-footer {
@extend %grid-row;
@include core-16;
margin-top: $gutter*1.5;
&.direction-rtl {
direction: rtl;
text-align: start;
}
.history-information {
min-height: 1em;
@include grid-column( 1 / 3 );
}
&.direction-rtl .history-information {
@include grid-column( 1 / 3, $float: right );
}
.related-information {
@include grid-column( 2 / 3 );
}
&.direction-rtl .related-information {
@include grid-column( 2 / 3, $float: right );
}
p, ol {
margin-bottom: $gutter-one-third;
@include media(tablet){
margin-bottom: $gutter-two-thirds;
}
}
.definition {
@include bold-24;
margin-top: 5px;
display: block;
a {
text-decoration: none;
}
}
.change-notes {
.timestamp {
display: block;
font-weight: bold;
}
li {
list-style: none;
margin-bottom: 5px;
@include media(tablet){
margin-bottom: $gutter-one-third;
}
}
}
}
| .govuk-document-footer {
@extend %grid-row;
@include core-16;
margin-top: $gutter * 1.5;
margin-bottom: $gutter * 1.5;
&.direction-rtl {
direction: rtl;
text-align: start;
}
.history-information {
min-height: 1em;
@include grid-column( 1 / 3 );
}
&.direction-rtl .history-information {
@include grid-column( 1 / 3, $float: right );
}
.related-information {
@include grid-column( 2 / 3 );
}
&.direction-rtl .related-information {
@include grid-column( 2 / 3, $float: right );
}
p, ol {
margin-bottom: $gutter-one-third;
@include media(tablet){
margin-bottom: $gutter-two-thirds;
}
}
.definition {
@include bold-24;
margin-top: 5px;
display: block;
a {
text-decoration: none;
}
}
.change-notes {
.timestamp {
display: block;
font-weight: bold;
}
li {
list-style: none;
margin-bottom: 5px;
@include media(tablet){
margin-bottom: $gutter-one-third;
}
}
}
}
| Add default bottom margin to document footer | Add default bottom margin to document footer
Document footers appear in a consistent place on a page. Add a default
bottom margin so that apps do not have to add manual styles to separate
it from the “report a problem” link.
Currently only used by specialist-frontend and government-frontend.
| SCSS | mit | alphagov/static,alphagov/static,alphagov/static,kalleth/static,kalleth/static,kalleth/static,kalleth/static | scss | ## Code Before:
.govuk-document-footer {
@extend %grid-row;
@include core-16;
margin-top: $gutter*1.5;
&.direction-rtl {
direction: rtl;
text-align: start;
}
.history-information {
min-height: 1em;
@include grid-column( 1 / 3 );
}
&.direction-rtl .history-information {
@include grid-column( 1 / 3, $float: right );
}
.related-information {
@include grid-column( 2 / 3 );
}
&.direction-rtl .related-information {
@include grid-column( 2 / 3, $float: right );
}
p, ol {
margin-bottom: $gutter-one-third;
@include media(tablet){
margin-bottom: $gutter-two-thirds;
}
}
.definition {
@include bold-24;
margin-top: 5px;
display: block;
a {
text-decoration: none;
}
}
.change-notes {
.timestamp {
display: block;
font-weight: bold;
}
li {
list-style: none;
margin-bottom: 5px;
@include media(tablet){
margin-bottom: $gutter-one-third;
}
}
}
}
## Instruction:
Add default bottom margin to document footer
Document footers appear in a consistent place on a page. Add a default
bottom margin so that apps do not have to add manual styles to separate
it from the “report a problem” link.
Currently only used by specialist-frontend and government-frontend.
## Code After:
.govuk-document-footer {
@extend %grid-row;
@include core-16;
margin-top: $gutter * 1.5;
margin-bottom: $gutter * 1.5;
&.direction-rtl {
direction: rtl;
text-align: start;
}
.history-information {
min-height: 1em;
@include grid-column( 1 / 3 );
}
&.direction-rtl .history-information {
@include grid-column( 1 / 3, $float: right );
}
.related-information {
@include grid-column( 2 / 3 );
}
&.direction-rtl .related-information {
@include grid-column( 2 / 3, $float: right );
}
p, ol {
margin-bottom: $gutter-one-third;
@include media(tablet){
margin-bottom: $gutter-two-thirds;
}
}
.definition {
@include bold-24;
margin-top: 5px;
display: block;
a {
text-decoration: none;
}
}
.change-notes {
.timestamp {
display: block;
font-weight: bold;
}
li {
list-style: none;
margin-bottom: 5px;
@include media(tablet){
margin-bottom: $gutter-one-third;
}
}
}
}
| .govuk-document-footer {
@extend %grid-row;
@include core-16;
- margin-top: $gutter*1.5;
+ margin-top: $gutter * 1.5;
? + +
+ margin-bottom: $gutter * 1.5;
&.direction-rtl {
direction: rtl;
text-align: start;
}
.history-information {
min-height: 1em;
@include grid-column( 1 / 3 );
}
&.direction-rtl .history-information {
@include grid-column( 1 / 3, $float: right );
}
.related-information {
@include grid-column( 2 / 3 );
}
&.direction-rtl .related-information {
@include grid-column( 2 / 3, $float: right );
}
p, ol {
margin-bottom: $gutter-one-third;
@include media(tablet){
margin-bottom: $gutter-two-thirds;
}
}
.definition {
@include bold-24;
margin-top: 5px;
display: block;
a {
text-decoration: none;
}
}
.change-notes {
.timestamp {
display: block;
font-weight: bold;
}
li {
list-style: none;
margin-bottom: 5px;
@include media(tablet){
margin-bottom: $gutter-one-third;
}
}
}
} | 3 | 0.052632 | 2 | 1 |
4e1b1049d2b07d625a5473fbde365987f69bf5ab | lib/core/window.js | lib/core/window.js | /* See license.txt for terms of usage */
"use strict";
const { Trace, TraceError } = require("../core/trace.js").get(module.id);
const { once } = require("sdk/dom/events.js");
const { getMostRecentBrowserWindow } = require("sdk/window/utils");
const { openTab } = require("sdk/tabs/utils");
// Module implementation
var Win = {};
/**
* Returns a promise that is resolved as soon as the window is loaded
* or immediately if the window is already loaded.
*
* @param {Window} The window object we need use when loaded.
*/
Win.loaded = win => new Promise(resolve => {
if (win.document.readyState === "complete") {
resolve(win.document);
}
else {
once(win, "load", event => resolve(event.target));
}
});
// xxxHonza: we might want to merge with Win.loaded
Win.domContentLoaded = win => new Promise(resolve => {
if (win.document.readyState === "complete") {
resolve(win.document);
}
else {
once(win, "DOMContentLoaded", event => resolve(event.target));
}
});
Win.openNewTab = function(url, options) {
let browser = getMostRecentBrowserWindow();
openTab(browser, url, options);
}
// Exports from this module
exports.Win = Win;
| /* See license.txt for terms of usage */
"use strict";
const { Cu } = require("chrome");
const { Trace, TraceError } = require("../core/trace.js").get(module.id);
const { once } = require("sdk/dom/events.js");
const { getMostRecentBrowserWindow } = require("sdk/window/utils");
const { openTab } = require("sdk/tabs/utils");
const { devtools } = Cu.import("resource://gre/modules/devtools/Loader.jsm", {});
const { makeInfallible } = devtools["require"]("devtools/toolkit/DevToolsUtils.js");
// Module implementation
var Win = {};
/**
* Returns a promise that is resolved as soon as the window is loaded
* or immediately if the window is already loaded.
*
* @param {Window} The window object we need use when loaded.
*/
Win.loaded = makeInfallible(win => new Promise(resolve => {
if (win.document.readyState === "complete") {
resolve(win.document);
}
else {
once(win, "load", event => resolve(event.target));
}
}));
// xxxHonza: we might want to merge with Win.loaded
Win.domContentLoaded = makeInfallible(win => new Promise(resolve => {
if (win.document.readyState === "complete") {
resolve(win.document);
}
else {
once(win, "DOMContentLoaded", event => resolve(event.target));
}
}));
Win.openNewTab = function(url, options) {
let browser = getMostRecentBrowserWindow();
openTab(browser, url, options);
}
// Exports from this module
exports.Win = Win;
| Use make infallible for win load handlers | Use make infallible for win load handlers
| JavaScript | bsd-3-clause | vlajos/firebug.next,firebug/firebug.next,bmdeveloper/firebug.next,firebug/firebug.next,bmdeveloper/firebug.next,vlajos/firebug.next | javascript | ## Code Before:
/* See license.txt for terms of usage */
"use strict";
const { Trace, TraceError } = require("../core/trace.js").get(module.id);
const { once } = require("sdk/dom/events.js");
const { getMostRecentBrowserWindow } = require("sdk/window/utils");
const { openTab } = require("sdk/tabs/utils");
// Module implementation
var Win = {};
/**
* Returns a promise that is resolved as soon as the window is loaded
* or immediately if the window is already loaded.
*
* @param {Window} The window object we need use when loaded.
*/
Win.loaded = win => new Promise(resolve => {
if (win.document.readyState === "complete") {
resolve(win.document);
}
else {
once(win, "load", event => resolve(event.target));
}
});
// xxxHonza: we might want to merge with Win.loaded
Win.domContentLoaded = win => new Promise(resolve => {
if (win.document.readyState === "complete") {
resolve(win.document);
}
else {
once(win, "DOMContentLoaded", event => resolve(event.target));
}
});
Win.openNewTab = function(url, options) {
let browser = getMostRecentBrowserWindow();
openTab(browser, url, options);
}
// Exports from this module
exports.Win = Win;
## Instruction:
Use make infallible for win load handlers
## Code After:
/* See license.txt for terms of usage */
"use strict";
const { Cu } = require("chrome");
const { Trace, TraceError } = require("../core/trace.js").get(module.id);
const { once } = require("sdk/dom/events.js");
const { getMostRecentBrowserWindow } = require("sdk/window/utils");
const { openTab } = require("sdk/tabs/utils");
const { devtools } = Cu.import("resource://gre/modules/devtools/Loader.jsm", {});
const { makeInfallible } = devtools["require"]("devtools/toolkit/DevToolsUtils.js");
// Module implementation
var Win = {};
/**
* Returns a promise that is resolved as soon as the window is loaded
* or immediately if the window is already loaded.
*
* @param {Window} The window object we need use when loaded.
*/
Win.loaded = makeInfallible(win => new Promise(resolve => {
if (win.document.readyState === "complete") {
resolve(win.document);
}
else {
once(win, "load", event => resolve(event.target));
}
}));
// xxxHonza: we might want to merge with Win.loaded
Win.domContentLoaded = makeInfallible(win => new Promise(resolve => {
if (win.document.readyState === "complete") {
resolve(win.document);
}
else {
once(win, "DOMContentLoaded", event => resolve(event.target));
}
}));
Win.openNewTab = function(url, options) {
let browser = getMostRecentBrowserWindow();
openTab(browser, url, options);
}
// Exports from this module
exports.Win = Win;
| /* See license.txt for terms of usage */
"use strict";
+
+ const { Cu } = require("chrome");
const { Trace, TraceError } = require("../core/trace.js").get(module.id);
const { once } = require("sdk/dom/events.js");
const { getMostRecentBrowserWindow } = require("sdk/window/utils");
const { openTab } = require("sdk/tabs/utils");
+
+ const { devtools } = Cu.import("resource://gre/modules/devtools/Loader.jsm", {});
+ const { makeInfallible } = devtools["require"]("devtools/toolkit/DevToolsUtils.js");
// Module implementation
var Win = {};
/**
* Returns a promise that is resolved as soon as the window is loaded
* or immediately if the window is already loaded.
*
* @param {Window} The window object we need use when loaded.
*/
- Win.loaded = win => new Promise(resolve => {
+ Win.loaded = makeInfallible(win => new Promise(resolve => {
? +++++++++++++++
if (win.document.readyState === "complete") {
resolve(win.document);
}
else {
once(win, "load", event => resolve(event.target));
}
- });
+ }));
? +
// xxxHonza: we might want to merge with Win.loaded
- Win.domContentLoaded = win => new Promise(resolve => {
+ Win.domContentLoaded = makeInfallible(win => new Promise(resolve => {
? +++++++++++++++
if (win.document.readyState === "complete") {
resolve(win.document);
}
else {
once(win, "DOMContentLoaded", event => resolve(event.target));
}
- });
+ }));
? +
Win.openNewTab = function(url, options) {
let browser = getMostRecentBrowserWindow();
openTab(browser, url, options);
}
// Exports from this module
exports.Win = Win; | 13 | 0.295455 | 9 | 4 |
88e309b62d5e2e0b3a9d661c663058589cbfbc6e | lib/cloudwatch_rails.rb | lib/cloudwatch_rails.rb | require 'cloudwatch_rails/version'
require 'cloudwatch_rails/railtie' if defined?(Rails)
require 'cloudwatch_rails/config'
require 'cloudwatch_rails/async_queue' if defined?(Rails)
module CloudwatchRails
class CollectorConfig
attr_accessor :collector, :queue
end
class << self
attr_accessor :config
def start
unless @config
@config = Config.new
end
end
def transactions
@transactions ||= {}
end
def collector_config
@collector_config ||= CloudwatchRails::CollectorConfig.new
end
def config
@config ||= CloudwatchRails::Config.new
end
end
end
CloudwatchRails.collector_config.queue = CloudwatchRails::AsyncQueue.new
| require 'cloudwatch_rails/version'
require 'cloudwatch_rails/railtie' if defined?(Rails)
require 'cloudwatch_rails/config'
require 'cloudwatch_rails/async_queue'
module CloudwatchRails
class CollectorConfig
attr_accessor :collector, :queue
end
class << self
attr_accessor :config
def start
unless @config
@config = Config.new
end
end
def transactions
@transactions ||= {}
end
def collector_config
@collector_config ||= CloudwatchRails::CollectorConfig.new
end
def config
@config ||= CloudwatchRails::Config.new
end
end
end
CloudwatchRails.collector_config.queue = CloudwatchRails::AsyncQueue.new
| Fix require for Async Queue | Fix require for Async Queue
| Ruby | mit | nickrobinson/cloudwatch_rails,nickrobinson/cloudwatch_rails | ruby | ## Code Before:
require 'cloudwatch_rails/version'
require 'cloudwatch_rails/railtie' if defined?(Rails)
require 'cloudwatch_rails/config'
require 'cloudwatch_rails/async_queue' if defined?(Rails)
module CloudwatchRails
class CollectorConfig
attr_accessor :collector, :queue
end
class << self
attr_accessor :config
def start
unless @config
@config = Config.new
end
end
def transactions
@transactions ||= {}
end
def collector_config
@collector_config ||= CloudwatchRails::CollectorConfig.new
end
def config
@config ||= CloudwatchRails::Config.new
end
end
end
CloudwatchRails.collector_config.queue = CloudwatchRails::AsyncQueue.new
## Instruction:
Fix require for Async Queue
## Code After:
require 'cloudwatch_rails/version'
require 'cloudwatch_rails/railtie' if defined?(Rails)
require 'cloudwatch_rails/config'
require 'cloudwatch_rails/async_queue'
module CloudwatchRails
class CollectorConfig
attr_accessor :collector, :queue
end
class << self
attr_accessor :config
def start
unless @config
@config = Config.new
end
end
def transactions
@transactions ||= {}
end
def collector_config
@collector_config ||= CloudwatchRails::CollectorConfig.new
end
def config
@config ||= CloudwatchRails::Config.new
end
end
end
CloudwatchRails.collector_config.queue = CloudwatchRails::AsyncQueue.new
| require 'cloudwatch_rails/version'
require 'cloudwatch_rails/railtie' if defined?(Rails)
require 'cloudwatch_rails/config'
- require 'cloudwatch_rails/async_queue' if defined?(Rails)
? -------------------
+ require 'cloudwatch_rails/async_queue'
module CloudwatchRails
class CollectorConfig
attr_accessor :collector, :queue
end
class << self
attr_accessor :config
def start
unless @config
@config = Config.new
end
end
def transactions
@transactions ||= {}
end
def collector_config
@collector_config ||= CloudwatchRails::CollectorConfig.new
end
def config
@config ||= CloudwatchRails::Config.new
end
end
end
CloudwatchRails.collector_config.queue = CloudwatchRails::AsyncQueue.new | 2 | 0.057143 | 1 | 1 |
9dd6dfac24fc2b5c6b509ed7bcf365654711e7e1 | app/src/test/java/com/kickstarter/libs/utils/StringUtilsTest.java | app/src/test/java/com/kickstarter/libs/utils/StringUtilsTest.java | package com.kickstarter.libs.utils;
import junit.framework.TestCase;
public class StringUtilsTest extends TestCase {
public void testIsEmail() {
assertTrue(StringUtils.isEmail("hello@kickstarter.com"));
}
public void testIsEmpty() {
assertTrue(StringUtils.isEmpty(""));
assertTrue(StringUtils.isEmpty(" "));
assertTrue(StringUtils.isEmpty(" "));
assertTrue(StringUtils.isEmpty(null));
assertFalse(StringUtils.isEmpty("a"));
assertFalse(StringUtils.isEmpty(" a "));
}
public void testIsPresent() {
assertFalse(StringUtils.isPresent(""));
assertFalse(StringUtils.isPresent(" "));
assertFalse(StringUtils.isPresent(" "));
assertFalse(StringUtils.isPresent(null));
assertTrue(StringUtils.isPresent("a"));
assertTrue(StringUtils.isPresent(" a "));
}
}
| package com.kickstarter.libs.utils;
import com.kickstarter.KSRobolectricTestCase;
import org.junit.Test;
public class StringUtilsTest extends KSRobolectricTestCase {
@Test
public void testIsEmail() {
assertTrue(StringUtils.isEmail("hello@kickstarter.com"));
assertFalse(StringUtils.isEmail("hello@kickstarter"));
}
@Test
public void testIsEmpty() {
assertTrue(StringUtils.isEmpty(""));
assertTrue(StringUtils.isEmpty(" "));
assertTrue(StringUtils.isEmpty(" "));
assertTrue(StringUtils.isEmpty(null));
assertFalse(StringUtils.isEmpty("a"));
assertFalse(StringUtils.isEmpty(" a "));
}
@Test
public void testIsPresent() {
assertFalse(StringUtils.isPresent(""));
assertFalse(StringUtils.isPresent(" "));
assertFalse(StringUtils.isPresent(" "));
assertFalse(StringUtils.isPresent(null));
assertTrue(StringUtils.isPresent("a"));
assertTrue(StringUtils.isPresent(" a "));
}
}
| Make into roboelectric test so pattern is available. | Make into roboelectric test so pattern is available.
| Java | apache-2.0 | kickstarter/android-oss,kickstarter/android-oss,kickstarter/android-oss,kickstarter/android-oss | java | ## Code Before:
package com.kickstarter.libs.utils;
import junit.framework.TestCase;
public class StringUtilsTest extends TestCase {
public void testIsEmail() {
assertTrue(StringUtils.isEmail("hello@kickstarter.com"));
}
public void testIsEmpty() {
assertTrue(StringUtils.isEmpty(""));
assertTrue(StringUtils.isEmpty(" "));
assertTrue(StringUtils.isEmpty(" "));
assertTrue(StringUtils.isEmpty(null));
assertFalse(StringUtils.isEmpty("a"));
assertFalse(StringUtils.isEmpty(" a "));
}
public void testIsPresent() {
assertFalse(StringUtils.isPresent(""));
assertFalse(StringUtils.isPresent(" "));
assertFalse(StringUtils.isPresent(" "));
assertFalse(StringUtils.isPresent(null));
assertTrue(StringUtils.isPresent("a"));
assertTrue(StringUtils.isPresent(" a "));
}
}
## Instruction:
Make into roboelectric test so pattern is available.
## Code After:
package com.kickstarter.libs.utils;
import com.kickstarter.KSRobolectricTestCase;
import org.junit.Test;
public class StringUtilsTest extends KSRobolectricTestCase {
@Test
public void testIsEmail() {
assertTrue(StringUtils.isEmail("hello@kickstarter.com"));
assertFalse(StringUtils.isEmail("hello@kickstarter"));
}
@Test
public void testIsEmpty() {
assertTrue(StringUtils.isEmpty(""));
assertTrue(StringUtils.isEmpty(" "));
assertTrue(StringUtils.isEmpty(" "));
assertTrue(StringUtils.isEmpty(null));
assertFalse(StringUtils.isEmpty("a"));
assertFalse(StringUtils.isEmpty(" a "));
}
@Test
public void testIsPresent() {
assertFalse(StringUtils.isPresent(""));
assertFalse(StringUtils.isPresent(" "));
assertFalse(StringUtils.isPresent(" "));
assertFalse(StringUtils.isPresent(null));
assertTrue(StringUtils.isPresent("a"));
assertTrue(StringUtils.isPresent(" a "));
}
}
| package com.kickstarter.libs.utils;
- import junit.framework.TestCase;
+ import com.kickstarter.KSRobolectricTestCase;
- public class StringUtilsTest extends TestCase {
+ import org.junit.Test;
+ public class StringUtilsTest extends KSRobolectricTestCase {
+
+ @Test
public void testIsEmail() {
assertTrue(StringUtils.isEmail("hello@kickstarter.com"));
+ assertFalse(StringUtils.isEmail("hello@kickstarter"));
}
+ @Test
public void testIsEmpty() {
assertTrue(StringUtils.isEmpty(""));
assertTrue(StringUtils.isEmpty(" "));
assertTrue(StringUtils.isEmpty(" "));
assertTrue(StringUtils.isEmpty(null));
assertFalse(StringUtils.isEmpty("a"));
assertFalse(StringUtils.isEmpty(" a "));
}
+ @Test
public void testIsPresent() {
assertFalse(StringUtils.isPresent(""));
assertFalse(StringUtils.isPresent(" "));
assertFalse(StringUtils.isPresent(" "));
assertFalse(StringUtils.isPresent(null));
assertTrue(StringUtils.isPresent("a"));
assertTrue(StringUtils.isPresent(" a "));
}
} | 10 | 0.357143 | 8 | 2 |
db851838b616c6070996070771daf74ea6e73764 | app/views/images/_new.html.erb | app/views/images/_new.html.erb | <div class="new-image">
<%= cloudinary_js_config %>
<%= form_for :image, id: "upload-form", url: album_images_path, html: { multipart: true } do |f| %>
<fieldset>
<legend>Plop an Image</legend>
<%= render "shared/errors" %>
<%= f.file_field(:image_url) %>
<label for="direct_url">Image Url</label>
<input type="text" name="image[direct_url]" id="direct_url">
<%= f.label :caption %>
<%= f.text_field :caption %>
<%= f.submit "Upload" %>
<% end %>
<% if @album.tag %>
<%= button_to "Search Instagram for Images with Tag: #{@album.tag}", search_path(@album.tag), method: :get %>
<% end %>
</fieldset>
</div>
| <div class="new-image">
<%= cloudinary_js_config %>
<%= form_for :image, id: "upload-form", url: album_images_path, html: { multipart: true } do |f| %>
<fieldset>
<legend>Plop an Image</legend>
<%= render "shared/errors" %>
<%= f.file_field(:image_url) %>
<label for="direct_url">Image Url</label>
<input type="text" name="image[direct_url]" id="direct_url">
<%= f.label :caption %>
<%= f.text_field :caption %>
<%= f.submit "Upload" %>
<% end %>
<% if @album.tag %>
<%= button_to "Search Instagram for Images with Your Tag", search_path(@album.tag), method: :get %>
<% end %>
</fieldset>
</div>
| Fix fieldset going off of page - due to instagram button | Fix fieldset going off of page - due to instagram button
| HTML+ERB | mit | chi-fiddler-crabs-2015/Photo-Plop,chi-fiddler-crabs-2015/Photo-Plop,chi-fiddler-crabs-2015/Photo-Plop | html+erb | ## Code Before:
<div class="new-image">
<%= cloudinary_js_config %>
<%= form_for :image, id: "upload-form", url: album_images_path, html: { multipart: true } do |f| %>
<fieldset>
<legend>Plop an Image</legend>
<%= render "shared/errors" %>
<%= f.file_field(:image_url) %>
<label for="direct_url">Image Url</label>
<input type="text" name="image[direct_url]" id="direct_url">
<%= f.label :caption %>
<%= f.text_field :caption %>
<%= f.submit "Upload" %>
<% end %>
<% if @album.tag %>
<%= button_to "Search Instagram for Images with Tag: #{@album.tag}", search_path(@album.tag), method: :get %>
<% end %>
</fieldset>
</div>
## Instruction:
Fix fieldset going off of page - due to instagram button
## Code After:
<div class="new-image">
<%= cloudinary_js_config %>
<%= form_for :image, id: "upload-form", url: album_images_path, html: { multipart: true } do |f| %>
<fieldset>
<legend>Plop an Image</legend>
<%= render "shared/errors" %>
<%= f.file_field(:image_url) %>
<label for="direct_url">Image Url</label>
<input type="text" name="image[direct_url]" id="direct_url">
<%= f.label :caption %>
<%= f.text_field :caption %>
<%= f.submit "Upload" %>
<% end %>
<% if @album.tag %>
<%= button_to "Search Instagram for Images with Your Tag", search_path(@album.tag), method: :get %>
<% end %>
</fieldset>
</div>
| <div class="new-image">
<%= cloudinary_js_config %>
<%= form_for :image, id: "upload-form", url: album_images_path, html: { multipart: true } do |f| %>
<fieldset>
<legend>Plop an Image</legend>
<%= render "shared/errors" %>
<%= f.file_field(:image_url) %>
<label for="direct_url">Image Url</label>
<input type="text" name="image[direct_url]" id="direct_url">
<%= f.label :caption %>
<%= f.text_field :caption %>
<%= f.submit "Upload" %>
<% end %>
<% if @album.tag %>
- <%= button_to "Search Instagram for Images with Tag: #{@album.tag}", search_path(@album.tag), method: :get %>
? ---------------
+ <%= button_to "Search Instagram for Images with Your Tag", search_path(@album.tag), method: :get %>
? +++++
<% end %>
</fieldset>
</div> | 2 | 0.083333 | 1 | 1 |
852f067c7aab6bdcaabf2550fc5a0995a7e9b0ae | maediprojects/__init__.py | maediprojects/__init__.py | from flask import Flask
from flask.ext.sqlalchemy import SQLAlchemy
from flask.ext.babel import Babel
import os
app = Flask(__name__.split('.')[0])
app.config.from_pyfile(os.path.join('..', 'config.py'))
db = SQLAlchemy(app)
babel = Babel(app)
import routes
@babel.localeselector
def get_locale():
return app.config["BABEL_DEFAULT_LOCALE"]
| from flask import Flask
from flask.ext.sqlalchemy import SQLAlchemy
from flask.ext.babel import Babel
from flask.ext.mail import Mail
import os
app = Flask(__name__.split('.')[0])
app.config.from_pyfile(os.path.join('..', 'config.py'))
db = SQLAlchemy(app)
babel = Babel(app)
mail = Mail(app)
import routes
@babel.localeselector
def get_locale():
return app.config["BABEL_DEFAULT_LOCALE"]
| Add flask-mail to use for emailing updates | Add flask-mail to use for emailing updates
| Python | agpl-3.0 | markbrough/maedi-projects,markbrough/maedi-projects,markbrough/maedi-projects | python | ## Code Before:
from flask import Flask
from flask.ext.sqlalchemy import SQLAlchemy
from flask.ext.babel import Babel
import os
app = Flask(__name__.split('.')[0])
app.config.from_pyfile(os.path.join('..', 'config.py'))
db = SQLAlchemy(app)
babel = Babel(app)
import routes
@babel.localeselector
def get_locale():
return app.config["BABEL_DEFAULT_LOCALE"]
## Instruction:
Add flask-mail to use for emailing updates
## Code After:
from flask import Flask
from flask.ext.sqlalchemy import SQLAlchemy
from flask.ext.babel import Babel
from flask.ext.mail import Mail
import os
app = Flask(__name__.split('.')[0])
app.config.from_pyfile(os.path.join('..', 'config.py'))
db = SQLAlchemy(app)
babel = Babel(app)
mail = Mail(app)
import routes
@babel.localeselector
def get_locale():
return app.config["BABEL_DEFAULT_LOCALE"]
| from flask import Flask
from flask.ext.sqlalchemy import SQLAlchemy
from flask.ext.babel import Babel
+ from flask.ext.mail import Mail
import os
app = Flask(__name__.split('.')[0])
app.config.from_pyfile(os.path.join('..', 'config.py'))
db = SQLAlchemy(app)
babel = Babel(app)
+ mail = Mail(app)
import routes
@babel.localeselector
def get_locale():
return app.config["BABEL_DEFAULT_LOCALE"] | 2 | 0.133333 | 2 | 0 |
84b48b9be466ac72bddf5ee6288ff48be26eed62 | tests/classifier/RandomForestClassifier/RandomForestClassifierPHPTest.py | tests/classifier/RandomForestClassifier/RandomForestClassifierPHPTest.py |
import unittest
from unittest import TestCase
from sklearn.ensemble import RandomForestClassifier
from ..Classifier import Classifier
from ...language.PHP import PHP
class RandomForestClassifierPHPTest(PHP, Classifier, TestCase):
def setUp(self):
super(RandomForestClassifierPHPTest, self).setUp()
self.mdl = RandomForestClassifier(n_estimators=100, random_state=0)
def tearDown(self):
super(RandomForestClassifierPHPTest, self).tearDown()
@unittest.skip('The generated code would be too large.')
def test_existing_features_w_digits_data(self):
pass
@unittest.skip('The generated code would be too large.')
def test_random_features_w_digits_data(self):
pass
|
import unittest
from unittest import TestCase
from sklearn.ensemble import RandomForestClassifier
from ..Classifier import Classifier
from ...language.PHP import PHP
class RandomForestClassifierPHPTest(PHP, Classifier, TestCase):
def setUp(self):
super(RandomForestClassifierPHPTest, self).setUp()
self.mdl = RandomForestClassifier(n_estimators=20, random_state=0)
def tearDown(self):
super(RandomForestClassifierPHPTest, self).tearDown()
| Reduce the number of trees | Reduce the number of trees
| Python | bsd-3-clause | nok/sklearn-porter | python | ## Code Before:
import unittest
from unittest import TestCase
from sklearn.ensemble import RandomForestClassifier
from ..Classifier import Classifier
from ...language.PHP import PHP
class RandomForestClassifierPHPTest(PHP, Classifier, TestCase):
def setUp(self):
super(RandomForestClassifierPHPTest, self).setUp()
self.mdl = RandomForestClassifier(n_estimators=100, random_state=0)
def tearDown(self):
super(RandomForestClassifierPHPTest, self).tearDown()
@unittest.skip('The generated code would be too large.')
def test_existing_features_w_digits_data(self):
pass
@unittest.skip('The generated code would be too large.')
def test_random_features_w_digits_data(self):
pass
## Instruction:
Reduce the number of trees
## Code After:
import unittest
from unittest import TestCase
from sklearn.ensemble import RandomForestClassifier
from ..Classifier import Classifier
from ...language.PHP import PHP
class RandomForestClassifierPHPTest(PHP, Classifier, TestCase):
def setUp(self):
super(RandomForestClassifierPHPTest, self).setUp()
self.mdl = RandomForestClassifier(n_estimators=20, random_state=0)
def tearDown(self):
super(RandomForestClassifierPHPTest, self).tearDown()
|
import unittest
from unittest import TestCase
from sklearn.ensemble import RandomForestClassifier
from ..Classifier import Classifier
from ...language.PHP import PHP
class RandomForestClassifierPHPTest(PHP, Classifier, TestCase):
def setUp(self):
super(RandomForestClassifierPHPTest, self).setUp()
- self.mdl = RandomForestClassifier(n_estimators=100, random_state=0)
? ^^
+ self.mdl = RandomForestClassifier(n_estimators=20, random_state=0)
? ^
def tearDown(self):
super(RandomForestClassifierPHPTest, self).tearDown()
-
- @unittest.skip('The generated code would be too large.')
- def test_existing_features_w_digits_data(self):
- pass
-
- @unittest.skip('The generated code would be too large.')
- def test_random_features_w_digits_data(self):
- pass | 10 | 0.384615 | 1 | 9 |
8f13251b0a65ca0b943255f7375514e24b378b1f | chef/cookbooks/pivotal_ci/metadata.rb | chef/cookbooks/pivotal_ci/metadata.rb | name "pivotal_ci"
maintainer "Pivotal Labs"
maintainer_email "commoncode+lobot@pivotallabs.com"
license "MIT"
description "Sets up Lobot"
version "0.1.0"
recipe "pivotal_ci::default", "Install Lobot"
supports "ubuntu", "12.04"
| name "pivotal_ci"
maintainer "Pivotal Labs"
maintainer_email "commoncode+lobot@pivotallabs.com"
license "MIT"
description "Sets up Lobot"
version "0.1.0"
recipe "pivotal_ci::default", "Install Lobot"
supports "ubuntu", "12.04"
depends "build-essential"
depends "chromium"
depends "cmake"
depends "firefox"
depends "imagemagick"
depends "libffi"
depends "libgdbm"
depends "libncurses"
depends "libossp-uuid"
depends "libqt4"
depends "mysql"
depends "networking_basic"
depends "nodejs"
depends "openssl"
depends "phantomjs"
depends "postgresql"
depends "ragel"
depends "ramfs"
depends "sysctl"
depends "unarchivers"
depends "xserver"
| Declare cookbook dependencies for Chef 11 | Declare cookbook dependencies for Chef 11
This appears to be necessary for attributes, libraries etc. to be loaded
in the correct order in Chef 11.
See
http://docs.opscode.com/breaking_changes_chef_11.html#non-recipe-file-evaluation-includes-dependencies
| Ruby | mit | pivotal/lobot,pivotal/ciborg,pivotal/ciborg,pivotal/ciborg,pivotal/lobot | ruby | ## Code Before:
name "pivotal_ci"
maintainer "Pivotal Labs"
maintainer_email "commoncode+lobot@pivotallabs.com"
license "MIT"
description "Sets up Lobot"
version "0.1.0"
recipe "pivotal_ci::default", "Install Lobot"
supports "ubuntu", "12.04"
## Instruction:
Declare cookbook dependencies for Chef 11
This appears to be necessary for attributes, libraries etc. to be loaded
in the correct order in Chef 11.
See
http://docs.opscode.com/breaking_changes_chef_11.html#non-recipe-file-evaluation-includes-dependencies
## Code After:
name "pivotal_ci"
maintainer "Pivotal Labs"
maintainer_email "commoncode+lobot@pivotallabs.com"
license "MIT"
description "Sets up Lobot"
version "0.1.0"
recipe "pivotal_ci::default", "Install Lobot"
supports "ubuntu", "12.04"
depends "build-essential"
depends "chromium"
depends "cmake"
depends "firefox"
depends "imagemagick"
depends "libffi"
depends "libgdbm"
depends "libncurses"
depends "libossp-uuid"
depends "libqt4"
depends "mysql"
depends "networking_basic"
depends "nodejs"
depends "openssl"
depends "phantomjs"
depends "postgresql"
depends "ragel"
depends "ramfs"
depends "sysctl"
depends "unarchivers"
depends "xserver"
| name "pivotal_ci"
maintainer "Pivotal Labs"
maintainer_email "commoncode+lobot@pivotallabs.com"
license "MIT"
description "Sets up Lobot"
version "0.1.0"
recipe "pivotal_ci::default", "Install Lobot"
supports "ubuntu", "12.04"
+
+ depends "build-essential"
+ depends "chromium"
+ depends "cmake"
+ depends "firefox"
+ depends "imagemagick"
+ depends "libffi"
+ depends "libgdbm"
+ depends "libncurses"
+ depends "libossp-uuid"
+ depends "libqt4"
+ depends "mysql"
+ depends "networking_basic"
+ depends "nodejs"
+ depends "openssl"
+ depends "phantomjs"
+ depends "postgresql"
+ depends "ragel"
+ depends "ramfs"
+ depends "sysctl"
+ depends "unarchivers"
+ depends "xserver" | 22 | 2.2 | 22 | 0 |
ccdf7a7c68492c624fbb4de6378230d9a5ff59ed | proto/eventmon/messages.go | proto/eventmon/messages.go | package eventmon
const (
ConnectString = "200 Connected to keymaster eventmon service"
HttpPath = "/eventmon/v0"
EventTypeSSHCert = "SSHCert"
EventTypeX509Cert = "X509Cert"
)
// Client sends no data. Server sends a sequence of events.
type EventV0 struct {
Type string
CertData []byte `json:",omitempty"`
}
| package eventmon
const (
ConnectString = "200 Connected to keymaster eventmon service"
HttpPath = "/eventmon/v0"
AuthTypePassword = "Password"
AuthTypeSymantecVIP = "SymantecVIP"
AuthTypeU2F = "U2F"
EventTypeSSHCert = "SSHCert"
EventTypeWebLogin = "WebLogin"
EventTypeX509Cert = "X509Cert"
)
// Client sends no data. Server sends a sequence of events.
type EventV0 struct {
Type string
// Present for SSH and X509 certificate events.
CertData []byte `json:",omitempty"`
// Present for Web login events.
AuthType string `json:",omitempty"`
Username string `json:",omitempty"`
}
| Add Web login events to eventmon protocol. | Add Web login events to eventmon protocol.
| Go | apache-2.0 | rgooch/keymaster,Symantec/keymaster,rgooch/keymaster,Symantec/keymaster,Symantec/keymaster,rgooch/keymaster | go | ## Code Before:
package eventmon
const (
ConnectString = "200 Connected to keymaster eventmon service"
HttpPath = "/eventmon/v0"
EventTypeSSHCert = "SSHCert"
EventTypeX509Cert = "X509Cert"
)
// Client sends no data. Server sends a sequence of events.
type EventV0 struct {
Type string
CertData []byte `json:",omitempty"`
}
## Instruction:
Add Web login events to eventmon protocol.
## Code After:
package eventmon
const (
ConnectString = "200 Connected to keymaster eventmon service"
HttpPath = "/eventmon/v0"
AuthTypePassword = "Password"
AuthTypeSymantecVIP = "SymantecVIP"
AuthTypeU2F = "U2F"
EventTypeSSHCert = "SSHCert"
EventTypeWebLogin = "WebLogin"
EventTypeX509Cert = "X509Cert"
)
// Client sends no data. Server sends a sequence of events.
type EventV0 struct {
Type string
// Present for SSH and X509 certificate events.
CertData []byte `json:",omitempty"`
// Present for Web login events.
AuthType string `json:",omitempty"`
Username string `json:",omitempty"`
}
| package eventmon
const (
ConnectString = "200 Connected to keymaster eventmon service"
HttpPath = "/eventmon/v0"
+ AuthTypePassword = "Password"
+ AuthTypeSymantecVIP = "SymantecVIP"
+ AuthTypeU2F = "U2F"
+
EventTypeSSHCert = "SSHCert"
+ EventTypeWebLogin = "WebLogin"
EventTypeX509Cert = "X509Cert"
)
// Client sends no data. Server sends a sequence of events.
type EventV0 struct {
- Type string
? ----
+ Type string
+
+ // Present for SSH and X509 certificate events.
CertData []byte `json:",omitempty"`
+
+ // Present for Web login events.
+ AuthType string `json:",omitempty"`
+ Username string `json:",omitempty"`
} | 13 | 0.8125 | 12 | 1 |
79ac2da646ed6285fc64d6c9e7ba13f1f237be0d | test/CodeGen/pr9614.c | test/CodeGen/pr9614.c | // RUN: %clang_cc1 -emit-llvm %s -o - | FileCheck %s
extern void foo_alias (void) __asm ("foo");
inline void foo (void) {
return foo_alias ();
}
extern void bar_alias (void) __asm ("bar");
inline __attribute__ ((__always_inline__)) void bar (void) {
return bar_alias ();
}
extern char *strrchr_foo (const char *__s, int __c) __asm ("strrchr");
extern inline __attribute__ ((__always_inline__)) __attribute__ ((__gnu_inline__)) char * strrchr_foo (const char *__s, int __c) {
return __builtin_strrchr (__s, __c);
}
void f(void) {
foo();
bar();
strrchr_foo("", '.');
}
// CHECK: define void @f()
// CHECK: call void @foo()
// CHECK-NEXT: call void @bar()
// CHECK-NEXT: call i8* @strrchr(
// CHECK-NEXT: ret void
// CHECK: declare void @foo()
// CHECK: declare void @bar()
// CHECK: declare i8* @strrchr(i8*, i32)
| // RUN: %clang_cc1 -triple x86_64-pc-linux -emit-llvm %s -o - | FileCheck %s
extern void foo_alias (void) __asm ("foo");
inline void foo (void) {
return foo_alias ();
}
extern void bar_alias (void) __asm ("bar");
inline __attribute__ ((__always_inline__)) void bar (void) {
return bar_alias ();
}
extern char *strrchr_foo (const char *__s, int __c) __asm ("strrchr");
extern inline __attribute__ ((__always_inline__)) __attribute__ ((__gnu_inline__)) char * strrchr_foo (const char *__s, int __c) {
return __builtin_strrchr (__s, __c);
}
void f(void) {
foo();
bar();
strrchr_foo("", '.');
}
// CHECK: define void @f()
// CHECK: call void @foo()
// CHECK-NEXT: call void @bar()
// CHECK-NEXT: call i8* @strrchr(
// CHECK-NEXT: ret void
// CHECK: declare void @foo()
// CHECK: declare void @bar()
// CHECK: declare i8* @strrchr(i8*, i32)
| Add a triple to the test. | Add a triple to the test.
git-svn-id: ffe668792ed300d6c2daa1f6eba2e0aa28d7ec6c@146871 91177308-0d34-0410-b5e6-96231b3b80d8
| C | apache-2.0 | apple/swift-clang,apple/swift-clang,llvm-mirror/clang,apple/swift-clang,llvm-mirror/clang,apple/swift-clang,apple/swift-clang,apple/swift-clang,apple/swift-clang,llvm-mirror/clang,llvm-mirror/clang,llvm-mirror/clang,llvm-mirror/clang,apple/swift-clang,apple/swift-clang,llvm-mirror/clang,llvm-mirror/clang,llvm-mirror/clang,apple/swift-clang,llvm-mirror/clang | c | ## Code Before:
// RUN: %clang_cc1 -emit-llvm %s -o - | FileCheck %s
extern void foo_alias (void) __asm ("foo");
inline void foo (void) {
return foo_alias ();
}
extern void bar_alias (void) __asm ("bar");
inline __attribute__ ((__always_inline__)) void bar (void) {
return bar_alias ();
}
extern char *strrchr_foo (const char *__s, int __c) __asm ("strrchr");
extern inline __attribute__ ((__always_inline__)) __attribute__ ((__gnu_inline__)) char * strrchr_foo (const char *__s, int __c) {
return __builtin_strrchr (__s, __c);
}
void f(void) {
foo();
bar();
strrchr_foo("", '.');
}
// CHECK: define void @f()
// CHECK: call void @foo()
// CHECK-NEXT: call void @bar()
// CHECK-NEXT: call i8* @strrchr(
// CHECK-NEXT: ret void
// CHECK: declare void @foo()
// CHECK: declare void @bar()
// CHECK: declare i8* @strrchr(i8*, i32)
## Instruction:
Add a triple to the test.
git-svn-id: ffe668792ed300d6c2daa1f6eba2e0aa28d7ec6c@146871 91177308-0d34-0410-b5e6-96231b3b80d8
## Code After:
// RUN: %clang_cc1 -triple x86_64-pc-linux -emit-llvm %s -o - | FileCheck %s
extern void foo_alias (void) __asm ("foo");
inline void foo (void) {
return foo_alias ();
}
extern void bar_alias (void) __asm ("bar");
inline __attribute__ ((__always_inline__)) void bar (void) {
return bar_alias ();
}
extern char *strrchr_foo (const char *__s, int __c) __asm ("strrchr");
extern inline __attribute__ ((__always_inline__)) __attribute__ ((__gnu_inline__)) char * strrchr_foo (const char *__s, int __c) {
return __builtin_strrchr (__s, __c);
}
void f(void) {
foo();
bar();
strrchr_foo("", '.');
}
// CHECK: define void @f()
// CHECK: call void @foo()
// CHECK-NEXT: call void @bar()
// CHECK-NEXT: call i8* @strrchr(
// CHECK-NEXT: ret void
// CHECK: declare void @foo()
// CHECK: declare void @bar()
// CHECK: declare i8* @strrchr(i8*, i32)
| - // RUN: %clang_cc1 -emit-llvm %s -o - | FileCheck %s
+ // RUN: %clang_cc1 -triple x86_64-pc-linux -emit-llvm %s -o - | FileCheck %s
? ++++++++++++++++++++++++
extern void foo_alias (void) __asm ("foo");
inline void foo (void) {
return foo_alias ();
}
extern void bar_alias (void) __asm ("bar");
inline __attribute__ ((__always_inline__)) void bar (void) {
return bar_alias ();
}
extern char *strrchr_foo (const char *__s, int __c) __asm ("strrchr");
extern inline __attribute__ ((__always_inline__)) __attribute__ ((__gnu_inline__)) char * strrchr_foo (const char *__s, int __c) {
return __builtin_strrchr (__s, __c);
}
void f(void) {
foo();
bar();
strrchr_foo("", '.');
}
// CHECK: define void @f()
// CHECK: call void @foo()
// CHECK-NEXT: call void @bar()
// CHECK-NEXT: call i8* @strrchr(
// CHECK-NEXT: ret void
// CHECK: declare void @foo()
// CHECK: declare void @bar()
// CHECK: declare i8* @strrchr(i8*, i32) | 2 | 0.068966 | 1 | 1 |
285f31bee2f3fc7fda87ec3c89a2111bd2bce3d4 | package.json | package.json | {
"name": "EspruinoHub",
"version": "0.0.0",
"description": "Linux based hub software for Puck.js / Espruino",
"main": "index.js",
"scripts": {
},
"author": "Gordon Williams (gw@pur3.co.uk)",
"license": "MPLv2",
"dependencies": {
"bleno": "^0.4.1",
"mqtt": "^2.0.1",
"noble": "^1.7.0"
}
}
| {
"name": "EspruinoHub",
"version": "0.0.0",
"description": "Linux based hub software for Puck.js / Espruino",
"main": "index.js",
"scripts": {
},
"author": "Gordon Williams (gw@pur3.co.uk)",
"license": "MPLv2",
"dependencies": {
"bleno": "^0.4.1",
"mqtt": "^2.0.1",
"noble": "^1.7.0",
"bluetooth-hci-socket": "^0.5"
}
}
| Add missing dependency to bluetooth-hci-socket | Add missing dependency to bluetooth-hci-socket | JSON | mpl-2.0 | tomgidden/EspruinoHub,tomgidden/EspruinoHub,tomgidden/EspruinoHub | json | ## Code Before:
{
"name": "EspruinoHub",
"version": "0.0.0",
"description": "Linux based hub software for Puck.js / Espruino",
"main": "index.js",
"scripts": {
},
"author": "Gordon Williams (gw@pur3.co.uk)",
"license": "MPLv2",
"dependencies": {
"bleno": "^0.4.1",
"mqtt": "^2.0.1",
"noble": "^1.7.0"
}
}
## Instruction:
Add missing dependency to bluetooth-hci-socket
## Code After:
{
"name": "EspruinoHub",
"version": "0.0.0",
"description": "Linux based hub software for Puck.js / Espruino",
"main": "index.js",
"scripts": {
},
"author": "Gordon Williams (gw@pur3.co.uk)",
"license": "MPLv2",
"dependencies": {
"bleno": "^0.4.1",
"mqtt": "^2.0.1",
"noble": "^1.7.0",
"bluetooth-hci-socket": "^0.5"
}
}
| {
"name": "EspruinoHub",
"version": "0.0.0",
"description": "Linux based hub software for Puck.js / Espruino",
"main": "index.js",
"scripts": {
},
"author": "Gordon Williams (gw@pur3.co.uk)",
"license": "MPLv2",
"dependencies": {
"bleno": "^0.4.1",
"mqtt": "^2.0.1",
- "noble": "^1.7.0"
+ "noble": "^1.7.0",
? +
+ "bluetooth-hci-socket": "^0.5"
}
} | 3 | 0.2 | 2 | 1 |
239fc296a53bafce25ce0050dac4055fa27d3e73 | features/apimgt/org.wso2.carbon.apimgt.store.feature/src/main/resources/devportal/source/src/app/components/Apis/Details/ApiConsole/SwaggerUI.jsx | features/apimgt/org.wso2.carbon.apimgt.store.feature/src/main/resources/devportal/source/src/app/components/Apis/Details/ApiConsole/SwaggerUI.jsx | import React from 'react';
import PropTypes from 'prop-types';
import 'swagger-ui/dist/swagger-ui.css';
import SwaggerUILib from './PatchedSwaggerUIReact';
const disableAuthorizeAndInfoPlugin = function () {
return {
wrapComponents: {
info: () => () => null,
},
};
};
/**
*
* @class SwaggerUI
* @extends {Component}
*/
const SwaggerUI = (props) => {
const { spec, accessTokenProvider, authorizationHeader, api } = props;
const componentProps = {
spec,
validatorUrl: null,
docExpansion: 'list',
defaultModelsExpandDepth: 0,
requestInterceptor: (req) => {
const { url } = req;
const patternToCheck = api.context + '/*';
req.headers[authorizationHeader] = 'Bearer ' + accessTokenProvider();
if (url.endsWith(patternToCheck)) {
req.url = url.substring(0, url.length - 2);
} else if (url.includes('/*?')) {
const splitTokens = url.split('/*?');
req.url = splitTokens.length > 1 ? splitTokens[0] + '?' + splitTokens[1] : splitTokens[0];
}
return req;
},
presets: [disableAuthorizeAndInfoPlugin],
plugins: null,
};
return <SwaggerUILib {...componentProps} />;
};
SwaggerUI.propTypes = {
spec: PropTypes.shape({}).isRequired,
};
export default SwaggerUI;
| import React from 'react';
import PropTypes from 'prop-types';
import 'swagger-ui/dist/swagger-ui.css';
import SwaggerUILib from './PatchedSwaggerUIReact';
const disableAuthorizeAndInfoPlugin = function () {
return {
wrapComponents: {
info: () => () => null,
},
};
};
/**
*
* @class SwaggerUI
* @extends {Component}
*/
const SwaggerUI = (props) => {
const { spec, accessTokenProvider, authorizationHeader, api } = props;
const componentProps = {
spec,
validatorUrl: null,
docExpansion: 'list',
defaultModelsExpandDepth: 0,
requestInterceptor: (req) => {
const { url } = req;
const patternToCheck = api.context + '/*';
req.headers[authorizationHeader] = 'Bearer ' + accessTokenProvider();
if (url.endsWith(patternToCheck)) {
req.url = url.substring(0, url.length - 2);
} else if (url.includes(patternToCheck + '?')) { // Check for query parameters.
const splitTokens = url.split('/*?');
req.url = splitTokens.length > 1 ? splitTokens[0] + '?' + splitTokens[1] : splitTokens[0];
}
return req;
},
presets: [disableAuthorizeAndInfoPlugin],
plugins: null,
};
return <SwaggerUILib {...componentProps} />;
};
SwaggerUI.propTypes = {
spec: PropTypes.shape({}).isRequired,
};
export default SwaggerUI;
| Change query params in swagger console | Change query params in swagger console
| JSX | apache-2.0 | uvindra/carbon-apimgt,bhathiya/carbon-apimgt,harsha89/carbon-apimgt,chamindias/carbon-apimgt,praminda/carbon-apimgt,nuwand/carbon-apimgt,fazlan-nazeem/carbon-apimgt,nuwand/carbon-apimgt,praminda/carbon-apimgt,harsha89/carbon-apimgt,chamilaadhi/carbon-apimgt,isharac/carbon-apimgt,jaadds/carbon-apimgt,praminda/carbon-apimgt,tharikaGitHub/carbon-apimgt,nuwand/carbon-apimgt,jaadds/carbon-apimgt,harsha89/carbon-apimgt,malinthaprasan/carbon-apimgt,sanjeewa-malalgoda/carbon-apimgt,Rajith90/carbon-apimgt,tharindu1st/carbon-apimgt,Rajith90/carbon-apimgt,tharindu1st/carbon-apimgt,bhathiya/carbon-apimgt,prasa7/carbon-apimgt,bhathiya/carbon-apimgt,bhathiya/carbon-apimgt,prasa7/carbon-apimgt,ruks/carbon-apimgt,jaadds/carbon-apimgt,harsha89/carbon-apimgt,prasa7/carbon-apimgt,tharindu1st/carbon-apimgt,chamilaadhi/carbon-apimgt,chamindias/carbon-apimgt,sanjeewa-malalgoda/carbon-apimgt,ruks/carbon-apimgt,tharikaGitHub/carbon-apimgt,wso2/carbon-apimgt,fazlan-nazeem/carbon-apimgt,uvindra/carbon-apimgt,wso2/carbon-apimgt,uvindra/carbon-apimgt,isharac/carbon-apimgt,prasa7/carbon-apimgt,isharac/carbon-apimgt,chamilaadhi/carbon-apimgt,sanjeewa-malalgoda/carbon-apimgt,fazlan-nazeem/carbon-apimgt,jaadds/carbon-apimgt,sanjeewa-malalgoda/carbon-apimgt,nuwand/carbon-apimgt,malinthaprasan/carbon-apimgt,wso2/carbon-apimgt,malinthaprasan/carbon-apimgt,chamindias/carbon-apimgt,uvindra/carbon-apimgt,ruks/carbon-apimgt,malinthaprasan/carbon-apimgt,Rajith90/carbon-apimgt,fazlan-nazeem/carbon-apimgt,tharikaGitHub/carbon-apimgt,wso2/carbon-apimgt,chamilaadhi/carbon-apimgt,tharindu1st/carbon-apimgt,ruks/carbon-apimgt,tharikaGitHub/carbon-apimgt,chamindias/carbon-apimgt,Rajith90/carbon-apimgt,isharac/carbon-apimgt | jsx | ## Code Before:
import React from 'react';
import PropTypes from 'prop-types';
import 'swagger-ui/dist/swagger-ui.css';
import SwaggerUILib from './PatchedSwaggerUIReact';
const disableAuthorizeAndInfoPlugin = function () {
return {
wrapComponents: {
info: () => () => null,
},
};
};
/**
*
* @class SwaggerUI
* @extends {Component}
*/
const SwaggerUI = (props) => {
const { spec, accessTokenProvider, authorizationHeader, api } = props;
const componentProps = {
spec,
validatorUrl: null,
docExpansion: 'list',
defaultModelsExpandDepth: 0,
requestInterceptor: (req) => {
const { url } = req;
const patternToCheck = api.context + '/*';
req.headers[authorizationHeader] = 'Bearer ' + accessTokenProvider();
if (url.endsWith(patternToCheck)) {
req.url = url.substring(0, url.length - 2);
} else if (url.includes('/*?')) {
const splitTokens = url.split('/*?');
req.url = splitTokens.length > 1 ? splitTokens[0] + '?' + splitTokens[1] : splitTokens[0];
}
return req;
},
presets: [disableAuthorizeAndInfoPlugin],
plugins: null,
};
return <SwaggerUILib {...componentProps} />;
};
SwaggerUI.propTypes = {
spec: PropTypes.shape({}).isRequired,
};
export default SwaggerUI;
## Instruction:
Change query params in swagger console
## Code After:
import React from 'react';
import PropTypes from 'prop-types';
import 'swagger-ui/dist/swagger-ui.css';
import SwaggerUILib from './PatchedSwaggerUIReact';
const disableAuthorizeAndInfoPlugin = function () {
return {
wrapComponents: {
info: () => () => null,
},
};
};
/**
*
* @class SwaggerUI
* @extends {Component}
*/
const SwaggerUI = (props) => {
const { spec, accessTokenProvider, authorizationHeader, api } = props;
const componentProps = {
spec,
validatorUrl: null,
docExpansion: 'list',
defaultModelsExpandDepth: 0,
requestInterceptor: (req) => {
const { url } = req;
const patternToCheck = api.context + '/*';
req.headers[authorizationHeader] = 'Bearer ' + accessTokenProvider();
if (url.endsWith(patternToCheck)) {
req.url = url.substring(0, url.length - 2);
} else if (url.includes(patternToCheck + '?')) { // Check for query parameters.
const splitTokens = url.split('/*?');
req.url = splitTokens.length > 1 ? splitTokens[0] + '?' + splitTokens[1] : splitTokens[0];
}
return req;
},
presets: [disableAuthorizeAndInfoPlugin],
plugins: null,
};
return <SwaggerUILib {...componentProps} />;
};
SwaggerUI.propTypes = {
spec: PropTypes.shape({}).isRequired,
};
export default SwaggerUI;
| import React from 'react';
import PropTypes from 'prop-types';
import 'swagger-ui/dist/swagger-ui.css';
import SwaggerUILib from './PatchedSwaggerUIReact';
const disableAuthorizeAndInfoPlugin = function () {
return {
wrapComponents: {
info: () => () => null,
},
};
};
/**
*
* @class SwaggerUI
* @extends {Component}
*/
const SwaggerUI = (props) => {
const { spec, accessTokenProvider, authorizationHeader, api } = props;
const componentProps = {
spec,
validatorUrl: null,
docExpansion: 'list',
defaultModelsExpandDepth: 0,
requestInterceptor: (req) => {
const { url } = req;
const patternToCheck = api.context + '/*';
req.headers[authorizationHeader] = 'Bearer ' + accessTokenProvider();
if (url.endsWith(patternToCheck)) {
req.url = url.substring(0, url.length - 2);
- } else if (url.includes('/*?')) {
+ } else if (url.includes(patternToCheck + '?')) { // Check for query parameters.
const splitTokens = url.split('/*?');
req.url = splitTokens.length > 1 ? splitTokens[0] + '?' + splitTokens[1] : splitTokens[0];
}
return req;
},
presets: [disableAuthorizeAndInfoPlugin],
plugins: null,
};
return <SwaggerUILib {...componentProps} />;
};
SwaggerUI.propTypes = {
spec: PropTypes.shape({}).isRequired,
};
export default SwaggerUI; | 2 | 0.040816 | 1 | 1 |
fce75fa9526b327a8281aaad0388b582ea1917be | ext/conf/tinylog-findbugs.xml | ext/conf/tinylog-findbugs.xml | <!-- Exclude list of warnings -->
<FindBugsFilter>
<Match>
<Or>
<Class name="org.pmw.tinylog.Logger" />
<Class name="org.apache.log4j.TinylogBride" />
</Or>
<!-- Catching of exceptions is required for the case that the fast but unofficial ways to get the stack trace will fail -->
<Bug pattern="REC_CATCH_EXCEPTION" />
</Match>
<Match>
<Class name="org.apache.log4j.Priority" />
<!-- Known as bad but source code is from Log4j -->
<Bug pattern="HE_EQUALS_USE_HASHCODE" />
</Match>
<Match>
<!-- Default encoding is wanted, even if the encoding vary between platforms -->
<Bug pattern="DM_DEFAULT_ENCODING" />
</Match>
</FindBugsFilter> | <!-- Exclude list of warnings -->
<FindBugsFilter>
<Match>
<Or>
<Class name="org.pmw.tinylog.Logger" />
<Class name="org.apache.log4j.TinylogBride" />
</Or>
<!-- Catching of exceptions is required for the case that the fast but unofficial ways to get the stack trace will fail -->
<Bug pattern="REC_CATCH_EXCEPTION" />
</Match>
<Match>
<Class name="org.apache.log4j.Priority" />
<!-- Known as bad but source code is from Log4j -->
<Bug pattern="HE_EQUALS_USE_HASHCODE" />
</Match>
<Match>
<!-- Default encoding is wanted, even if the encoding will be vary between platforms -->
<Bug pattern="DM_DEFAULT_ENCODING" />
</Match>
<Match>
<Class name="org.pmw.tinylog.writers.LogEntry" />
<!-- Problem with "java.util.Date" is known but currently there is no better alternative in the Java library -->
<Bug pattern="EI_EXPOSE_REP, EI_EXPOSE_REP2" />
</Match>
<Match>
<Class name="org.pmw.tinylog.writers.SharedFileWriter" />
<!-- "file.delete()" can be successful (=> create a new log file) as well as it can fail (=> join existing log file) -->
<Bug pattern="RV_RETURN_VALUE_IGNORED_BAD_PRACTICE" />
</Match>
</FindBugsFilter> | Update exclude list for findbugs | Update exclude list for findbugs | XML | apache-2.0 | yarish/tinylog,yarish/tinylog,robymus/tinylog,robymus/tinylog | xml | ## Code Before:
<!-- Exclude list of warnings -->
<FindBugsFilter>
<Match>
<Or>
<Class name="org.pmw.tinylog.Logger" />
<Class name="org.apache.log4j.TinylogBride" />
</Or>
<!-- Catching of exceptions is required for the case that the fast but unofficial ways to get the stack trace will fail -->
<Bug pattern="REC_CATCH_EXCEPTION" />
</Match>
<Match>
<Class name="org.apache.log4j.Priority" />
<!-- Known as bad but source code is from Log4j -->
<Bug pattern="HE_EQUALS_USE_HASHCODE" />
</Match>
<Match>
<!-- Default encoding is wanted, even if the encoding vary between platforms -->
<Bug pattern="DM_DEFAULT_ENCODING" />
</Match>
</FindBugsFilter>
## Instruction:
Update exclude list for findbugs
## Code After:
<!-- Exclude list of warnings -->
<FindBugsFilter>
<Match>
<Or>
<Class name="org.pmw.tinylog.Logger" />
<Class name="org.apache.log4j.TinylogBride" />
</Or>
<!-- Catching of exceptions is required for the case that the fast but unofficial ways to get the stack trace will fail -->
<Bug pattern="REC_CATCH_EXCEPTION" />
</Match>
<Match>
<Class name="org.apache.log4j.Priority" />
<!-- Known as bad but source code is from Log4j -->
<Bug pattern="HE_EQUALS_USE_HASHCODE" />
</Match>
<Match>
<!-- Default encoding is wanted, even if the encoding will be vary between platforms -->
<Bug pattern="DM_DEFAULT_ENCODING" />
</Match>
<Match>
<Class name="org.pmw.tinylog.writers.LogEntry" />
<!-- Problem with "java.util.Date" is known but currently there is no better alternative in the Java library -->
<Bug pattern="EI_EXPOSE_REP, EI_EXPOSE_REP2" />
</Match>
<Match>
<Class name="org.pmw.tinylog.writers.SharedFileWriter" />
<!-- "file.delete()" can be successful (=> create a new log file) as well as it can fail (=> join existing log file) -->
<Bug pattern="RV_RETURN_VALUE_IGNORED_BAD_PRACTICE" />
</Match>
</FindBugsFilter> | <!-- Exclude list of warnings -->
<FindBugsFilter>
<Match>
<Or>
<Class name="org.pmw.tinylog.Logger" />
<Class name="org.apache.log4j.TinylogBride" />
</Or>
<!-- Catching of exceptions is required for the case that the fast but unofficial ways to get the stack trace will fail -->
<Bug pattern="REC_CATCH_EXCEPTION" />
</Match>
<Match>
<Class name="org.apache.log4j.Priority" />
<!-- Known as bad but source code is from Log4j -->
<Bug pattern="HE_EQUALS_USE_HASHCODE" />
</Match>
<Match>
- <!-- Default encoding is wanted, even if the encoding vary between platforms -->
+ <!-- Default encoding is wanted, even if the encoding will be vary between platforms -->
? ++++++++
<Bug pattern="DM_DEFAULT_ENCODING" />
</Match>
+ <Match>
+ <Class name="org.pmw.tinylog.writers.LogEntry" />
+ <!-- Problem with "java.util.Date" is known but currently there is no better alternative in the Java library -->
+ <Bug pattern="EI_EXPOSE_REP, EI_EXPOSE_REP2" />
+ </Match>
+ <Match>
+ <Class name="org.pmw.tinylog.writers.SharedFileWriter" />
+ <!-- "file.delete()" can be successful (=> create a new log file) as well as it can fail (=> join existing log file) -->
+ <Bug pattern="RV_RETURN_VALUE_IGNORED_BAD_PRACTICE" />
+ </Match>
</FindBugsFilter> | 12 | 0.6 | 11 | 1 |
d2843d1e41abb2190aad51125db8aebd875e8125 | src/main/resources/application.yml | src/main/resources/application.yml | endpoints:
hypermedia:
enabled: true
info:
app:
encoding: @project.build.sourceEncoding@
java:
source: @maven.compiler.source@
target: @maven.compiler.target@
project:
name: '@project.name@'
groupId: @project.groupId@
artifactId: @project.artifactId@
version: @project.version@
management:
port: 9898
security:
enabled: true
security:
basic:
enabled: false
headers:
# disables pragma no-cache header
cache: false
server:
contextPath: /iiif
server-header: "IIIF Server Demo v@project.version@"
spring:
profiles:
active: PROD
thymeleaf:
cache: true
---
spring:
profiles: local
thymeleaf:
cache: false
management:
security:
enabled: false | endpoints:
hypermedia:
enabled: true
info:
app:
encoding: @project.build.sourceEncoding@
java:
source: @maven.compiler.source@
target: @maven.compiler.target@
project:
name: '@project.name@'
groupId: @project.groupId@
artifactId: @project.artifactId@
version: @project.version@
management:
port: 9898
security:
enabled: true
security:
basic:
enabled: false
headers:
# disables pragma no-cache header
cache: false
server:
contextPath: /iiif
server-header: "IIIF Server Demo v@project.version@"
spring:
profiles:
active: PROD
thymeleaf:
cache: true
---
spring:
profiles: local
thymeleaf:
cache: false
management:
security:
enabled: false
---
spring:
profiles: PROD
security:
user:
name: admin
password: secret | Add credentials for the actuator sites | Add credentials for the actuator sites
| YAML | mit | dbmdz/iiif-server-demo,dbmdz/iiif-server-demo | yaml | ## Code Before:
endpoints:
hypermedia:
enabled: true
info:
app:
encoding: @project.build.sourceEncoding@
java:
source: @maven.compiler.source@
target: @maven.compiler.target@
project:
name: '@project.name@'
groupId: @project.groupId@
artifactId: @project.artifactId@
version: @project.version@
management:
port: 9898
security:
enabled: true
security:
basic:
enabled: false
headers:
# disables pragma no-cache header
cache: false
server:
contextPath: /iiif
server-header: "IIIF Server Demo v@project.version@"
spring:
profiles:
active: PROD
thymeleaf:
cache: true
---
spring:
profiles: local
thymeleaf:
cache: false
management:
security:
enabled: false
## Instruction:
Add credentials for the actuator sites
## Code After:
endpoints:
hypermedia:
enabled: true
info:
app:
encoding: @project.build.sourceEncoding@
java:
source: @maven.compiler.source@
target: @maven.compiler.target@
project:
name: '@project.name@'
groupId: @project.groupId@
artifactId: @project.artifactId@
version: @project.version@
management:
port: 9898
security:
enabled: true
security:
basic:
enabled: false
headers:
# disables pragma no-cache header
cache: false
server:
contextPath: /iiif
server-header: "IIIF Server Demo v@project.version@"
spring:
profiles:
active: PROD
thymeleaf:
cache: true
---
spring:
profiles: local
thymeleaf:
cache: false
management:
security:
enabled: false
---
spring:
profiles: PROD
security:
user:
name: admin
password: secret | endpoints:
hypermedia:
enabled: true
info:
app:
encoding: @project.build.sourceEncoding@
java:
source: @maven.compiler.source@
target: @maven.compiler.target@
project:
name: '@project.name@'
groupId: @project.groupId@
artifactId: @project.artifactId@
version: @project.version@
management:
port: 9898
security:
enabled: true
security:
basic:
enabled: false
headers:
# disables pragma no-cache header
cache: false
server:
contextPath: /iiif
server-header: "IIIF Server Demo v@project.version@"
spring:
profiles:
active: PROD
thymeleaf:
cache: true
---
spring:
profiles: local
thymeleaf:
cache: false
management:
security:
enabled: false
+
+ ---
+
+ spring:
+ profiles: PROD
+
+ security:
+ user:
+ name: admin
+ password: secret | 10 | 0.208333 | 10 | 0 |
db399a1fe74789f80ec57f309119a655bce660f3 | build.sh | build.sh |
python3.6 -m venv thinglang-env
source thinglang-env/bin/activate
pip install -r requirements.txt
mkdir /app-src/build
cd /app-src/build
cmake ..
make
/app-src/build/thinglang --build-only
cd /app-src
pytest tests/integration |
python3.6 -m venv thinglang-env
source thinglang-env/bin/activate
pip install -r requirements.txt
mkdir /app-src/build
cd /app-src/build
cmake ..
make
/app-src/build/thinglang --build-only
export PATH=$PATH:/app-src/build/
cd /app-src
pytest tests/integration | Add thinglang executable to path | Add thinglang executable to path
| Shell | mit | ytanay/thinglang,ytanay/thinglang,ytanay/thinglang,ytanay/thinglang | shell | ## Code Before:
python3.6 -m venv thinglang-env
source thinglang-env/bin/activate
pip install -r requirements.txt
mkdir /app-src/build
cd /app-src/build
cmake ..
make
/app-src/build/thinglang --build-only
cd /app-src
pytest tests/integration
## Instruction:
Add thinglang executable to path
## Code After:
python3.6 -m venv thinglang-env
source thinglang-env/bin/activate
pip install -r requirements.txt
mkdir /app-src/build
cd /app-src/build
cmake ..
make
/app-src/build/thinglang --build-only
export PATH=$PATH:/app-src/build/
cd /app-src
pytest tests/integration |
python3.6 -m venv thinglang-env
source thinglang-env/bin/activate
pip install -r requirements.txt
mkdir /app-src/build
cd /app-src/build
cmake ..
make
/app-src/build/thinglang --build-only
+ export PATH=$PATH:/app-src/build/
+
cd /app-src
pytest tests/integration | 2 | 0.153846 | 2 | 0 |
44d4f29cbe74fd76a9883601591aef45883621ab | README.md | README.md | Expose
======
Creates a flat list of album suitable to be imported with iTunes, from
the documents stared in Picasa.
### Prerequisites:
* It is recomanded backup of your photos in oreder to avoid any data loss
* This software requires [Java runtime 1.7](http://java.com/en/download)
### Command line:
`java -jar expose.jar <source> <target>`
* source: directory of the Picasa document collection
* target: directory where stared document are copied. This directory
must be initially empty. The target must be in the same
filesystem as the source and the file system must support hardlinks.
**Warning:** The target is purged for the documents that are not seend stared in
the source. It means the files and directory in the target can be lost.
### Details:
First the program scans recursively the directory hierarchy, looking for
Picasa '.ini' files. Then the files are parsed in order to extract the list
of stared documents. Now a directory is created for every album that hold
at least one stared file. The album is populated with links to the original
documents. Finally the files & directories of the target that no longer match
a stared document of the input are removed.
| Expose
======
Creates a flat list of album suitable to be imported with iTunes, from
the documents stared in Picasa.
### Prerequisites:
* It is recomanded backup of your photos in oreder to avoid any data loss
* This software requires [Java runtime 1.7](http://java.com/en/download)
### Command line:
`java -jar expose.jar <source> <target>`
* source: directory of the Picasa document collection
* target: directory where stared document are copied. This directory
must be initially empty. The target must be in the same
filesystem as the source and the file system must support hardlinks.
**Warning:** The target is purged for the documents that are not seend stared in
the source. It means the files and directory in the target can be lost.
### Details:
First the program scans recursively the directory hierarchy, looking for
Picasa '.ini' files. Then the files are parsed in order to extract the list
of stared documents. Now a directory is created for every album that hold
at least one stared file. The album is populated with links to the original
documents. Finally the files & directories of the target that no longer match
a stared document of the input are removed.
### Screenshot:

| Add a screenshot to readme | Add a screenshot to readme | Markdown | mit | piwicode/expose | markdown | ## Code Before:
Expose
======
Creates a flat list of album suitable to be imported with iTunes, from
the documents stared in Picasa.
### Prerequisites:
* It is recomanded backup of your photos in oreder to avoid any data loss
* This software requires [Java runtime 1.7](http://java.com/en/download)
### Command line:
`java -jar expose.jar <source> <target>`
* source: directory of the Picasa document collection
* target: directory where stared document are copied. This directory
must be initially empty. The target must be in the same
filesystem as the source and the file system must support hardlinks.
**Warning:** The target is purged for the documents that are not seend stared in
the source. It means the files and directory in the target can be lost.
### Details:
First the program scans recursively the directory hierarchy, looking for
Picasa '.ini' files. Then the files are parsed in order to extract the list
of stared documents. Now a directory is created for every album that hold
at least one stared file. The album is populated with links to the original
documents. Finally the files & directories of the target that no longer match
a stared document of the input are removed.
## Instruction:
Add a screenshot to readme
## Code After:
Expose
======
Creates a flat list of album suitable to be imported with iTunes, from
the documents stared in Picasa.
### Prerequisites:
* It is recomanded backup of your photos in oreder to avoid any data loss
* This software requires [Java runtime 1.7](http://java.com/en/download)
### Command line:
`java -jar expose.jar <source> <target>`
* source: directory of the Picasa document collection
* target: directory where stared document are copied. This directory
must be initially empty. The target must be in the same
filesystem as the source and the file system must support hardlinks.
**Warning:** The target is purged for the documents that are not seend stared in
the source. It means the files and directory in the target can be lost.
### Details:
First the program scans recursively the directory hierarchy, looking for
Picasa '.ini' files. Then the files are parsed in order to extract the list
of stared documents. Now a directory is created for every album that hold
at least one stared file. The album is populated with links to the original
documents. Finally the files & directories of the target that no longer match
a stared document of the input are removed.
### Screenshot:

| Expose
======
Creates a flat list of album suitable to be imported with iTunes, from
the documents stared in Picasa.
### Prerequisites:
+
* It is recomanded backup of your photos in oreder to avoid any data loss
* This software requires [Java runtime 1.7](http://java.com/en/download)
+ ### Command line:
- ### Command line:
`java -jar expose.jar <source> <target>`
* source: directory of the Picasa document collection
* target: directory where stared document are copied. This directory
must be initially empty. The target must be in the same
filesystem as the source and the file system must support hardlinks.
**Warning:** The target is purged for the documents that are not seend stared in
the source. It means the files and directory in the target can be lost.
### Details:
+
First the program scans recursively the directory hierarchy, looking for
Picasa '.ini' files. Then the files are parsed in order to extract the list
of stared documents. Now a directory is created for every album that hold
at least one stared file. The album is populated with links to the original
documents. Finally the files & directories of the target that no longer match
a stared document of the input are removed.
+ ### Screenshot:
+
+  | 7 | 0.233333 | 6 | 1 |
330114d17efbaa3acf11c6bec6a114c14d660bb6 | irrverbs/config.go | irrverbs/config.go | package main
import "code.google.com/p/gcfg"
// Config is config
type Config struct {
Telegram struct {
Token string
Username string
}
}
func getConfig() (Config, error) {
var cfg Config
err := gcfg.ReadFileInto(&cfg, "./config.cfg")
return cfg, err
}
| package main
import "gopkg.in/gcfg.v1"
// Config is config
type Config struct {
Telegram struct {
Token string
Username string
}
}
func getConfig() (Config, error) {
var cfg Config
err := gcfg.ReadFileInto(&cfg, "/tmp/config.cfg")
return cfg, err
}
| Change imprort name of gcfg to make project compile | Change imprort name of gcfg to make project compile
| Go | mit | sanchopanca/irrverbs | go | ## Code Before:
package main
import "code.google.com/p/gcfg"
// Config is config
type Config struct {
Telegram struct {
Token string
Username string
}
}
func getConfig() (Config, error) {
var cfg Config
err := gcfg.ReadFileInto(&cfg, "./config.cfg")
return cfg, err
}
## Instruction:
Change imprort name of gcfg to make project compile
## Code After:
package main
import "gopkg.in/gcfg.v1"
// Config is config
type Config struct {
Telegram struct {
Token string
Username string
}
}
func getConfig() (Config, error) {
var cfg Config
err := gcfg.ReadFileInto(&cfg, "/tmp/config.cfg")
return cfg, err
}
| package main
- import "code.google.com/p/gcfg"
+ import "gopkg.in/gcfg.v1"
// Config is config
type Config struct {
Telegram struct {
Token string
Username string
}
}
func getConfig() (Config, error) {
var cfg Config
- err := gcfg.ReadFileInto(&cfg, "./config.cfg")
? ^
+ err := gcfg.ReadFileInto(&cfg, "/tmp/config.cfg")
? ^^^^
return cfg, err
} | 4 | 0.235294 | 2 | 2 |
3e159dede448e7e73e7f06f7ff467dacc9307bcc | em-twitter.gemspec | em-twitter.gemspec | lib = File.expand_path('../lib', __FILE__)
$LOAD_PATH.unshift(lib) unless $LOAD_PATH.include?(lib)
require 'em-twitter/version'
Gem::Specification.new do |spec|
spec.name = 'em-twitter'
spec.version = EventMachine::Twitter::VERSION
spec.homepage = 'https://github.com/spagalloco/em-twitter'
spec.licenses = ['MIT']
spec.authors = ["Steve Agalloco"]
spec.email = ['steve.agalloco@gmail.com']
spec.description = %q{Twitter Streaming API client for EventMachine}
spec.summary = spec.description
spec.add_dependency 'eventmachine', '~> 1.0'
spec.add_dependency 'http_parser.rb', '~> 0.5'
spec.add_dependency 'simple_oauth', '~> 0.2'
spec.add_development_dependency 'bundler', '~> 1.0'
spec.files = %w(.yardopts CONTRIBUTING.md LICENSE.md README.md Rakefile em-twitter.gemspec)
spec.files += Dir.glob("lib/**/*.rb")
spec.files += Dir.glob("spec/**/*")
spec.require_paths = ['lib']
end
| lib = File.expand_path('../lib', __FILE__)
$LOAD_PATH.unshift(lib) unless $LOAD_PATH.include?(lib)
require 'em-twitter/version'
Gem::Specification.new do |spec|
spec.name = 'em-twitter'
spec.version = EventMachine::Twitter::VERSION
spec.homepage = 'https://github.com/spagalloco/em-twitter'
spec.licenses = ['MIT']
spec.authors = ["Steve Agalloco"]
spec.email = ['steve.agalloco@gmail.com']
spec.description = %q{Twitter Streaming API client for EventMachine}
spec.summary = spec.description
spec.add_dependency 'eventmachine', '~> 1.0'
spec.add_dependency 'http_parser.rb', ['>= 0.6.0.beta.2', '< 0.7']
spec.add_dependency 'simple_oauth', '~> 0.2'
spec.add_development_dependency 'bundler', '~> 1.0'
spec.files = %w(.yardopts CONTRIBUTING.md LICENSE.md README.md Rakefile em-twitter.gemspec)
spec.files += Dir.glob("lib/**/*.rb")
spec.files += Dir.glob("spec/**/*")
spec.require_paths = ['lib']
end
| Update http_parser.rb dependency to allow latest beta | Update http_parser.rb dependency to allow latest beta
| Ruby | mit | chgu82837/em-twitter,tweetstream/em-twitter | ruby | ## Code Before:
lib = File.expand_path('../lib', __FILE__)
$LOAD_PATH.unshift(lib) unless $LOAD_PATH.include?(lib)
require 'em-twitter/version'
Gem::Specification.new do |spec|
spec.name = 'em-twitter'
spec.version = EventMachine::Twitter::VERSION
spec.homepage = 'https://github.com/spagalloco/em-twitter'
spec.licenses = ['MIT']
spec.authors = ["Steve Agalloco"]
spec.email = ['steve.agalloco@gmail.com']
spec.description = %q{Twitter Streaming API client for EventMachine}
spec.summary = spec.description
spec.add_dependency 'eventmachine', '~> 1.0'
spec.add_dependency 'http_parser.rb', '~> 0.5'
spec.add_dependency 'simple_oauth', '~> 0.2'
spec.add_development_dependency 'bundler', '~> 1.0'
spec.files = %w(.yardopts CONTRIBUTING.md LICENSE.md README.md Rakefile em-twitter.gemspec)
spec.files += Dir.glob("lib/**/*.rb")
spec.files += Dir.glob("spec/**/*")
spec.require_paths = ['lib']
end
## Instruction:
Update http_parser.rb dependency to allow latest beta
## Code After:
lib = File.expand_path('../lib', __FILE__)
$LOAD_PATH.unshift(lib) unless $LOAD_PATH.include?(lib)
require 'em-twitter/version'
Gem::Specification.new do |spec|
spec.name = 'em-twitter'
spec.version = EventMachine::Twitter::VERSION
spec.homepage = 'https://github.com/spagalloco/em-twitter'
spec.licenses = ['MIT']
spec.authors = ["Steve Agalloco"]
spec.email = ['steve.agalloco@gmail.com']
spec.description = %q{Twitter Streaming API client for EventMachine}
spec.summary = spec.description
spec.add_dependency 'eventmachine', '~> 1.0'
spec.add_dependency 'http_parser.rb', ['>= 0.6.0.beta.2', '< 0.7']
spec.add_dependency 'simple_oauth', '~> 0.2'
spec.add_development_dependency 'bundler', '~> 1.0'
spec.files = %w(.yardopts CONTRIBUTING.md LICENSE.md README.md Rakefile em-twitter.gemspec)
spec.files += Dir.glob("lib/**/*.rb")
spec.files += Dir.glob("spec/**/*")
spec.require_paths = ['lib']
end
| lib = File.expand_path('../lib', __FILE__)
$LOAD_PATH.unshift(lib) unless $LOAD_PATH.include?(lib)
require 'em-twitter/version'
Gem::Specification.new do |spec|
spec.name = 'em-twitter'
spec.version = EventMachine::Twitter::VERSION
spec.homepage = 'https://github.com/spagalloco/em-twitter'
spec.licenses = ['MIT']
spec.authors = ["Steve Agalloco"]
spec.email = ['steve.agalloco@gmail.com']
spec.description = %q{Twitter Streaming API client for EventMachine}
spec.summary = spec.description
spec.add_dependency 'eventmachine', '~> 1.0'
- spec.add_dependency 'http_parser.rb', '~> 0.5'
? - ^
+ spec.add_dependency 'http_parser.rb', ['>= 0.6.0.beta.2', '< 0.7']
? + + ^^^^^^^^^^ ++++++++++
spec.add_dependency 'simple_oauth', '~> 0.2'
spec.add_development_dependency 'bundler', '~> 1.0'
spec.files = %w(.yardopts CONTRIBUTING.md LICENSE.md README.md Rakefile em-twitter.gemspec)
spec.files += Dir.glob("lib/**/*.rb")
spec.files += Dir.glob("spec/**/*")
spec.require_paths = ['lib']
end | 2 | 0.074074 | 1 | 1 |
c4d8e08a68a938d4832c29edbc282abb8bd060b0 | locales/zh-TW/notes.properties | locales/zh-TW/notes.properties | welcomeTitle2=哈囉!
welcomeText2=歡迎使用這套 Firefox 內建的一頁式記事本。上網時快速記下筆記,就是這麼簡單。
emptyPlaceHolder=寫點筆記…
giveFeedback=點擊此處,提供意見回饋給我們
openingLogin=正在開啟登入畫面…
forgetEmail=忘記此地址
syncNotReady2=很抱歉,筆記同步功能還沒開發完成。我們將把您的點擊次數記錄下來,當作投票來加速我們的開發!
syncNotes=同步您的筆記
syncProgress=正在同步變更…
# LOCALIZATION NOTE (disableSync): Sync is intended as a generic
# synchronization, not Firefox Sync.
disableSync=停用同步
# LOCALIZATION NOTE (syncComplete): {date} is the date of last sync. If this
# structure doesn't work for your locale, you can translate this as "Last sync:
# {date}".
syncComplete=同步於 {date}
# Tooltips for toolbar buttons
fontSizeTitle=字型大小
boldTitle=粗體
italicTitle=斜體
strikethroughTitle=刪除線
numberedListTitle=編號清單
bulletedListTitle=項目清單
textDirectionTitle=文字方向
# Settings page labels
themeLegend=佈景主題
defaultThemeTitle=預設
darkThemeTitle=暗色
| welcomeTitle2=哈囉!
welcomeText2=歡迎使用這套 Firefox 內建的一頁式記事本。上網時快速記下筆記,就是這麼簡單。
emptyPlaceHolder=寫點筆記…
giveFeedback=點擊此處,提供意見回饋給我們
feedback=意見回饋
openingLogin=正在開啟登入畫面…
forgetEmail=忘記此地址
syncNotReady2=很抱歉,筆記同步功能還沒開發完成。我們將把您的點擊次數記錄下來,當作投票來加速我們的開發!
syncNotes=同步您的筆記
syncProgress=正在同步變更…
# LOCALIZATION NOTE (disableSync): Sync is intended as a generic
# synchronization, not Firefox Sync.
disableSync=停用同步
# LOCALIZATION NOTE (syncComplete): {date} is the date of last sync. If this
# structure doesn't work for your locale, you can translate this as "Last sync:
# {date}".
syncComplete=同步於 {date}
# Tooltips for toolbar buttons
fontSizeTitle=字型大小
boldTitle=粗體
italicTitle=斜體
strikethroughTitle=刪除線
numberedListTitle=編號清單
bulletedListTitle=項目清單
textDirectionTitle=文字方向
# Settings page labels
themeLegend=佈景主題
defaultThemeTitle=預設
darkThemeTitle=暗色
| Update Chinese (Taiwan) (zh-TW) localization of Test Pilot: Notes | Pontoon: Update Chinese (Taiwan) (zh-TW) localization of Test Pilot: Notes
Localization authors:
- Pin-guang Chen <petercpg@mail.moztw.org>
| INI | mpl-2.0 | cedricium/notes,cedricium/notes,cedricium/notes,cedricium/notes,cedricium/notes | ini | ## Code Before:
welcomeTitle2=哈囉!
welcomeText2=歡迎使用這套 Firefox 內建的一頁式記事本。上網時快速記下筆記,就是這麼簡單。
emptyPlaceHolder=寫點筆記…
giveFeedback=點擊此處,提供意見回饋給我們
openingLogin=正在開啟登入畫面…
forgetEmail=忘記此地址
syncNotReady2=很抱歉,筆記同步功能還沒開發完成。我們將把您的點擊次數記錄下來,當作投票來加速我們的開發!
syncNotes=同步您的筆記
syncProgress=正在同步變更…
# LOCALIZATION NOTE (disableSync): Sync is intended as a generic
# synchronization, not Firefox Sync.
disableSync=停用同步
# LOCALIZATION NOTE (syncComplete): {date} is the date of last sync. If this
# structure doesn't work for your locale, you can translate this as "Last sync:
# {date}".
syncComplete=同步於 {date}
# Tooltips for toolbar buttons
fontSizeTitle=字型大小
boldTitle=粗體
italicTitle=斜體
strikethroughTitle=刪除線
numberedListTitle=編號清單
bulletedListTitle=項目清單
textDirectionTitle=文字方向
# Settings page labels
themeLegend=佈景主題
defaultThemeTitle=預設
darkThemeTitle=暗色
## Instruction:
Pontoon: Update Chinese (Taiwan) (zh-TW) localization of Test Pilot: Notes
Localization authors:
- Pin-guang Chen <petercpg@mail.moztw.org>
## Code After:
welcomeTitle2=哈囉!
welcomeText2=歡迎使用這套 Firefox 內建的一頁式記事本。上網時快速記下筆記,就是這麼簡單。
emptyPlaceHolder=寫點筆記…
giveFeedback=點擊此處,提供意見回饋給我們
feedback=意見回饋
openingLogin=正在開啟登入畫面…
forgetEmail=忘記此地址
syncNotReady2=很抱歉,筆記同步功能還沒開發完成。我們將把您的點擊次數記錄下來,當作投票來加速我們的開發!
syncNotes=同步您的筆記
syncProgress=正在同步變更…
# LOCALIZATION NOTE (disableSync): Sync is intended as a generic
# synchronization, not Firefox Sync.
disableSync=停用同步
# LOCALIZATION NOTE (syncComplete): {date} is the date of last sync. If this
# structure doesn't work for your locale, you can translate this as "Last sync:
# {date}".
syncComplete=同步於 {date}
# Tooltips for toolbar buttons
fontSizeTitle=字型大小
boldTitle=粗體
italicTitle=斜體
strikethroughTitle=刪除線
numberedListTitle=編號清單
bulletedListTitle=項目清單
textDirectionTitle=文字方向
# Settings page labels
themeLegend=佈景主題
defaultThemeTitle=預設
darkThemeTitle=暗色
| welcomeTitle2=哈囉!
welcomeText2=歡迎使用這套 Firefox 內建的一頁式記事本。上網時快速記下筆記,就是這麼簡單。
emptyPlaceHolder=寫點筆記…
giveFeedback=點擊此處,提供意見回饋給我們
+ feedback=意見回饋
openingLogin=正在開啟登入畫面…
forgetEmail=忘記此地址
syncNotReady2=很抱歉,筆記同步功能還沒開發完成。我們將把您的點擊次數記錄下來,當作投票來加速我們的開發!
syncNotes=同步您的筆記
syncProgress=正在同步變更…
# LOCALIZATION NOTE (disableSync): Sync is intended as a generic
# synchronization, not Firefox Sync.
disableSync=停用同步
# LOCALIZATION NOTE (syncComplete): {date} is the date of last sync. If this
# structure doesn't work for your locale, you can translate this as "Last sync:
# {date}".
syncComplete=同步於 {date}
# Tooltips for toolbar buttons
fontSizeTitle=字型大小
boldTitle=粗體
italicTitle=斜體
strikethroughTitle=刪除線
numberedListTitle=編號清單
bulletedListTitle=項目清單
textDirectionTitle=文字方向
# Settings page labels
themeLegend=佈景主題
defaultThemeTitle=預設
darkThemeTitle=暗色 | 1 | 0.029412 | 1 | 0 |
ee4d87538cc36c445a2694eabe7c9d1d52d9e6f2 | app/controllers/meals_controller.rb | app/controllers/meals_controller.rb | class MealsController < MyplaceonlineController
def model
Meal
end
def display_obj(obj)
Myp.display_datetime_short(obj.meal_time, User.current_user)
end
protected
def sorts
["meals.meal_time DESC"]
end
def obj_params
params.require(:meal).permit(
:meal_time,
:notes,
:price,
:calories,
select_or_create_permit(:meal, :location_attributes, LocationsController.param_names)
)
end
end
| class MealsController < MyplaceonlineController
def model
Meal
end
def display_obj(obj)
Myp.display_datetime_short(obj.meal_time, User.current_user)
end
protected
def sorts
["meals.meal_time DESC"]
end
def obj_params
params.require(:meal).permit(
:meal_time,
:notes,
:price,
:calories,
select_or_create_permit(:meal, :location_attributes, LocationsController.param_names)
)
end
def new_obj_initialize
@obj.meal_time = DateTime.now
end
end
| Initialize meal time to current time | Initialize meal time to current time
| Ruby | agpl-3.0 | myplaceonline/myplaceonline_rails,myplaceonline/myplaceonline_rails,myplaceonline/myplaceonline_rails | ruby | ## Code Before:
class MealsController < MyplaceonlineController
def model
Meal
end
def display_obj(obj)
Myp.display_datetime_short(obj.meal_time, User.current_user)
end
protected
def sorts
["meals.meal_time DESC"]
end
def obj_params
params.require(:meal).permit(
:meal_time,
:notes,
:price,
:calories,
select_or_create_permit(:meal, :location_attributes, LocationsController.param_names)
)
end
end
## Instruction:
Initialize meal time to current time
## Code After:
class MealsController < MyplaceonlineController
def model
Meal
end
def display_obj(obj)
Myp.display_datetime_short(obj.meal_time, User.current_user)
end
protected
def sorts
["meals.meal_time DESC"]
end
def obj_params
params.require(:meal).permit(
:meal_time,
:notes,
:price,
:calories,
select_or_create_permit(:meal, :location_attributes, LocationsController.param_names)
)
end
def new_obj_initialize
@obj.meal_time = DateTime.now
end
end
| class MealsController < MyplaceonlineController
def model
Meal
end
def display_obj(obj)
Myp.display_datetime_short(obj.meal_time, User.current_user)
end
protected
def sorts
["meals.meal_time DESC"]
end
def obj_params
params.require(:meal).permit(
:meal_time,
:notes,
:price,
:calories,
select_or_create_permit(:meal, :location_attributes, LocationsController.param_names)
)
end
+
+ def new_obj_initialize
+ @obj.meal_time = DateTime.now
+ end
end | 4 | 0.166667 | 4 | 0 |
ddcb4237332c35f558b91095413c9b59e8f38e9c | style.css | style.css | div.container {
margin-top: 30px;
}
p {
margin-top: 20px;
margin-bottom: 20px;
}
h2 {
padding-top: 7px;
font-size: 14px;
}
table {
padding: 10px;
}
#features td {
padding: 5px;
border: none;
}
pre {
padding: 15px;
}
#logo {
margin: 0px 30px 30px 0px;
}
ul.menu {
margin-left: 0.5em;
padding-left: 0.5em;
}
ul.menu li {
list-style: none;
}
ul.menu li.menu-header {
margin-top: 1em;
margin-bottom: 2em;
}
ul.menu li.menu-header>ul {
margin-top: 1em;
margin-left: 0em;
padding-left: 0em;
}
ul.menu li.release>ul {
margin-left: 1em;
}
ul.menu li.release>ul>li {
display: inline;
}
| div.container {
margin-top: 30px;
}
p {
margin-top: 20px;
margin-bottom: 20px;
}
h2 {
padding-top: 7px;
font-size: 14px;
}
table {
padding: 10px;
}
#features td {
padding: 5px;
border: none;
}
pre {
padding: 15px;
}
#logo {
margin: 0px 30px 30px 0px;
}
ul.menu {
margin-left: 0.5em;
padding-left: 0.5em;
}
ul.menu li {
list-style: none;
margin-bottom: .2em;
}
ul.menu li.menu-header {
margin-top: 1em;
margin-bottom: 2em;
}
ul.menu li.menu-header>ul {
margin-top: 1em;
margin-left: 0em;
padding-left: 0em;
}
ul.menu li.release>ul {
margin-left: 1em;
}
ul.menu li.release>ul>li {
display: inline;
}
| Put a little more spacing between menu items | Put a little more spacing between menu items
| CSS | apache-2.0 | jFransham/rust-www,ituxbag/rust-www,ituxbag/rust-www,jFransham/rust-www,starsep/rust-www,azerupi/rust-www,ituxbag/rust-www,AndrewBrinker/rust-www,starsep/rust-www,jFransham/rust-www,azerupi/rust-www,AndrewBrinker/rust-www,ituxbag/rust-www,AndrewBrinker/rust-www,starsep/rust-www,AndrewBrinker/rust-www,AndrewBrinker/rust-www,jFransham/rust-www,ituxbag/rust-www,starsep/rust-www,azerupi/rust-www,starsep/rust-www,starsep/rust-www,ituxbag/rust-www,AndrewBrinker/rust-www,azerupi/rust-www,azerupi/rust-www | css | ## Code Before:
div.container {
margin-top: 30px;
}
p {
margin-top: 20px;
margin-bottom: 20px;
}
h2 {
padding-top: 7px;
font-size: 14px;
}
table {
padding: 10px;
}
#features td {
padding: 5px;
border: none;
}
pre {
padding: 15px;
}
#logo {
margin: 0px 30px 30px 0px;
}
ul.menu {
margin-left: 0.5em;
padding-left: 0.5em;
}
ul.menu li {
list-style: none;
}
ul.menu li.menu-header {
margin-top: 1em;
margin-bottom: 2em;
}
ul.menu li.menu-header>ul {
margin-top: 1em;
margin-left: 0em;
padding-left: 0em;
}
ul.menu li.release>ul {
margin-left: 1em;
}
ul.menu li.release>ul>li {
display: inline;
}
## Instruction:
Put a little more spacing between menu items
## Code After:
div.container {
margin-top: 30px;
}
p {
margin-top: 20px;
margin-bottom: 20px;
}
h2 {
padding-top: 7px;
font-size: 14px;
}
table {
padding: 10px;
}
#features td {
padding: 5px;
border: none;
}
pre {
padding: 15px;
}
#logo {
margin: 0px 30px 30px 0px;
}
ul.menu {
margin-left: 0.5em;
padding-left: 0.5em;
}
ul.menu li {
list-style: none;
margin-bottom: .2em;
}
ul.menu li.menu-header {
margin-top: 1em;
margin-bottom: 2em;
}
ul.menu li.menu-header>ul {
margin-top: 1em;
margin-left: 0em;
padding-left: 0em;
}
ul.menu li.release>ul {
margin-left: 1em;
}
ul.menu li.release>ul>li {
display: inline;
}
| div.container {
margin-top: 30px;
}
p {
margin-top: 20px;
margin-bottom: 20px;
}
h2 {
padding-top: 7px;
font-size: 14px;
}
table {
padding: 10px;
}
#features td {
padding: 5px;
border: none;
}
pre {
padding: 15px;
}
#logo {
margin: 0px 30px 30px 0px;
}
ul.menu {
margin-left: 0.5em;
padding-left: 0.5em;
}
ul.menu li {
list-style: none;
+ margin-bottom: .2em;
}
ul.menu li.menu-header {
margin-top: 1em;
margin-bottom: 2em;
}
ul.menu li.menu-header>ul {
margin-top: 1em;
margin-left: 0em;
padding-left: 0em;
}
ul.menu li.release>ul {
margin-left: 1em;
}
ul.menu li.release>ul>li {
display: inline;
} | 1 | 0.021277 | 1 | 0 |
540a72dd268b3e3ba3dd90718a56f5abd4154d5c | package.json | package.json | {
"name": "erickmerchant.com-source",
"private": true,
"devDependencies": {
"@erickmerchant/assets": "^1.0.0",
"@erickmerchant/copy-files": "^1.0.0",
"@erickmerchant/html": "^2.0.0",
"@erickmerchant/ift": "^2.1.1",
"@erickmerchant/serve-files": "^1.0.0",
"basscss": "^8.0.0",
"basscss-basic": "^1.0.0",
"chalk": "^1.0.0",
"geomicons-open": "^2.0.0",
"lodash.groupby": "^4.6.0",
"moment-timezone": "^0.5.11",
"slug": "^0.9.1",
"standard": "^10.0.0"
},
"dependencies": {},
"scripts": {
"test": "standard",
"build": "copy-files 'base/*' build && copy-files node_modules/geomicons-open/dist/geomicons.svg build && assets build && html build",
"watch": "copy-files 'base/*' build -w & assets build -w --no-min & html build -w --no-min & serve-files build"
}
}
| {
"name": "erickmerchant.com-source",
"private": true,
"devDependencies": {
"@erickmerchant/assets": "^1.0.0",
"@erickmerchant/copy-files": "^1.0.0",
"@erickmerchant/html": "^2.0.0",
"@erickmerchant/ift": "^2.1.1",
"@erickmerchant/serve-files": "^1.0.0",
"basscss": "^8.0.0",
"basscss-basic": "^1.0.0",
"chalk": "^1.0.0",
"geomicons-open": "^2.0.0",
"lodash.groupby": "^4.6.0",
"moment-timezone": "^0.5.11",
"slug": "^0.9.1",
"standard": "^10.0.0"
},
"dependencies": {},
"scripts": {
"test": "standard",
"build": "copy-files node_modules/geomicons-open/dist/geomicons.svg build && copy-files 'base/*' build && assets build && html build",
"watch": "copy-files node_modules/geomicons-open/dist/geomicons.svg build & copy-files 'base/*' build -w & assets build -w --no-min & html build -w --no-min & serve-files build"
}
}
| Copy icons when watching too | Copy icons when watching too
| JSON | mit | erickmerchant/erickmerchant.com-source,erickmerchant/erickmerchant.com-source | json | ## Code Before:
{
"name": "erickmerchant.com-source",
"private": true,
"devDependencies": {
"@erickmerchant/assets": "^1.0.0",
"@erickmerchant/copy-files": "^1.0.0",
"@erickmerchant/html": "^2.0.0",
"@erickmerchant/ift": "^2.1.1",
"@erickmerchant/serve-files": "^1.0.0",
"basscss": "^8.0.0",
"basscss-basic": "^1.0.0",
"chalk": "^1.0.0",
"geomicons-open": "^2.0.0",
"lodash.groupby": "^4.6.0",
"moment-timezone": "^0.5.11",
"slug": "^0.9.1",
"standard": "^10.0.0"
},
"dependencies": {},
"scripts": {
"test": "standard",
"build": "copy-files 'base/*' build && copy-files node_modules/geomicons-open/dist/geomicons.svg build && assets build && html build",
"watch": "copy-files 'base/*' build -w & assets build -w --no-min & html build -w --no-min & serve-files build"
}
}
## Instruction:
Copy icons when watching too
## Code After:
{
"name": "erickmerchant.com-source",
"private": true,
"devDependencies": {
"@erickmerchant/assets": "^1.0.0",
"@erickmerchant/copy-files": "^1.0.0",
"@erickmerchant/html": "^2.0.0",
"@erickmerchant/ift": "^2.1.1",
"@erickmerchant/serve-files": "^1.0.0",
"basscss": "^8.0.0",
"basscss-basic": "^1.0.0",
"chalk": "^1.0.0",
"geomicons-open": "^2.0.0",
"lodash.groupby": "^4.6.0",
"moment-timezone": "^0.5.11",
"slug": "^0.9.1",
"standard": "^10.0.0"
},
"dependencies": {},
"scripts": {
"test": "standard",
"build": "copy-files node_modules/geomicons-open/dist/geomicons.svg build && copy-files 'base/*' build && assets build && html build",
"watch": "copy-files node_modules/geomicons-open/dist/geomicons.svg build & copy-files 'base/*' build -w & assets build -w --no-min & html build -w --no-min & serve-files build"
}
}
| {
"name": "erickmerchant.com-source",
"private": true,
"devDependencies": {
"@erickmerchant/assets": "^1.0.0",
"@erickmerchant/copy-files": "^1.0.0",
"@erickmerchant/html": "^2.0.0",
"@erickmerchant/ift": "^2.1.1",
"@erickmerchant/serve-files": "^1.0.0",
"basscss": "^8.0.0",
"basscss-basic": "^1.0.0",
"chalk": "^1.0.0",
"geomicons-open": "^2.0.0",
"lodash.groupby": "^4.6.0",
"moment-timezone": "^0.5.11",
"slug": "^0.9.1",
"standard": "^10.0.0"
},
"dependencies": {},
"scripts": {
"test": "standard",
- "build": "copy-files 'base/*' build && copy-files node_modules/geomicons-open/dist/geomicons.svg build && assets build && html build",
? -----------------------------
+ "build": "copy-files node_modules/geomicons-open/dist/geomicons.svg build && copy-files 'base/*' build && assets build && html build",
? ++++++++++++++++++++++++++++++
- "watch": "copy-files 'base/*' build -w & assets build -w --no-min & html build -w --no-min & serve-files build"
+ "watch": "copy-files node_modules/geomicons-open/dist/geomicons.svg build & copy-files 'base/*' build -w & assets build -w --no-min & html build -w --no-min & serve-files build"
? +++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
}
} | 4 | 0.16 | 2 | 2 |
3c22fb745f49a6d9d2189b83a76becc2fbb18c36 | examples/color/app.go | examples/color/app.go | package main
import (
"log"
"time"
"github.com/briandowns/spinner"
)
func main() {
s := spinner.New(spinner.CharSets[0], 100*time.Millisecond)
s.Prefix = "Colors: "
if err := s.Color("yellow"); err != nil {
log.Fatalln(err)
}
s.Start()
time.Sleep(4 * time.Second)
s.Color("red")
s.UpdateCharSet(spinner.CharSets[20])
s.Reverse()
s.Restart()
time.Sleep(4 * time.Second)
s.Color("blue")
s.UpdateCharSet(spinner.CharSets[3])
s.Restart()
time.Sleep(4 * time.Second)
s.Color("cyan")
s.UpdateCharSet(spinner.CharSets[28])
s.Reverse()
s.Restart()
time.Sleep(4 * time.Second)
s.Color("green")
s.UpdateCharSet(spinner.CharSets[25])
s.Restart()
time.Sleep(4 * time.Second)
s.Stop()
}
| package main
import (
"log"
"time"
"github.com/briandowns/spinner"
)
func main() {
s := spinner.New(spinner.CharSets[0], 100*time.Millisecond)
s.Prefix = "Colors: "
if err := s.Color("yellow"); err != nil {
log.Fatalln(err)
}
s.Start()
time.Sleep(4 * time.Second)
if err := s.Color("red"); err != nil {
log.Fatalln(err)
}
s.UpdateCharSet(spinner.CharSets[20])
s.Reverse()
s.Restart()
time.Sleep(4 * time.Second)
if err := s.Color("blue"); err != nil {
log.Fatalln(err)
}
s.UpdateCharSet(spinner.CharSets[3])
s.Restart()
time.Sleep(4 * time.Second)
if err := s.Color("cyan"); err != nil {
log.Fatalln(err)
}
s.UpdateCharSet(spinner.CharSets[28])
s.Reverse()
s.Restart()
time.Sleep(4 * time.Second)
if err := s.Color("green"); err != nil {
log.Fatalln(err)
}
s.UpdateCharSet(spinner.CharSets[25])
s.Restart()
time.Sleep(4 * time.Second)
s.Stop()
}
| Update color example to work the way it should. | Update color example to work the way it should.
| Go | apache-2.0 | diddledan/spinner,briandowns/spinner,mjmac/spinner | go | ## Code Before:
package main
import (
"log"
"time"
"github.com/briandowns/spinner"
)
func main() {
s := spinner.New(spinner.CharSets[0], 100*time.Millisecond)
s.Prefix = "Colors: "
if err := s.Color("yellow"); err != nil {
log.Fatalln(err)
}
s.Start()
time.Sleep(4 * time.Second)
s.Color("red")
s.UpdateCharSet(spinner.CharSets[20])
s.Reverse()
s.Restart()
time.Sleep(4 * time.Second)
s.Color("blue")
s.UpdateCharSet(spinner.CharSets[3])
s.Restart()
time.Sleep(4 * time.Second)
s.Color("cyan")
s.UpdateCharSet(spinner.CharSets[28])
s.Reverse()
s.Restart()
time.Sleep(4 * time.Second)
s.Color("green")
s.UpdateCharSet(spinner.CharSets[25])
s.Restart()
time.Sleep(4 * time.Second)
s.Stop()
}
## Instruction:
Update color example to work the way it should.
## Code After:
package main
import (
"log"
"time"
"github.com/briandowns/spinner"
)
func main() {
s := spinner.New(spinner.CharSets[0], 100*time.Millisecond)
s.Prefix = "Colors: "
if err := s.Color("yellow"); err != nil {
log.Fatalln(err)
}
s.Start()
time.Sleep(4 * time.Second)
if err := s.Color("red"); err != nil {
log.Fatalln(err)
}
s.UpdateCharSet(spinner.CharSets[20])
s.Reverse()
s.Restart()
time.Sleep(4 * time.Second)
if err := s.Color("blue"); err != nil {
log.Fatalln(err)
}
s.UpdateCharSet(spinner.CharSets[3])
s.Restart()
time.Sleep(4 * time.Second)
if err := s.Color("cyan"); err != nil {
log.Fatalln(err)
}
s.UpdateCharSet(spinner.CharSets[28])
s.Reverse()
s.Restart()
time.Sleep(4 * time.Second)
if err := s.Color("green"); err != nil {
log.Fatalln(err)
}
s.UpdateCharSet(spinner.CharSets[25])
s.Restart()
time.Sleep(4 * time.Second)
s.Stop()
}
| package main
import (
"log"
"time"
"github.com/briandowns/spinner"
)
func main() {
s := spinner.New(spinner.CharSets[0], 100*time.Millisecond)
s.Prefix = "Colors: "
if err := s.Color("yellow"); err != nil {
log.Fatalln(err)
}
s.Start()
time.Sleep(4 * time.Second)
-
- s.Color("red")
+ if err := s.Color("red"); err != nil {
+ log.Fatalln(err)
+ }
s.UpdateCharSet(spinner.CharSets[20])
s.Reverse()
s.Restart()
time.Sleep(4 * time.Second)
-
- s.Color("blue")
+ if err := s.Color("blue"); err != nil {
+ log.Fatalln(err)
+ }
s.UpdateCharSet(spinner.CharSets[3])
s.Restart()
time.Sleep(4 * time.Second)
-
- s.Color("cyan")
+ if err := s.Color("cyan"); err != nil {
+ log.Fatalln(err)
+ }
s.UpdateCharSet(spinner.CharSets[28])
s.Reverse()
s.Restart()
time.Sleep(4 * time.Second)
-
- s.Color("green")
+ if err := s.Color("green"); err != nil {
+ log.Fatalln(err)
+ }
s.UpdateCharSet(spinner.CharSets[25])
s.Restart()
time.Sleep(4 * time.Second)
-
s.Stop()
} | 21 | 0.5 | 12 | 9 |
67c1ccee504ae9189bfea7cb5bc9bf8517bf852c | src/make_initial_modules.sh | src/make_initial_modules.sh |
for arch in ptx ptx_dev arm x86; do
C_STUB=architecture.${arch}.stdlib.cpp
LL_STUB=architecture.${arch}.stdlib.ll
RESULT=architecture.${arch}.initmod.c
clang -emit-llvm -S $C_STUB -o - \
| grep -v "^target triple" \
| grep -v "^target datalayout" \
| grep -v "^; ModuleID" \
| cat - $LL_STUB \
| llvm-as - -o - \
| python bitcode2cpp.py ${arch} \
> $RESULT
done |
for arch in ptx ptx_dev arm x86; do
C_STUB=architecture.${arch}.stdlib.cpp
LL_STUB=architecture.${arch}.stdlib.ll
RESULT=architecture.${arch}.initmod.c
if [[ $arch == "ptx" ]]; then
LL_STUB="$LL_STUB architecture.x86.stdlib.ll"
fi
clang -emit-llvm -S $C_STUB -o - \
| grep -v "^target triple" \
| grep -v "^target datalayout" \
| grep -v "^; ModuleID" \
| cat - $LL_STUB \
| llvm-as - -o - \
| python bitcode2cpp.py ${arch} \
> $RESULT
done | Make PTX host initial module include x86 ll | Make PTX host initial module include x86 ll | Shell | mit | rodrigob/Halide,delcypher/Halide,damienfir/Halide,gchauras/Halide,ronen/Halide,jiawen/Halide,ronen/Halide,damienfir/Halide,damienfir/Halide,dan-tull/Halide,fengzhyuan/Halide,kgnk/Halide,lglucin/Halide,dan-tull/Halide,gchauras/Halide,jiawen/Halide,psuriana/Halide,ronen/Halide,dougkwan/Halide,ayanazmat/Halide,damienfir/Halide,ayanazmat/Halide,psuriana/Halide,damienfir/Halide,smxlong/Halide,dougkwan/Halide,fengzhyuan/Halide,ayanazmat/Halide,rodrigob/Halide,psuriana/Halide,aam/Halide,aam/Halide,mcanthony/Halide,lglucin/Halide,ronen/Halide,smxlong/Halide,myrtleTree33/Halide,tdenniston/Halide,tdenniston/Halide,kenkuang1213/Halide,myrtleTree33/Halide,gchauras/Halide,rodrigob/Halide,kgnk/Halide,lglucin/Halide,myrtleTree33/Halide,jiawen/Halide,kgnk/Halide,mikeseven/Halide,mikeseven/Halide,rodrigob/Halide,mikeseven/Halide,mikeseven/Halide,mcanthony/Halide,psuriana/Halide,kenkuang1213/Halide,mcanthony/Halide,gchauras/Halide,dougkwan/Halide,adasworks/Halide,mcanthony/Halide,kenkuang1213/Halide,fengzhyuan/Halide,lglucin/Halide,tdenniston/Halide,delcypher/Halide,adasworks/Halide,ayanazmat/Halide,jiawen/Halide,dan-tull/Halide,fengzhyuan/Halide,tdenniston/Halide,tdenniston/Halide,adasworks/Halide,damienfir/Halide,delcypher/Halide,kgnk/Halide,fengzhyuan/Halide,tdenniston/Halide,jiawen/Halide,adasworks/Halide,ayanazmat/Halide,dougkwan/Halide,ronen/Halide,kenkuang1213/Halide,smxlong/Halide,myrtleTree33/Halide,lglucin/Halide,lglucin/Halide,mcanthony/Halide,smxlong/Halide,mcanthony/Halide,myrtleTree33/Halide,dougkwan/Halide,kgnk/Halide,fengzhyuan/Halide,gchauras/Halide,adasworks/Halide,dougkwan/Halide,ronen/Halide,fengzhyuan/Halide,delcypher/Halide,mikeseven/Halide,dan-tull/Halide,ayanazmat/Halide,fengzhyuan/Halide,myrtleTree33/Halide,rodrigob/Halide,mcanthony/Halide,kenkuang1213/Halide,ayanazmat/Halide,adasworks/Halide,rodrigob/Halide,dan-tull/Halide,myrtleTree33/Halide,smxlong/Halide,jiawen/Halide,aam/Halide,adasworks/Halide,kgnk/Halide,dan-tull/Halide,aam/Halide,tdenniston/Halide,delcypher/Halide,rodrigob/Halide,jiawen/Halide,dougkwan/Halide,adasworks/Halide,lglucin/Halide,kenkuang1213/Halide,smxlong/Halide,kgnk/Halide,rodrigob/Halide,delcypher/Halide,aam/Halide,kenkuang1213/Halide,dan-tull/Halide,aam/Halide,gchauras/Halide,psuriana/Halide,aam/Halide,smxlong/Halide,dan-tull/Halide,ronen/Halide,damienfir/Halide,tdenniston/Halide,kenkuang1213/Halide,dougkwan/Halide,mcanthony/Halide,psuriana/Halide,ayanazmat/Halide,smxlong/Halide,damienfir/Halide,kgnk/Halide,ronen/Halide,delcypher/Halide,delcypher/Halide,psuriana/Halide,myrtleTree33/Halide | shell | ## Code Before:
for arch in ptx ptx_dev arm x86; do
C_STUB=architecture.${arch}.stdlib.cpp
LL_STUB=architecture.${arch}.stdlib.ll
RESULT=architecture.${arch}.initmod.c
clang -emit-llvm -S $C_STUB -o - \
| grep -v "^target triple" \
| grep -v "^target datalayout" \
| grep -v "^; ModuleID" \
| cat - $LL_STUB \
| llvm-as - -o - \
| python bitcode2cpp.py ${arch} \
> $RESULT
done
## Instruction:
Make PTX host initial module include x86 ll
## Code After:
for arch in ptx ptx_dev arm x86; do
C_STUB=architecture.${arch}.stdlib.cpp
LL_STUB=architecture.${arch}.stdlib.ll
RESULT=architecture.${arch}.initmod.c
if [[ $arch == "ptx" ]]; then
LL_STUB="$LL_STUB architecture.x86.stdlib.ll"
fi
clang -emit-llvm -S $C_STUB -o - \
| grep -v "^target triple" \
| grep -v "^target datalayout" \
| grep -v "^; ModuleID" \
| cat - $LL_STUB \
| llvm-as - -o - \
| python bitcode2cpp.py ${arch} \
> $RESULT
done |
for arch in ptx ptx_dev arm x86; do
C_STUB=architecture.${arch}.stdlib.cpp
LL_STUB=architecture.${arch}.stdlib.ll
RESULT=architecture.${arch}.initmod.c
-
+
+ if [[ $arch == "ptx" ]]; then
+ LL_STUB="$LL_STUB architecture.x86.stdlib.ll"
+ fi
+
clang -emit-llvm -S $C_STUB -o - \
| grep -v "^target triple" \
| grep -v "^target datalayout" \
| grep -v "^; ModuleID" \
| cat - $LL_STUB \
| llvm-as - -o - \
| python bitcode2cpp.py ${arch} \
> $RESULT
done | 6 | 0.352941 | 5 | 1 |
2b2ee232bcda04418dd47b64df3e4776ae364a92 | .travis.yml | .travis.yml | language: java
jdk:
- oraclejdk8
install:
- mvn -s .travis.maven.settings.xml -version -B
script:
- mvn -fae -s .travis.maven.settings.xml clean install
notifications:
email: false
env:
global:
- secure: h2UBs+L83eGd8+vTvd4B5sh+ixjp67JB/5Ko/zGAEt6a3hIgrhqyjiXJmaEH3+jzJZKXUcUC6uOuxXHwZ4ODeF+McHViOjqqUGEynJ7AsLXqSr5SSBKZuY/pb69phQnh7Br5w2D+5IFHWEcjrSo/kHDbeZBVAMFRDeGVwpHy2yA=
- secure: BADgz+otMMO+gUlFNduU/mgFm1dOfv2ZWAYD3nO1nEwu7Y1QVUsqimHdlqntUcinILDVeD33oT27kElERvr7KyjfDEPSDAVUg5Y/5soezg/tBu0zi3iodAP2GynOMohS/Rfm4+3ajWuEe/Y2SYjFXgldMR9qjP+/31FsVtW8Op0=
after_success:
- test ${TRAVIS_BRANCH} = master && mvn -s .travis.maven.settings.xml deploy
| language: java
jdk:
- oraclejdk8
install:
- mvn -s .travis.maven.settings.xml -version -B
script:
- mvn -fae -s .travis.maven.settings.xml clean install
notifications:
email: false
env:
global:
- secure: h2UBs+L83eGd8+vTvd4B5sh+ixjp67JB/5Ko/zGAEt6a3hIgrhqyjiXJmaEH3+jzJZKXUcUC6uOuxXHwZ4ODeF+McHViOjqqUGEynJ7AsLXqSr5SSBKZuY/pb69phQnh7Br5w2D+5IFHWEcjrSo/kHDbeZBVAMFRDeGVwpHy2yA=
- secure: BADgz+otMMO+gUlFNduU/mgFm1dOfv2ZWAYD3nO1nEwu7Y1QVUsqimHdlqntUcinILDVeD33oT27kElERvr7KyjfDEPSDAVUg5Y/5soezg/tBu0zi3iodAP2GynOMohS/Rfm4+3ajWuEe/Y2SYjFXgldMR9qjP+/31FsVtW8Op0=
after_success:
- test "${TRAVIS_BRANCH}" = "master" && test "${TRAVIS_PULL_REQUEST}" = "false" && mvn -s .travis.maven.settings.xml deploy
| Add one more test to ensure publishing does not happen for pull requests against the master branch. | Add one more test to ensure publishing does not happen for pull requests against the master branch.
| YAML | apache-2.0 | ammendonca/hawkular,theute/hawkular,panossot/hawkular,mtho11/hawkular,mwringe/hawkular,tsegismont/hawkular,jshaughn/hawkular,vrockai/hawkular,theute/hawkular,panossot/hawkular,mwringe/hawkular,pavolloffay/hawkular,jkandasa/hawkular,ammendonca/hawkular,tsegismont/hawkular,ammendonca/hawkular,panossot/hawkular,metlos/hawkular,ppalaga/hawkular,hawkular/hawkular,jshaughn/hawkular,lucasponce/hawkular,metlos/hawkular,theute/hawkular,panossot/hawkular,Jiri-Kremser/hawkular,ammendonca/hawkular,vrockai/hawkular,pavolloffay/hawkular,lucasponce/hawkular,vrockai/hawkular,mtho11/hawkular,pilhuhn/hawkular,Jiri-Kremser/hawkular,ammendonca/hawkular,metlos/hawkular,jkandasa/hawkular,hawkular/hawkular,jpkrohling/hawkular,jshaughn/hawkular,metlos/hawkular,hawkular/hawkular,Jiri-Kremser/hawkular,jkandasa/hawkular,tsegismont/hawkular,mwringe/hawkular,hawkular/hawkular,hawkular/hawkular-ui,pavolloffay/hawkular,jshaughn/hawkular,jpkrohling/hawkular,hawkular/hawkular,mtho11/hawkular,pilhuhn/hawkular,hawkular/hawkular-ui,metlos/hawkular,ppalaga/hawkular,Jiri-Kremser/hawkular,jpkrohling/hawkular,pavolloffay/hawkular,tsegismont/hawkular,lucasponce/hawkular,jpkrohling/hawkular,hawkular/hawkular-ui,lucasponce/hawkular,vrockai/hawkular,ppalaga/hawkular,lucasponce/hawkular,mtho11/hawkular,jpkrohling/hawkular,jkandasa/hawkular,tsegismont/hawkular,theute/hawkular,ppalaga/hawkular,Jiri-Kremser/hawkular,pilhuhn/hawkular,mwringe/hawkular,pilhuhn/hawkular | yaml | ## Code Before:
language: java
jdk:
- oraclejdk8
install:
- mvn -s .travis.maven.settings.xml -version -B
script:
- mvn -fae -s .travis.maven.settings.xml clean install
notifications:
email: false
env:
global:
- secure: h2UBs+L83eGd8+vTvd4B5sh+ixjp67JB/5Ko/zGAEt6a3hIgrhqyjiXJmaEH3+jzJZKXUcUC6uOuxXHwZ4ODeF+McHViOjqqUGEynJ7AsLXqSr5SSBKZuY/pb69phQnh7Br5w2D+5IFHWEcjrSo/kHDbeZBVAMFRDeGVwpHy2yA=
- secure: BADgz+otMMO+gUlFNduU/mgFm1dOfv2ZWAYD3nO1nEwu7Y1QVUsqimHdlqntUcinILDVeD33oT27kElERvr7KyjfDEPSDAVUg5Y/5soezg/tBu0zi3iodAP2GynOMohS/Rfm4+3ajWuEe/Y2SYjFXgldMR9qjP+/31FsVtW8Op0=
after_success:
- test ${TRAVIS_BRANCH} = master && mvn -s .travis.maven.settings.xml deploy
## Instruction:
Add one more test to ensure publishing does not happen for pull requests against the master branch.
## Code After:
language: java
jdk:
- oraclejdk8
install:
- mvn -s .travis.maven.settings.xml -version -B
script:
- mvn -fae -s .travis.maven.settings.xml clean install
notifications:
email: false
env:
global:
- secure: h2UBs+L83eGd8+vTvd4B5sh+ixjp67JB/5Ko/zGAEt6a3hIgrhqyjiXJmaEH3+jzJZKXUcUC6uOuxXHwZ4ODeF+McHViOjqqUGEynJ7AsLXqSr5SSBKZuY/pb69phQnh7Br5w2D+5IFHWEcjrSo/kHDbeZBVAMFRDeGVwpHy2yA=
- secure: BADgz+otMMO+gUlFNduU/mgFm1dOfv2ZWAYD3nO1nEwu7Y1QVUsqimHdlqntUcinILDVeD33oT27kElERvr7KyjfDEPSDAVUg5Y/5soezg/tBu0zi3iodAP2GynOMohS/Rfm4+3ajWuEe/Y2SYjFXgldMR9qjP+/31FsVtW8Op0=
after_success:
- test "${TRAVIS_BRANCH}" = "master" && test "${TRAVIS_PULL_REQUEST}" = "false" && mvn -s .travis.maven.settings.xml deploy
| language: java
jdk:
- oraclejdk8
install:
- mvn -s .travis.maven.settings.xml -version -B
script:
- mvn -fae -s .travis.maven.settings.xml clean install
notifications:
email: false
env:
global:
- secure: h2UBs+L83eGd8+vTvd4B5sh+ixjp67JB/5Ko/zGAEt6a3hIgrhqyjiXJmaEH3+jzJZKXUcUC6uOuxXHwZ4ODeF+McHViOjqqUGEynJ7AsLXqSr5SSBKZuY/pb69phQnh7Br5w2D+5IFHWEcjrSo/kHDbeZBVAMFRDeGVwpHy2yA=
- secure: BADgz+otMMO+gUlFNduU/mgFm1dOfv2ZWAYD3nO1nEwu7Y1QVUsqimHdlqntUcinILDVeD33oT27kElERvr7KyjfDEPSDAVUg5Y/5soezg/tBu0zi3iodAP2GynOMohS/Rfm4+3ajWuEe/Y2SYjFXgldMR9qjP+/31FsVtW8Op0=
after_success:
- - test ${TRAVIS_BRANCH} = master && mvn -s .travis.maven.settings.xml deploy
+ - test "${TRAVIS_BRANCH}" = "master" && test "${TRAVIS_PULL_REQUEST}" = "false" && mvn -s .travis.maven.settings.xml deploy
? + + + ++++++++++++++++++++++++++++++++++++++++++++
| 2 | 0.125 | 1 | 1 |
11119c98b403117c5621da00ec1f3ae31b105a9f | SQL-Script-2.0.0/script/sspBatchMPU.sql | SQL-Script-2.0.0/script/sspBatchMPU.sql | --drop proc sspBatchMPU -- Monthly Paying User
create proc sspBatchMPU
as
set nocount on
declare @Day30DT datetimeoffset(7)
declare @CurrentDT datetimeoffset(7)
declare @nowdt datetime
declare @MPU bigint
set @nowdt = (select getutcdate())
set @CurrentDT = ((SELECT DATETIMEFROMPARTS (DATEPART(year, @nowdt), DATEPART(month,@nowdt), DATEPART(day, @nowdt), DATEPART(hour, @nowdt), 0, 0, 0 )))
set @Day30DT = (dateadd(day, -30, @CurrentDT))
set @MPU = (select count(*) from MemberItemPurchases where PurchaseDT between @Day30DT and @CurrentDT AND PurchaseCancelYN like 'N')
insert into StatsData(CategoryName, CountNum, Fields, Groups) values('MPU', @MPU, CONVERT(nvarchar(8), GETUTCDATE(), 112), '')
GO | --drop proc sspBatchMPU -- Monthly Paying User
create proc sspBatchMPU
as
set nocount on
declare @Day30DT datetimeoffset(7)
declare @CurrentDT datetimeoffset(7)
declare @nowdt datetime
declare @MPU bigint
set @nowdt = (select getutcdate())
set @CurrentDT = ((SELECT DATETIMEFROMPARTS (DATEPART(year, @nowdt), DATEPART(month,@nowdt), DATEPART(day, @nowdt), DATEPART(hour, @nowdt), 0, 0, 0 )))
set @Day30DT = (dateadd(day, -30, @CurrentDT))
set @MPU = (select count(*) from MemberItemPurchases where PurchaseDT between @Day30DT and @CurrentDT AND PurchaseCancelYN like 'N')
insert into StatsData(CategoryName, CountNum, Fields, Groups) values('MPU', @MPU, CONVERT(nvarchar(8), GETUTCDATE(), 112), '')
GO
------------------------------------------------------------------
-- run test
--exec sspBatchMPU
------------------------------------------------------------------
/*
select * from StatsData order by createdat desc
select * from Members
select * from MemberItemPurchases
select count(*) from MemberItemPurchases where PurchaseDT between '2016-05-15 15:00:03.1749825 +00:00' and sysutcdatetime()
-- test data value
update Members set LastLoginDT = sysutcdatetime() where memberid like 'bbb'
update Members set LastLoginDT = sysutcdatetime() where memberid like 'ccc'
update Members set LastLoginDT = sysutcdatetime() where memberid like 'ddd'
select sysutcdatetime()
select dateadd(day, -30, sysutcdatetime())
select CONVERT(nvarchar(20), getutcdate(), 112)
declare @nowdt datetime
set @nowdt = (select getutcdate())
SELECT DATEPART(year, @nowdt) + '-' + DATEPART(month,@nowdt) + '-' + DATEPART(day, @nowdt);
SELECT convert(datetime, getutcdate(), 121) -- yyyy-mm-dd hh:mm:ss.mmm
*/ | Test the simple MPU query | SQL: Test the simple MPU query
| SQL | mit | CloudBreadProject/CloudBread-DB-Install-Script | sql | ## Code Before:
--drop proc sspBatchMPU -- Monthly Paying User
create proc sspBatchMPU
as
set nocount on
declare @Day30DT datetimeoffset(7)
declare @CurrentDT datetimeoffset(7)
declare @nowdt datetime
declare @MPU bigint
set @nowdt = (select getutcdate())
set @CurrentDT = ((SELECT DATETIMEFROMPARTS (DATEPART(year, @nowdt), DATEPART(month,@nowdt), DATEPART(day, @nowdt), DATEPART(hour, @nowdt), 0, 0, 0 )))
set @Day30DT = (dateadd(day, -30, @CurrentDT))
set @MPU = (select count(*) from MemberItemPurchases where PurchaseDT between @Day30DT and @CurrentDT AND PurchaseCancelYN like 'N')
insert into StatsData(CategoryName, CountNum, Fields, Groups) values('MPU', @MPU, CONVERT(nvarchar(8), GETUTCDATE(), 112), '')
GO
## Instruction:
SQL: Test the simple MPU query
## Code After:
--drop proc sspBatchMPU -- Monthly Paying User
create proc sspBatchMPU
as
set nocount on
declare @Day30DT datetimeoffset(7)
declare @CurrentDT datetimeoffset(7)
declare @nowdt datetime
declare @MPU bigint
set @nowdt = (select getutcdate())
set @CurrentDT = ((SELECT DATETIMEFROMPARTS (DATEPART(year, @nowdt), DATEPART(month,@nowdt), DATEPART(day, @nowdt), DATEPART(hour, @nowdt), 0, 0, 0 )))
set @Day30DT = (dateadd(day, -30, @CurrentDT))
set @MPU = (select count(*) from MemberItemPurchases where PurchaseDT between @Day30DT and @CurrentDT AND PurchaseCancelYN like 'N')
insert into StatsData(CategoryName, CountNum, Fields, Groups) values('MPU', @MPU, CONVERT(nvarchar(8), GETUTCDATE(), 112), '')
GO
------------------------------------------------------------------
-- run test
--exec sspBatchMPU
------------------------------------------------------------------
/*
select * from StatsData order by createdat desc
select * from Members
select * from MemberItemPurchases
select count(*) from MemberItemPurchases where PurchaseDT between '2016-05-15 15:00:03.1749825 +00:00' and sysutcdatetime()
-- test data value
update Members set LastLoginDT = sysutcdatetime() where memberid like 'bbb'
update Members set LastLoginDT = sysutcdatetime() where memberid like 'ccc'
update Members set LastLoginDT = sysutcdatetime() where memberid like 'ddd'
select sysutcdatetime()
select dateadd(day, -30, sysutcdatetime())
select CONVERT(nvarchar(20), getutcdate(), 112)
declare @nowdt datetime
set @nowdt = (select getutcdate())
SELECT DATEPART(year, @nowdt) + '-' + DATEPART(month,@nowdt) + '-' + DATEPART(day, @nowdt);
SELECT convert(datetime, getutcdate(), 121) -- yyyy-mm-dd hh:mm:ss.mmm
*/ | --drop proc sspBatchMPU -- Monthly Paying User
create proc sspBatchMPU
as
set nocount on
declare @Day30DT datetimeoffset(7)
declare @CurrentDT datetimeoffset(7)
declare @nowdt datetime
declare @MPU bigint
set @nowdt = (select getutcdate())
set @CurrentDT = ((SELECT DATETIMEFROMPARTS (DATEPART(year, @nowdt), DATEPART(month,@nowdt), DATEPART(day, @nowdt), DATEPART(hour, @nowdt), 0, 0, 0 )))
set @Day30DT = (dateadd(day, -30, @CurrentDT))
set @MPU = (select count(*) from MemberItemPurchases where PurchaseDT between @Day30DT and @CurrentDT AND PurchaseCancelYN like 'N')
insert into StatsData(CategoryName, CountNum, Fields, Groups) values('MPU', @MPU, CONVERT(nvarchar(8), GETUTCDATE(), 112), '')
GO
+
+ ------------------------------------------------------------------
+ -- run test
+ --exec sspBatchMPU
+ ------------------------------------------------------------------
+
+ /*
+ select * from StatsData order by createdat desc
+ select * from Members
+ select * from MemberItemPurchases
+ select count(*) from MemberItemPurchases where PurchaseDT between '2016-05-15 15:00:03.1749825 +00:00' and sysutcdatetime()
+
+ -- test data value
+ update Members set LastLoginDT = sysutcdatetime() where memberid like 'bbb'
+ update Members set LastLoginDT = sysutcdatetime() where memberid like 'ccc'
+ update Members set LastLoginDT = sysutcdatetime() where memberid like 'ddd'
+
+ select sysutcdatetime()
+ select dateadd(day, -30, sysutcdatetime())
+ select CONVERT(nvarchar(20), getutcdate(), 112)
+
+ declare @nowdt datetime
+ set @nowdt = (select getutcdate())
+ SELECT DATEPART(year, @nowdt) + '-' + DATEPART(month,@nowdt) + '-' + DATEPART(day, @nowdt);
+ SELECT convert(datetime, getutcdate(), 121) -- yyyy-mm-dd hh:mm:ss.mmm
+ */ | 26 | 1.625 | 26 | 0 |
f636420211821faeb3e26a501fbe5a9a7e3eef5e | normal_admin/user_admin.py | normal_admin/user_admin.py | __author__ = 'weijia'
from django.contrib.admin.sites import AdminSite
from django.contrib.admin.forms import AdminAuthenticationForm
from django import forms
from django.utils.translation import ugettext_lazy as _
from django.contrib.auth import authenticate
ERROR_MESSAGE = _("Please enter the correct username and password "
"for a staff account. Note that both fields are case-sensitive.")
class UserAdminAuthenticationForm(AdminAuthenticationForm):
"""
Same as Django's AdminAuthenticationForm but allows to login
any user who is not staff.
"""
def clean(self):
username = self.cleaned_data.get('username')
password = self.cleaned_data.get('password')
message = ERROR_MESSAGE
if username and password:
try:
self.user_cache = authenticate(username=username, password=password)
except:
# The following is for userena as it uses different param
self.user_cache = authenticate(identification=username, password=password)
if self.user_cache is None:
raise forms.ValidationError(message)
elif not self.user_cache.is_active:
raise forms.ValidationError(message)
self.check_for_test_cookie()
return self.cleaned_data
class UserAdmin(AdminSite):
# Anything we wish to add or override
login_form = UserAdminAuthenticationForm
def has_permission(self, request):
return request.user.is_active
| from django.contrib.admin.sites import AdminSite
from django.contrib.admin.forms import AdminAuthenticationForm
from django import forms
from django.utils.translation import ugettext_lazy as _
from django.contrib.auth import authenticate
__author__ = 'weijia'
ERROR_MESSAGE = _("Please enter the correct username and password "
"for a staff account. Note that both fields are case-sensitive.")
class UserAdminAuthenticationForm(AdminAuthenticationForm):
"""
Same as Django's AdminAuthenticationForm but allows to login
any user who is not staff.
"""
def clean(self):
try:
return super(UserAdminAuthenticationForm, self).clean()
except:
username = self.cleaned_data.get('username')
password = self.cleaned_data.get('password')
message = ERROR_MESSAGE
if username and password:
try:
self.user_cache = authenticate(username=username, password=password)
except:
# The following is for userena as it uses different param
self.user_cache = authenticate(identification=username, password=password)
if self.user_cache is None:
raise forms.ValidationError(message)
elif not self.user_cache.is_active:
raise forms.ValidationError(message)
self.check_for_test_cookie()
return self.cleaned_data
# For Django 1.8
def confirm_login_allowed(self, user):
if not user.is_active:
raise forms.ValidationError(
self.error_messages['invalid_login'],
code='invalid_login',
params={'username': self.username_field.verbose_name}
)
class UserAdmin(AdminSite):
# Anything we wish to add or override
login_form = UserAdminAuthenticationForm
def has_permission(self, request):
return request.user.is_active
| Fix login error in Django 1.8. | Fix login error in Django 1.8.
| Python | bsd-3-clause | weijia/normal_admin,weijia/normal_admin | python | ## Code Before:
__author__ = 'weijia'
from django.contrib.admin.sites import AdminSite
from django.contrib.admin.forms import AdminAuthenticationForm
from django import forms
from django.utils.translation import ugettext_lazy as _
from django.contrib.auth import authenticate
ERROR_MESSAGE = _("Please enter the correct username and password "
"for a staff account. Note that both fields are case-sensitive.")
class UserAdminAuthenticationForm(AdminAuthenticationForm):
"""
Same as Django's AdminAuthenticationForm but allows to login
any user who is not staff.
"""
def clean(self):
username = self.cleaned_data.get('username')
password = self.cleaned_data.get('password')
message = ERROR_MESSAGE
if username and password:
try:
self.user_cache = authenticate(username=username, password=password)
except:
# The following is for userena as it uses different param
self.user_cache = authenticate(identification=username, password=password)
if self.user_cache is None:
raise forms.ValidationError(message)
elif not self.user_cache.is_active:
raise forms.ValidationError(message)
self.check_for_test_cookie()
return self.cleaned_data
class UserAdmin(AdminSite):
# Anything we wish to add or override
login_form = UserAdminAuthenticationForm
def has_permission(self, request):
return request.user.is_active
## Instruction:
Fix login error in Django 1.8.
## Code After:
from django.contrib.admin.sites import AdminSite
from django.contrib.admin.forms import AdminAuthenticationForm
from django import forms
from django.utils.translation import ugettext_lazy as _
from django.contrib.auth import authenticate
__author__ = 'weijia'
ERROR_MESSAGE = _("Please enter the correct username and password "
"for a staff account. Note that both fields are case-sensitive.")
class UserAdminAuthenticationForm(AdminAuthenticationForm):
"""
Same as Django's AdminAuthenticationForm but allows to login
any user who is not staff.
"""
def clean(self):
try:
return super(UserAdminAuthenticationForm, self).clean()
except:
username = self.cleaned_data.get('username')
password = self.cleaned_data.get('password')
message = ERROR_MESSAGE
if username and password:
try:
self.user_cache = authenticate(username=username, password=password)
except:
# The following is for userena as it uses different param
self.user_cache = authenticate(identification=username, password=password)
if self.user_cache is None:
raise forms.ValidationError(message)
elif not self.user_cache.is_active:
raise forms.ValidationError(message)
self.check_for_test_cookie()
return self.cleaned_data
# For Django 1.8
def confirm_login_allowed(self, user):
if not user.is_active:
raise forms.ValidationError(
self.error_messages['invalid_login'],
code='invalid_login',
params={'username': self.username_field.verbose_name}
)
class UserAdmin(AdminSite):
# Anything we wish to add or override
login_form = UserAdminAuthenticationForm
def has_permission(self, request):
return request.user.is_active
| - __author__ = 'weijia'
from django.contrib.admin.sites import AdminSite
from django.contrib.admin.forms import AdminAuthenticationForm
from django import forms
from django.utils.translation import ugettext_lazy as _
from django.contrib.auth import authenticate
+
+ __author__ = 'weijia'
ERROR_MESSAGE = _("Please enter the correct username and password "
"for a staff account. Note that both fields are case-sensitive.")
class UserAdminAuthenticationForm(AdminAuthenticationForm):
"""
Same as Django's AdminAuthenticationForm but allows to login
any user who is not staff.
"""
+ def clean(self):
+ try:
+ return super(UserAdminAuthenticationForm, self).clean()
+ except:
+ username = self.cleaned_data.get('username')
+ password = self.cleaned_data.get('password')
+ message = ERROR_MESSAGE
- def clean(self):
- username = self.cleaned_data.get('username')
- password = self.cleaned_data.get('password')
- message = ERROR_MESSAGE
+ if username and password:
+ try:
+ self.user_cache = authenticate(username=username, password=password)
+ except:
+ # The following is for userena as it uses different param
+ self.user_cache = authenticate(identification=username, password=password)
+ if self.user_cache is None:
+ raise forms.ValidationError(message)
+ elif not self.user_cache.is_active:
+ raise forms.ValidationError(message)
+ self.check_for_test_cookie()
+ return self.cleaned_data
+ # For Django 1.8
+ def confirm_login_allowed(self, user):
+ if not user.is_active:
- if username and password:
- try:
- self.user_cache = authenticate(username=username, password=password)
- except:
- # The following is for userena as it uses different param
- self.user_cache = authenticate(identification=username, password=password)
- if self.user_cache is None:
- raise forms.ValidationError(message)
? ---- --------
+ raise forms.ValidationError(
- elif not self.user_cache.is_active:
- raise forms.ValidationError(message)
- self.check_for_test_cookie()
- return self.cleaned_data
+ self.error_messages['invalid_login'],
+ code='invalid_login',
+ params={'username': self.username_field.verbose_name}
+ )
class UserAdmin(AdminSite):
# Anything we wish to add or override
login_form = UserAdminAuthenticationForm
def has_permission(self, request):
return request.user.is_active | 46 | 1.095238 | 29 | 17 |
35dbfc2202db2f2b3976096c48f85c1df0db7d53 | README.md | README.md |
Welcome to XEd. This is the source for a Docker image I'm working on to be a starting point for setting up an Oracle XE instance.
## Docker cheat sheet
```bash
docker build -t tschf/xed
docker run --shm-size=2g -it tschf/xed
# Removing, per: https://stackoverflow.com/questions/21398087/how-can-i-delete-dockers-images
# Containers
docker rm $(docker ps -a -q)
# Images
docker rmi $(docker images -q)
```
# License
The Unlicense
# Author
Trent Schafer
|
Welcome to XEd. This is the source for a Docker image I'm working on to be a starting point for setting up an Oracle XE instance.
## Docker cheat sheet
```bash
docker build -t tschf/xed
docker run --shm-size=2g -it tschf/xed
# Removing, per: https://stackoverflow.com/questions/21398087/how-can-i-delete-dockers-images
# Containers
docker rm $(docker ps -a -q)
# Images
docker rmi $(docker images -q)
# Stop running containers
docker stop $(docker ps -q)
# Delete images with no tag/name
docker rmi $(docker images -f "dangling=true -q)
```
# License
The Unlicense
# Author
Trent Schafer
| Add more docker commands. Remove dangling images and stop running containers | Add more docker commands.
Remove dangling images and stop running containers
| Markdown | unlicense | tschf/xed | markdown | ## Code Before:
Welcome to XEd. This is the source for a Docker image I'm working on to be a starting point for setting up an Oracle XE instance.
## Docker cheat sheet
```bash
docker build -t tschf/xed
docker run --shm-size=2g -it tschf/xed
# Removing, per: https://stackoverflow.com/questions/21398087/how-can-i-delete-dockers-images
# Containers
docker rm $(docker ps -a -q)
# Images
docker rmi $(docker images -q)
```
# License
The Unlicense
# Author
Trent Schafer
## Instruction:
Add more docker commands.
Remove dangling images and stop running containers
## Code After:
Welcome to XEd. This is the source for a Docker image I'm working on to be a starting point for setting up an Oracle XE instance.
## Docker cheat sheet
```bash
docker build -t tschf/xed
docker run --shm-size=2g -it tschf/xed
# Removing, per: https://stackoverflow.com/questions/21398087/how-can-i-delete-dockers-images
# Containers
docker rm $(docker ps -a -q)
# Images
docker rmi $(docker images -q)
# Stop running containers
docker stop $(docker ps -q)
# Delete images with no tag/name
docker rmi $(docker images -f "dangling=true -q)
```
# License
The Unlicense
# Author
Trent Schafer
|
Welcome to XEd. This is the source for a Docker image I'm working on to be a starting point for setting up an Oracle XE instance.
## Docker cheat sheet
```bash
docker build -t tschf/xed
docker run --shm-size=2g -it tschf/xed
# Removing, per: https://stackoverflow.com/questions/21398087/how-can-i-delete-dockers-images
# Containers
docker rm $(docker ps -a -q)
# Images
docker rmi $(docker images -q)
+ # Stop running containers
+ docker stop $(docker ps -q)
+ # Delete images with no tag/name
+ docker rmi $(docker images -f "dangling=true -q)
```
# License
The Unlicense
# Author
Trent Schafer
| 4 | 0.16 | 4 | 0 |
0eb7ddce9f425c30c70bc1442618deb72c530911 | networks/models.py | networks/models.py | from django.db import models
from helpers import models as helpermodels
# Create your models here.
class Networks(models.Model):
POLICIES = (
('reject', 'Reject'),
('drop', 'Ignore'),
('accept', 'Accept'),
)
name = models.CharField(max_length=30)
interface = models.CharField(max_length=10)
ip_range = helpermodels.IPNetworkField()
policy = models.CharField("default policy", choices=POLICIES, max_length=6)
| from django.db import models
from helpers.models import IPNetworkField
# Create your models here.
class Network(models.Model):
POLICIES = (
('reject', 'Reject'),
('drop', 'Ignore'),
('accept', 'Accept'),
)
name = models.CharField(max_length=30)
interface = models.CharField(max_length=10)
ip_range = IPNetworkField()
policy = models.CharField("default policy", choices=POLICIES, max_length=6)
| Fix Network model name; better import | Fix Network model name; better import
| Python | mit | Kromey/piroute,Kromey/piroute,Kromey/piroute | python | ## Code Before:
from django.db import models
from helpers import models as helpermodels
# Create your models here.
class Networks(models.Model):
POLICIES = (
('reject', 'Reject'),
('drop', 'Ignore'),
('accept', 'Accept'),
)
name = models.CharField(max_length=30)
interface = models.CharField(max_length=10)
ip_range = helpermodels.IPNetworkField()
policy = models.CharField("default policy", choices=POLICIES, max_length=6)
## Instruction:
Fix Network model name; better import
## Code After:
from django.db import models
from helpers.models import IPNetworkField
# Create your models here.
class Network(models.Model):
POLICIES = (
('reject', 'Reject'),
('drop', 'Ignore'),
('accept', 'Accept'),
)
name = models.CharField(max_length=30)
interface = models.CharField(max_length=10)
ip_range = IPNetworkField()
policy = models.CharField("default policy", choices=POLICIES, max_length=6)
| from django.db import models
- from helpers import models as helpermodels
+ from helpers.models import IPNetworkField
# Create your models here.
- class Networks(models.Model):
? -
+ class Network(models.Model):
POLICIES = (
('reject', 'Reject'),
('drop', 'Ignore'),
('accept', 'Accept'),
)
name = models.CharField(max_length=30)
interface = models.CharField(max_length=10)
- ip_range = helpermodels.IPNetworkField()
? -------------
+ ip_range = IPNetworkField()
policy = models.CharField("default policy", choices=POLICIES, max_length=6)
| 6 | 0.333333 | 3 | 3 |
31f9ab16a1370751841abd5b0852ab040b5a7654 | src/res/load_resources.asm | src/res/load_resources.asm | // load sid music
.var music = LoadSid("res/jeff_donald.sid")
//.var music = LoadSid("res/demo.sid")
.pc = music.location
.fill music.size, music.getData(i)
.var picture = LoadBinary("res/dcc.prg")
.pc = $2000 - 2 "Bitmap Data"
.fill picture.getSize(), picture.get(i) | // load sid music
//.var music = LoadSid("res/jeff_donald.sid")
.var music = LoadSid("res/demo.sid")
.pc = music.location "Music"
.fill music.size, music.getData(i)
.pc = $2000 - 2 "Bitmap Data"
.var picture = LoadBinary("res/dcc.prg")
.fill picture.getSize(), picture.get(i)
//----------------------------------------------------------
// Print the music info while assembling
.print ""
.print "SID Data"
.print "--------"
.print "location=$"+toHexString(music.location)
.print "init=$"+toHexString(music.init)
.print "play=$"+toHexString(music.play)
.print "songs="+music.songs
.print "startSong="+music.startSong
.print "size=$"+toHexString(music.size)
.print "name="+music.name
.print "author="+music.author
.print "copyright="+music.copyright
.print ""
.print "Additional tech data"
.print "--------------------"
.print "header="+music.header
.print "header version="+music.version
.print "flags="+toBinaryString(music.flags)
.print "speed="+toBinaryString(music.speed)
.print "startpage="+music.startpage
.print "pagelength="+music.pagelength
| Print additional info about SID file | Print additional info about SID file
| Assembly | cc0-1.0 | c64scene-ar/c64demo,c64scene-ar/c64demo,c64scene-ar/c64demo | assembly | ## Code Before:
// load sid music
.var music = LoadSid("res/jeff_donald.sid")
//.var music = LoadSid("res/demo.sid")
.pc = music.location
.fill music.size, music.getData(i)
.var picture = LoadBinary("res/dcc.prg")
.pc = $2000 - 2 "Bitmap Data"
.fill picture.getSize(), picture.get(i)
## Instruction:
Print additional info about SID file
## Code After:
// load sid music
//.var music = LoadSid("res/jeff_donald.sid")
.var music = LoadSid("res/demo.sid")
.pc = music.location "Music"
.fill music.size, music.getData(i)
.pc = $2000 - 2 "Bitmap Data"
.var picture = LoadBinary("res/dcc.prg")
.fill picture.getSize(), picture.get(i)
//----------------------------------------------------------
// Print the music info while assembling
.print ""
.print "SID Data"
.print "--------"
.print "location=$"+toHexString(music.location)
.print "init=$"+toHexString(music.init)
.print "play=$"+toHexString(music.play)
.print "songs="+music.songs
.print "startSong="+music.startSong
.print "size=$"+toHexString(music.size)
.print "name="+music.name
.print "author="+music.author
.print "copyright="+music.copyright
.print ""
.print "Additional tech data"
.print "--------------------"
.print "header="+music.header
.print "header version="+music.version
.print "flags="+toBinaryString(music.flags)
.print "speed="+toBinaryString(music.speed)
.print "startpage="+music.startpage
.print "pagelength="+music.pagelength
| // load sid music
- .var music = LoadSid("res/jeff_donald.sid")
+ //.var music = LoadSid("res/jeff_donald.sid")
? ++
- //.var music = LoadSid("res/demo.sid")
? --
+ .var music = LoadSid("res/demo.sid")
- .pc = music.location
+ .pc = music.location "Music"
? ++++++++
.fill music.size, music.getData(i)
- .var picture = LoadBinary("res/dcc.prg")
.pc = $2000 - 2 "Bitmap Data"
+ .var picture = LoadBinary("res/dcc.prg")
.fill picture.getSize(), picture.get(i)
+
+ //----------------------------------------------------------
+ // Print the music info while assembling
+ .print ""
+ .print "SID Data"
+ .print "--------"
+ .print "location=$"+toHexString(music.location)
+ .print "init=$"+toHexString(music.init)
+ .print "play=$"+toHexString(music.play)
+ .print "songs="+music.songs
+ .print "startSong="+music.startSong
+ .print "size=$"+toHexString(music.size)
+ .print "name="+music.name
+ .print "author="+music.author
+ .print "copyright="+music.copyright
+
+ .print ""
+ .print "Additional tech data"
+ .print "--------------------"
+ .print "header="+music.header
+ .print "header version="+music.version
+ .print "flags="+toBinaryString(music.flags)
+ .print "speed="+toBinaryString(music.speed)
+ .print "startpage="+music.startpage
+ .print "pagelength="+music.pagelength | 33 | 3.3 | 29 | 4 |
7410758343071dd92a77d63b9c1124805eeef8f2 | jobs/haproxy/templates/haproxy.config.erb | jobs/haproxy/templates/haproxy.config.erb | global
log 127.0.0.1 local1 info
daemon
user vcap
group vcap
maxconn 64000
defaults
log global
timeout connect 30000ms
timeout client <%= p("request_timeout_in_seconds").to_i * 1000 %>ms
timeout server <%= p("request_timeout_in_seconds").to_i * 1000 %>ms
listen mysql-cluster
stick-table type ip size 1
stick on dst
bind 0.0.0.0:3306
option httpchk GET / HTTP/1.1\r\nHost:\ www
mode tcp
option tcplog
<% p('mysql_ips').each do |ip| %>
server mysql-0 <%= ip %>:3306 check port 9200 inter 5000 rise 2 fall 1
<% end %>
listen stats :1936
mode http
stats enable
stats uri /
stats auth admin:<%= p("haproxy_stats_password") %>
| global
log 127.0.0.1 local1 info
daemon
user vcap
group vcap
maxconn 64000
defaults
log global
timeout connect 30000ms
timeout client <%= p("request_timeout_in_seconds").to_i * 1000 %>ms
timeout server <%= p("request_timeout_in_seconds").to_i * 1000 %>ms
listen mysql-cluster
stick-table type ip size 65000
stick on dst
bind 0.0.0.0:3306
option httpchk GET / HTTP/1.1\r\nHost:\ www
mode tcp
option tcplog
<% p('mysql_ips').each_with_index do |ip, idx| %>
server mysql-<%= idx %> <%= ip %>:3306 check port 9200 inter 5000 rise 2 fall 1
<% end %>
listen stats :1936
mode http
stats enable
stats uri /
stats auth admin:<%= p("haproxy_stats_password") %>
| Increase HAProxy stick table size. | Increase HAProxy stick table size.
[#77823944]
| HTML+ERB | apache-2.0 | krishna-mk/cf-mysql-release,cloudfoundry/cf-mysql-release,krishna-mk/cf-mysql-release,CloudCredo/cf-mysql-release,ruanbinfeng/cf-mysql-release,cloudfoundry/cf-mysql-release,CloudCredo/cf-mysql-release,ruanbinfeng/cf-mysql-release,kbastani/cf-mysql-release,krishna-mk/cf-mysql-release,ruanbinfeng/cf-mysql-release,cloudfoundry/cf-mysql-release,kbastani/cf-mysql-release,CloudCredo/cf-mysql-release,cloudfoundry/cf-mysql-release,kbastani/cf-mysql-release,ruanbinfeng/cf-mysql-release | html+erb | ## Code Before:
global
log 127.0.0.1 local1 info
daemon
user vcap
group vcap
maxconn 64000
defaults
log global
timeout connect 30000ms
timeout client <%= p("request_timeout_in_seconds").to_i * 1000 %>ms
timeout server <%= p("request_timeout_in_seconds").to_i * 1000 %>ms
listen mysql-cluster
stick-table type ip size 1
stick on dst
bind 0.0.0.0:3306
option httpchk GET / HTTP/1.1\r\nHost:\ www
mode tcp
option tcplog
<% p('mysql_ips').each do |ip| %>
server mysql-0 <%= ip %>:3306 check port 9200 inter 5000 rise 2 fall 1
<% end %>
listen stats :1936
mode http
stats enable
stats uri /
stats auth admin:<%= p("haproxy_stats_password") %>
## Instruction:
Increase HAProxy stick table size.
[#77823944]
## Code After:
global
log 127.0.0.1 local1 info
daemon
user vcap
group vcap
maxconn 64000
defaults
log global
timeout connect 30000ms
timeout client <%= p("request_timeout_in_seconds").to_i * 1000 %>ms
timeout server <%= p("request_timeout_in_seconds").to_i * 1000 %>ms
listen mysql-cluster
stick-table type ip size 65000
stick on dst
bind 0.0.0.0:3306
option httpchk GET / HTTP/1.1\r\nHost:\ www
mode tcp
option tcplog
<% p('mysql_ips').each_with_index do |ip, idx| %>
server mysql-<%= idx %> <%= ip %>:3306 check port 9200 inter 5000 rise 2 fall 1
<% end %>
listen stats :1936
mode http
stats enable
stats uri /
stats auth admin:<%= p("haproxy_stats_password") %>
| global
log 127.0.0.1 local1 info
daemon
user vcap
group vcap
maxconn 64000
defaults
log global
timeout connect 30000ms
timeout client <%= p("request_timeout_in_seconds").to_i * 1000 %>ms
timeout server <%= p("request_timeout_in_seconds").to_i * 1000 %>ms
listen mysql-cluster
- stick-table type ip size 1
? ^
+ stick-table type ip size 65000
? ^^^^^
stick on dst
bind 0.0.0.0:3306
option httpchk GET / HTTP/1.1\r\nHost:\ www
mode tcp
option tcplog
- <% p('mysql_ips').each do |ip| %>
+ <% p('mysql_ips').each_with_index do |ip, idx| %>
? +++++++++++ +++++
- server mysql-0 <%= ip %>:3306 check port 9200 inter 5000 rise 2 fall 1
? ^
+ server mysql-<%= idx %> <%= ip %>:3306 check port 9200 inter 5000 rise 2 fall 1
? ^^^^^^^^^^
<% end %>
listen stats :1936
mode http
stats enable
stats uri /
stats auth admin:<%= p("haproxy_stats_password") %> | 6 | 0.206897 | 3 | 3 |
6a07e4b3b0c67c442f0501f19c0253b4c99637cd | setup.py | setup.py | from setuptools import setup
import transip
setup(
name = transip.__name__,
version = transip.__version__,
author = transip.__author__,
author_email = transip.__email__,
license = transip.__license__,
description = transip.__doc__.splitlines()[0],
long_description = open('README.rst').read(),
url = 'http://github.com/goabout/transip-backup',
download_url = 'http://github.com/goabout/transip-backup/archives/master',
packages = ['transip', 'transip.service'],
include_package_data = True,
zip_safe = False,
platforms = ['all'],
test_suite = 'tests',
entry_points = {
'console_scripts': [
'transip-api = transip.transip_cli:main',
],
},
install_requires = [
'requests',
],
)
| from setuptools import setup
import transip
setup(
name = transip.__name__,
version = transip.__version__,
author = transip.__author__,
author_email = transip.__email__,
license = transip.__license__,
description = transip.__doc__.splitlines()[0],
long_description = open('README.rst').read(),
url = 'http://github.com/goabout/transip-backup',
download_url = 'http://github.com/goabout/transip-backup/archives/master',
packages = ['transip', 'transip.service'],
include_package_data = True,
zip_safe = False,
platforms = ['all'],
test_suite = 'tests',
entry_points = {
'console_scripts': [
'transip-api = transip.transip_cli:main',
],
},
install_requires = [
'requests',
'rsa',
'suds',
],
)
| Add more requirements that are needed for this module | Add more requirements that are needed for this module
| Python | mit | benkonrath/transip-api,benkonrath/transip-api | python | ## Code Before:
from setuptools import setup
import transip
setup(
name = transip.__name__,
version = transip.__version__,
author = transip.__author__,
author_email = transip.__email__,
license = transip.__license__,
description = transip.__doc__.splitlines()[0],
long_description = open('README.rst').read(),
url = 'http://github.com/goabout/transip-backup',
download_url = 'http://github.com/goabout/transip-backup/archives/master',
packages = ['transip', 'transip.service'],
include_package_data = True,
zip_safe = False,
platforms = ['all'],
test_suite = 'tests',
entry_points = {
'console_scripts': [
'transip-api = transip.transip_cli:main',
],
},
install_requires = [
'requests',
],
)
## Instruction:
Add more requirements that are needed for this module
## Code After:
from setuptools import setup
import transip
setup(
name = transip.__name__,
version = transip.__version__,
author = transip.__author__,
author_email = transip.__email__,
license = transip.__license__,
description = transip.__doc__.splitlines()[0],
long_description = open('README.rst').read(),
url = 'http://github.com/goabout/transip-backup',
download_url = 'http://github.com/goabout/transip-backup/archives/master',
packages = ['transip', 'transip.service'],
include_package_data = True,
zip_safe = False,
platforms = ['all'],
test_suite = 'tests',
entry_points = {
'console_scripts': [
'transip-api = transip.transip_cli:main',
],
},
install_requires = [
'requests',
'rsa',
'suds',
],
)
| from setuptools import setup
import transip
setup(
name = transip.__name__,
version = transip.__version__,
author = transip.__author__,
author_email = transip.__email__,
license = transip.__license__,
description = transip.__doc__.splitlines()[0],
long_description = open('README.rst').read(),
url = 'http://github.com/goabout/transip-backup',
download_url = 'http://github.com/goabout/transip-backup/archives/master',
packages = ['transip', 'transip.service'],
include_package_data = True,
zip_safe = False,
platforms = ['all'],
test_suite = 'tests',
entry_points = {
'console_scripts': [
'transip-api = transip.transip_cli:main',
],
},
install_requires = [
'requests',
+ 'rsa',
+ 'suds',
],
)
| 2 | 0.068966 | 2 | 0 |
8b30c221c20d35e3d217b7c695bcf0d490511076 | .travis.yml | .travis.yml | language: go
go:
- tip
install:
- go get github.com/onsi/gomega
- go get github.com/onsi/ginkgo/ginkgo
script:
- $GOPATH/bin/ginkgo github.com/petergtz/pegomock/pegomock/mockgen
- $GOPATH/bin/ginkgo -r --randomizeAllSpecs --randomizeSuites --race --trace
| language: go
go:
- tip
install:
- go get github.com/onsi/gomega
- go get github.com/onsi/ginkgo/ginkgo
- go get gopkg.in/alecthomas/kingpin.v2
script:
- $GOPATH/bin/ginkgo github.com/petergtz/pegomock/pegomock/mockgen
- $GOPATH/bin/ginkgo -r --randomizeAllSpecs --randomizeSuites --race --trace
| Add kingpin as a dependency | Add kingpin as a dependency
| YAML | apache-2.0 | petergtz/pegomock,petergtz/pegomock | yaml | ## Code Before:
language: go
go:
- tip
install:
- go get github.com/onsi/gomega
- go get github.com/onsi/ginkgo/ginkgo
script:
- $GOPATH/bin/ginkgo github.com/petergtz/pegomock/pegomock/mockgen
- $GOPATH/bin/ginkgo -r --randomizeAllSpecs --randomizeSuites --race --trace
## Instruction:
Add kingpin as a dependency
## Code After:
language: go
go:
- tip
install:
- go get github.com/onsi/gomega
- go get github.com/onsi/ginkgo/ginkgo
- go get gopkg.in/alecthomas/kingpin.v2
script:
- $GOPATH/bin/ginkgo github.com/petergtz/pegomock/pegomock/mockgen
- $GOPATH/bin/ginkgo -r --randomizeAllSpecs --randomizeSuites --race --trace
| language: go
go:
- tip
install:
- go get github.com/onsi/gomega
- go get github.com/onsi/ginkgo/ginkgo
+ - go get gopkg.in/alecthomas/kingpin.v2
script:
- $GOPATH/bin/ginkgo github.com/petergtz/pegomock/pegomock/mockgen
- $GOPATH/bin/ginkgo -r --randomizeAllSpecs --randomizeSuites --race --trace | 1 | 0.090909 | 1 | 0 |
417ff7688af3c9646b3c55eaaf6b18818f1cfd8f | packages/sandforms-test-helpers/server/test-helpers.js | packages/sandforms-test-helpers/server/test-helpers.js | function clearAllCollections() {
Prompts.remove({});
Submissions.remove({});
}
Meteor.methods({
clearAllCollections: clearAllCollections,
setupFixtures: function() {
clearAllCollections();
var id = Prompts.insert("test prompt");
Submissions.insert({
promptId: id,
response: 'test response'
});
}
});
| function clearAllCollections() {
Prompts.remove({});
Submissions.remove({});
}
Meteor.methods({
clearAllCollections: clearAllCollections,
setupFixtures: function() {
clearAllCollections();
var id = Prompts.create("test prompt");
Submissions.insert({
responses: [{
promptId: id,
response: 'test response'
}]
});
}
}); | Fix JSON formation in test helper methods | Berke+Michael: Fix JSON formation in test helper methods
| JavaScript | apache-2.0 | sandforms/sandforms,sandforms/sandforms,sandforms/sandforms,sandforms/sandforms | javascript | ## Code Before:
function clearAllCollections() {
Prompts.remove({});
Submissions.remove({});
}
Meteor.methods({
clearAllCollections: clearAllCollections,
setupFixtures: function() {
clearAllCollections();
var id = Prompts.insert("test prompt");
Submissions.insert({
promptId: id,
response: 'test response'
});
}
});
## Instruction:
Berke+Michael: Fix JSON formation in test helper methods
## Code After:
function clearAllCollections() {
Prompts.remove({});
Submissions.remove({});
}
Meteor.methods({
clearAllCollections: clearAllCollections,
setupFixtures: function() {
clearAllCollections();
var id = Prompts.create("test prompt");
Submissions.insert({
responses: [{
promptId: id,
response: 'test response'
}]
});
}
}); | function clearAllCollections() {
Prompts.remove({});
Submissions.remove({});
}
Meteor.methods({
clearAllCollections: clearAllCollections,
setupFixtures: function() {
clearAllCollections();
- var id = Prompts.insert("test prompt");
? ^^^ ^
+ var id = Prompts.create("test prompt");
? ^^ ^ +
Submissions.insert({
+ responses: [{
- promptId: id,
+ promptId: id,
? ++
- response: 'test response'
+ response: 'test response'
? ++
+ }]
});
}
}); | 8 | 0.421053 | 5 | 3 |
f94c323fa6d36563e8b0bcab30cb9fd0579279cd | concepts/Views/Partials.md | concepts/Views/Partials.md |
TODO: document how partials work w/ EJS, and clarify that partial syntax in other view engines must follow the conventions/syntax of that view engine. Only EJS partials are officially documented (since EJS is the default view engine in Sails)
<docmeta name="uniqueID" value="Partials610916">
<docmeta name="displayName" value="Partials">
|
Sails uses `ejs-locals` in its view rendering code, so in your views you can do:
```
<%- partial ('foo.ejs') %>
```
to render a partial located at `/views/foo.ejs`. All of your locals will be sent to the partial automatically.
the paths are relative to the view, that is loading the partial. So if you have a a user view at `/views/users/view.ejs` and want to load `/views/partials/widget.ejs` then you would use:
```
<%- partial ('../../partials/widget.ejs') %>
```
One thing to note: partials are rendered synchronously, so they will block Sails from serving more requests until they're done loading. It's something to keep in mind while developing your app, especially if you anticipate a large number of connections.
NOTE: When using other templating languages than ejs, their syntax for loading partials or block, etc. will be used. Please refer to their documentation for more information on their syntax and conventions
<docmeta name="uniqueID" value="Partials610916">
<docmeta name="displayName" value="Partials">
| Add documentation for ejs partials | Add documentation for ejs partials
We need to show the documentation for how to load partials when using ejs. As according to the default ejs partial syntax, the usage here is slightly different due to the use of ejs-locals and thay may not be immediately apparent to end users. | Markdown | mit | yoshioka-s/sails-docs-ja | markdown | ## Code Before:
TODO: document how partials work w/ EJS, and clarify that partial syntax in other view engines must follow the conventions/syntax of that view engine. Only EJS partials are officially documented (since EJS is the default view engine in Sails)
<docmeta name="uniqueID" value="Partials610916">
<docmeta name="displayName" value="Partials">
## Instruction:
Add documentation for ejs partials
We need to show the documentation for how to load partials when using ejs. As according to the default ejs partial syntax, the usage here is slightly different due to the use of ejs-locals and thay may not be immediately apparent to end users.
## Code After:
Sails uses `ejs-locals` in its view rendering code, so in your views you can do:
```
<%- partial ('foo.ejs') %>
```
to render a partial located at `/views/foo.ejs`. All of your locals will be sent to the partial automatically.
the paths are relative to the view, that is loading the partial. So if you have a a user view at `/views/users/view.ejs` and want to load `/views/partials/widget.ejs` then you would use:
```
<%- partial ('../../partials/widget.ejs') %>
```
One thing to note: partials are rendered synchronously, so they will block Sails from serving more requests until they're done loading. It's something to keep in mind while developing your app, especially if you anticipate a large number of connections.
NOTE: When using other templating languages than ejs, their syntax for loading partials or block, etc. will be used. Please refer to their documentation for more information on their syntax and conventions
<docmeta name="uniqueID" value="Partials610916">
<docmeta name="displayName" value="Partials">
|
- TODO: document how partials work w/ EJS, and clarify that partial syntax in other view engines must follow the conventions/syntax of that view engine. Only EJS partials are officially documented (since EJS is the default view engine in Sails)
+ Sails uses `ejs-locals` in its view rendering code, so in your views you can do:
+ ```
+ <%- partial ('foo.ejs') %>
+ ```
+
+ to render a partial located at `/views/foo.ejs`. All of your locals will be sent to the partial automatically.
+
+ the paths are relative to the view, that is loading the partial. So if you have a a user view at `/views/users/view.ejs` and want to load `/views/partials/widget.ejs` then you would use:
+
+ ```
+ <%- partial ('../../partials/widget.ejs') %>
+ ```
+
+ One thing to note: partials are rendered synchronously, so they will block Sails from serving more requests until they're done loading. It's something to keep in mind while developing your app, especially if you anticipate a large number of connections.
+
+ NOTE: When using other templating languages than ejs, their syntax for loading partials or block, etc. will be used. Please refer to their documentation for more information on their syntax and conventions
<docmeta name="uniqueID" value="Partials610916">
<docmeta name="displayName" value="Partials">
| 17 | 2.428571 | 16 | 1 |
8350be69c8afec613758e8dc3649eded52dcda0f | README.md | README.md |
remiii-aws-download-check is licensed under the MIT license (see LICENSE.md file).
## Authors
* Rémi barbe (aka Remiii)
|
remiii-aws-download-check is licensed under the MIT license (see LICENSE.md file).
## Authors
* Rémi barbe (aka Remiii)
* Laurent (aka lologhi)
| Add lologhi to contrib list | Add lologhi to contrib list
| Markdown | mit | Remiii/remiii-aws-download-check | markdown | ## Code Before:
remiii-aws-download-check is licensed under the MIT license (see LICENSE.md file).
## Authors
* Rémi barbe (aka Remiii)
## Instruction:
Add lologhi to contrib list
## Code After:
remiii-aws-download-check is licensed under the MIT license (see LICENSE.md file).
## Authors
* Rémi barbe (aka Remiii)
* Laurent (aka lologhi)
|
remiii-aws-download-check is licensed under the MIT license (see LICENSE.md file).
## Authors
* Rémi barbe (aka Remiii)
+ * Laurent (aka lologhi)
| 1 | 0.142857 | 1 | 0 |
65e6a1b4c944c89258ca6ad07b618042d9908d9d | .travis.yml | .travis.yml | sudo: required
services:
- docker
env:
- NODE_VERSION=latest
- NODE_VERSION=lts
- NODE_VERSION=old
script:
- make build-$NODE_VERSION
- make test-all-$NODE_VERSION
deploy:
provider: script
script: docker login -e $DOCKER_EMAIL -u $DOCKER_USERNAME -p $DOCKER_PASSWORD && make push-$NODE_VERSION
on:
repo: 'umweltdk/docker-node'
branch: master
| sudo: required
services:
- docker
env:
- NODE_VERSION=latest
- NODE_VERSION=lts
- NODE_VERSION=old
- NODE_VERSION=4.3.2
- NODE_VERSION=6.10.2
script:
- make build-$NODE_VERSION
- make test-all-$NODE_VERSION
deploy:
provider: script
script: docker login -e $DOCKER_EMAIL -u $DOCKER_USERNAME -p $DOCKER_PASSWORD && make push-$NODE_VERSION
on:
repo: 'umweltdk/docker-node'
branch: master
| Build aws lambda specific node versions | Build aws lambda specific node versions | YAML | mit | umweltdk/docker-node | yaml | ## Code Before:
sudo: required
services:
- docker
env:
- NODE_VERSION=latest
- NODE_VERSION=lts
- NODE_VERSION=old
script:
- make build-$NODE_VERSION
- make test-all-$NODE_VERSION
deploy:
provider: script
script: docker login -e $DOCKER_EMAIL -u $DOCKER_USERNAME -p $DOCKER_PASSWORD && make push-$NODE_VERSION
on:
repo: 'umweltdk/docker-node'
branch: master
## Instruction:
Build aws lambda specific node versions
## Code After:
sudo: required
services:
- docker
env:
- NODE_VERSION=latest
- NODE_VERSION=lts
- NODE_VERSION=old
- NODE_VERSION=4.3.2
- NODE_VERSION=6.10.2
script:
- make build-$NODE_VERSION
- make test-all-$NODE_VERSION
deploy:
provider: script
script: docker login -e $DOCKER_EMAIL -u $DOCKER_USERNAME -p $DOCKER_PASSWORD && make push-$NODE_VERSION
on:
repo: 'umweltdk/docker-node'
branch: master
| sudo: required
services:
- docker
env:
- NODE_VERSION=latest
- NODE_VERSION=lts
- NODE_VERSION=old
+ - NODE_VERSION=4.3.2
+ - NODE_VERSION=6.10.2
script:
- make build-$NODE_VERSION
- make test-all-$NODE_VERSION
deploy:
provider: script
script: docker login -e $DOCKER_EMAIL -u $DOCKER_USERNAME -p $DOCKER_PASSWORD && make push-$NODE_VERSION
on:
repo: 'umweltdk/docker-node'
branch: master | 2 | 0.1 | 2 | 0 |
79737d525ee22b2602082b6cfe81aaa3982ea71b | README.md | README.md |
Lets build something like that:
<a href="http://www.youtube.com/watch?feature=player_embedded&v=_RvHz3vJ3kA
" target="_blank"><img src="http://img.youtube.com/vi/_RvHz3vJ3kA/0.jpg"
alt="IMAGE ALT TEXT HERE" width="240" height="180" border="10" /></a>
[Emacs expand-region](https://github.com/magnars/expand-region.el)
[Extend Selection by Semantic Unit](http://ergoemacs.org/emacs/syntax_tree_walk.html) | [](https://travis-ci.org/aronwoost/sublime-expand-region)
# Work in progress - DON'T USE
Lets build something like that:
<a href="http://www.youtube.com/watch?feature=player_embedded&v=_RvHz3vJ3kA
" target="_blank"><img src="http://img.youtube.com/vi/_RvHz3vJ3kA/0.jpg"
alt="IMAGE ALT TEXT HERE" width="240" height="180" border="10" /></a>
[Emacs expand-region](https://github.com/magnars/expand-region.el)
[Extend Selection by Semantic Unit](http://ergoemacs.org/emacs/syntax_tree_walk.html) | Add travis build status image | Add travis build status image
| Markdown | mit | johyphenel/sublime-expand-region,johyphenel/sublime-expand-region,aronwoost/sublime-expand-region | markdown | ## Code Before:
Lets build something like that:
<a href="http://www.youtube.com/watch?feature=player_embedded&v=_RvHz3vJ3kA
" target="_blank"><img src="http://img.youtube.com/vi/_RvHz3vJ3kA/0.jpg"
alt="IMAGE ALT TEXT HERE" width="240" height="180" border="10" /></a>
[Emacs expand-region](https://github.com/magnars/expand-region.el)
[Extend Selection by Semantic Unit](http://ergoemacs.org/emacs/syntax_tree_walk.html)
## Instruction:
Add travis build status image
## Code After:
[](https://travis-ci.org/aronwoost/sublime-expand-region)
# Work in progress - DON'T USE
Lets build something like that:
<a href="http://www.youtube.com/watch?feature=player_embedded&v=_RvHz3vJ3kA
" target="_blank"><img src="http://img.youtube.com/vi/_RvHz3vJ3kA/0.jpg"
alt="IMAGE ALT TEXT HERE" width="240" height="180" border="10" /></a>
[Emacs expand-region](https://github.com/magnars/expand-region.el)
[Extend Selection by Semantic Unit](http://ergoemacs.org/emacs/syntax_tree_walk.html) | + [](https://travis-ci.org/aronwoost/sublime-expand-region)
+
+ # Work in progress - DON'T USE
Lets build something like that:
<a href="http://www.youtube.com/watch?feature=player_embedded&v=_RvHz3vJ3kA
" target="_blank"><img src="http://img.youtube.com/vi/_RvHz3vJ3kA/0.jpg"
alt="IMAGE ALT TEXT HERE" width="240" height="180" border="10" /></a>
[Emacs expand-region](https://github.com/magnars/expand-region.el)
[Extend Selection by Semantic Unit](http://ergoemacs.org/emacs/syntax_tree_walk.html) | 3 | 0.375 | 3 | 0 |
d1ec7464e780ddba2a763eb0bd97d00d7477b103 | GTProgressBarDirection.swift | GTProgressBarDirection.swift | //
// GTProgressBarDirection.swift
// GTProgressBar
//
// Created by greg on 15/11/2017.
//
import Foundation
public enum GTProgressBarDirection: Int {
case anticlocwise
}
| //
// GTProgressBarDirection.swift
// GTProgressBar
//
// Created by greg on 15/11/2017.
//
import Foundation
public enum GTProgressBarDirection: Int {
case clockwise
case anticlockwise
}
| Correct typo and add clockwise direction | Correct typo and add clockwise direction
| Swift | mit | gregttn/GTProgressBar,gregttn/GTProgressBar | swift | ## Code Before:
//
// GTProgressBarDirection.swift
// GTProgressBar
//
// Created by greg on 15/11/2017.
//
import Foundation
public enum GTProgressBarDirection: Int {
case anticlocwise
}
## Instruction:
Correct typo and add clockwise direction
## Code After:
//
// GTProgressBarDirection.swift
// GTProgressBar
//
// Created by greg on 15/11/2017.
//
import Foundation
public enum GTProgressBarDirection: Int {
case clockwise
case anticlockwise
}
| //
// GTProgressBarDirection.swift
// GTProgressBar
//
// Created by greg on 15/11/2017.
//
import Foundation
public enum GTProgressBarDirection: Int {
+ case clockwise
- case anticlocwise
+ case anticlockwise
? +
} | 3 | 0.25 | 2 | 1 |
3d9f31ac4f19c761fe554d4065acb598f590e00b | Gruntfile.js | Gruntfile.js | module.exports = function (grunt) {
'use strict';
var config = require('./grunt_tasks/config');
grunt.config.init(config);
var packages = grunt.file.readJSON('./package.json');
Object.keys(packages.devDependencies).forEach(function (p) {
if (p.match(/^grunt-/)) {
grunt.loadNpmTasks(p);
}
});
grunt.registerTask('test', ['espower', 'karma']);
};
| module.exports = function (grunt) {
'use strict';
var config = require('./grunt_tasks/config');
grunt.config.init(config);
var packages = grunt.file.readJSON('./package.json');
Object.keys(packages.devDependencies).forEach(function (p) {
if (p.match(/^grunt-/)) {
grunt.loadNpmTasks(p);
}
});
grunt.registerTask('test', ['espower', 'karma']);
grunt.registerTask('development', ['sass:development']);
grunt.registerTask('production', ['sass:production']);
};
| Add tasks (development & production) | Add tasks (development & production)
| JavaScript | mit | kubosho/simple-music-player | javascript | ## Code Before:
module.exports = function (grunt) {
'use strict';
var config = require('./grunt_tasks/config');
grunt.config.init(config);
var packages = grunt.file.readJSON('./package.json');
Object.keys(packages.devDependencies).forEach(function (p) {
if (p.match(/^grunt-/)) {
grunt.loadNpmTasks(p);
}
});
grunt.registerTask('test', ['espower', 'karma']);
};
## Instruction:
Add tasks (development & production)
## Code After:
module.exports = function (grunt) {
'use strict';
var config = require('./grunt_tasks/config');
grunt.config.init(config);
var packages = grunt.file.readJSON('./package.json');
Object.keys(packages.devDependencies).forEach(function (p) {
if (p.match(/^grunt-/)) {
grunt.loadNpmTasks(p);
}
});
grunt.registerTask('test', ['espower', 'karma']);
grunt.registerTask('development', ['sass:development']);
grunt.registerTask('production', ['sass:production']);
};
| module.exports = function (grunt) {
'use strict';
var config = require('./grunt_tasks/config');
grunt.config.init(config);
var packages = grunt.file.readJSON('./package.json');
Object.keys(packages.devDependencies).forEach(function (p) {
if (p.match(/^grunt-/)) {
grunt.loadNpmTasks(p);
}
});
grunt.registerTask('test', ['espower', 'karma']);
+ grunt.registerTask('development', ['sass:development']);
+ grunt.registerTask('production', ['sass:production']);
}; | 2 | 0.133333 | 2 | 0 |
4df150257e276f22e4c125c13c276d0f8316f8a9 | src/Mo/FlashCardsBundle/Controller/ApiController.php | src/Mo/FlashCardsBundle/Controller/ApiController.php | <?php
namespace Mo\FlashCardsBundle\Controller;
use Symfony\Bundle\FrameworkBundle\Controller\Controller;
use Symfony\Component\HttpFoundation\JsonResponse;
class ApiController extends Controller
{
public function cardsAction($group)
{
$cards = array(
'dog' => 'doggy',
'cat' => 'kitty'
);
// setup response
$response = new JsonResponse();
$response->setData($cards);
return $response;
}
}
| <?php
namespace Mo\FlashCardsBundle\Controller;
use Symfony\Bundle\FrameworkBundle\Controller\Controller;
use Symfony\Component\HttpFoundation\JsonResponse;
/**
* Actions handling the data flow.
*/
class ApiController extends Controller
{
/**
* @param string $group
* @return \Symfony\Component\HttpFoundation\JsonResponse
*/
public function cardsAction($group)
{
$cards = array(
'dog' => 'doggy',
'cat' => 'kitty'
);
// setup response
$response = new JsonResponse();
$response->setData($cards);
return $response;
}
}
| Add comments to the API controller. | Add comments to the API controller.
| PHP | mit | mishedone/mo-flash-cards,mishedone/mo-flash-cards | php | ## Code Before:
<?php
namespace Mo\FlashCardsBundle\Controller;
use Symfony\Bundle\FrameworkBundle\Controller\Controller;
use Symfony\Component\HttpFoundation\JsonResponse;
class ApiController extends Controller
{
public function cardsAction($group)
{
$cards = array(
'dog' => 'doggy',
'cat' => 'kitty'
);
// setup response
$response = new JsonResponse();
$response->setData($cards);
return $response;
}
}
## Instruction:
Add comments to the API controller.
## Code After:
<?php
namespace Mo\FlashCardsBundle\Controller;
use Symfony\Bundle\FrameworkBundle\Controller\Controller;
use Symfony\Component\HttpFoundation\JsonResponse;
/**
* Actions handling the data flow.
*/
class ApiController extends Controller
{
/**
* @param string $group
* @return \Symfony\Component\HttpFoundation\JsonResponse
*/
public function cardsAction($group)
{
$cards = array(
'dog' => 'doggy',
'cat' => 'kitty'
);
// setup response
$response = new JsonResponse();
$response->setData($cards);
return $response;
}
}
| <?php
namespace Mo\FlashCardsBundle\Controller;
use Symfony\Bundle\FrameworkBundle\Controller\Controller;
use Symfony\Component\HttpFoundation\JsonResponse;
+ /**
+ * Actions handling the data flow.
+ */
class ApiController extends Controller
{
+ /**
+ * @param string $group
+ * @return \Symfony\Component\HttpFoundation\JsonResponse
+ */
public function cardsAction($group)
{
$cards = array(
'dog' => 'doggy',
'cat' => 'kitty'
);
// setup response
$response = new JsonResponse();
$response->setData($cards);
return $response;
}
} | 7 | 0.304348 | 7 | 0 |
6163994ab506933332f19a66d93fee76a060d6f0 | .travis.yml | .travis.yml | language: go
go:
- 1.3
- 1.4
- 1.5
- 1.6
- tip
before_install:
- go get -v ./...
- go get github.com/stretchr/testify/assert
- go get -u github.com/golang/lint/golint
services:
- rabbitmq
env:
- AMQP_URL=amqp://guest:guest@127.0.0.1:5672/
script:
- go build
- golint ./...
- go vet ./...
- go build examples/consumer/consumer.go
- go build examples/publisher/publisher.go
- go test -tags integration ./...
| language: go
go:
- 1.5
- 1.6
- tip
before_install:
- go get -v ./...
- go get github.com/stretchr/testify/assert
- go get -u github.com/golang/lint/golint
services:
- rabbitmq
env:
- AMQP_URL=amqp://guest:guest@127.0.0.1:5672/
script:
- go build
- golint ./...
- go vet ./...
- go build examples/consumer/consumer.go
- go build examples/publisher/publisher.go
- go test -tags integration ./...
| Remove old version of golang. go vet needs 1.5 | Remove old version of golang. go vet needs 1.5
| YAML | mit | aleasoluciones/simpleamqp | yaml | ## Code Before:
language: go
go:
- 1.3
- 1.4
- 1.5
- 1.6
- tip
before_install:
- go get -v ./...
- go get github.com/stretchr/testify/assert
- go get -u github.com/golang/lint/golint
services:
- rabbitmq
env:
- AMQP_URL=amqp://guest:guest@127.0.0.1:5672/
script:
- go build
- golint ./...
- go vet ./...
- go build examples/consumer/consumer.go
- go build examples/publisher/publisher.go
- go test -tags integration ./...
## Instruction:
Remove old version of golang. go vet needs 1.5
## Code After:
language: go
go:
- 1.5
- 1.6
- tip
before_install:
- go get -v ./...
- go get github.com/stretchr/testify/assert
- go get -u github.com/golang/lint/golint
services:
- rabbitmq
env:
- AMQP_URL=amqp://guest:guest@127.0.0.1:5672/
script:
- go build
- golint ./...
- go vet ./...
- go build examples/consumer/consumer.go
- go build examples/publisher/publisher.go
- go test -tags integration ./...
| language: go
go:
- - 1.3
- - 1.4
- 1.5
- 1.6
- tip
before_install:
- go get -v ./...
- go get github.com/stretchr/testify/assert
- go get -u github.com/golang/lint/golint
services:
- rabbitmq
env:
- AMQP_URL=amqp://guest:guest@127.0.0.1:5672/
script:
- go build
- golint ./...
- go vet ./...
- go build examples/consumer/consumer.go
- go build examples/publisher/publisher.go
- go test -tags integration ./... | 2 | 0.074074 | 0 | 2 |
8dcb43a4272563971cd22f32106731519451f83a | desktop/app/components/tasks/directive.js | desktop/app/components/tasks/directive.js | angular.module('MainApp')
.directive('uaf', () => { // uaf = Update after find
return (scope, el, attrs) => {
scope.$watch('tasks', (newVal) => {
el[0].children[0].children[0].innerHTML = newVal[attrs.uaf];
}, true);
};
});
| angular.module('MainApp')
.directive('uaf', () => // uaf = Update after find
(scope, el, attrs) => {
scope.$watch('tasks', (newVal) => {
el[0].children[0].children[0].innerHTML = newVal[attrs.uaf];
}, true);
}
);
| Fix eslint errors - 6 | Fix eslint errors - 6
| JavaScript | mit | mkermani144/wanna,mkermani144/wanna | javascript | ## Code Before:
angular.module('MainApp')
.directive('uaf', () => { // uaf = Update after find
return (scope, el, attrs) => {
scope.$watch('tasks', (newVal) => {
el[0].children[0].children[0].innerHTML = newVal[attrs.uaf];
}, true);
};
});
## Instruction:
Fix eslint errors - 6
## Code After:
angular.module('MainApp')
.directive('uaf', () => // uaf = Update after find
(scope, el, attrs) => {
scope.$watch('tasks', (newVal) => {
el[0].children[0].children[0].innerHTML = newVal[attrs.uaf];
}, true);
}
);
| angular.module('MainApp')
- .directive('uaf', () => { // uaf = Update after find
? -
+ .directive('uaf', () => // uaf = Update after find
- return (scope, el, attrs) => {
? -------
+ (scope, el, attrs) => {
scope.$watch('tasks', (newVal) => {
el[0].children[0].children[0].innerHTML = newVal[attrs.uaf];
}, true);
- };
? -
+ }
- });
? -
+ ); | 8 | 1 | 4 | 4 |
57eaa33d78c35d918516319752eeb15886c6b552 | etc/qed.rb | etc/qed.rb | config :qed, :profile=>:cov do
require 'simplecov'
SimpleCov.start do
coverage_dir 'log/coverage'
#add_group "RSpec", "lib/assay/rspec.rb"
end
end
|
QED.configure 'cov' do
require 'simplecov'
SimpleCov.command_name 'qed'
SimpleCov.start do
add_filter '/demo/'
coverage_dir 'log/coverage'
#add_group "Label", "lib/qed/directory"
end
end
| Move config for QED to etc directory. | :admin: Move config for QED to etc directory.
| Ruby | bsd-2-clause | rubyworks/ansi,rubyworks/ansi,rubyworks/ansi | ruby | ## Code Before:
config :qed, :profile=>:cov do
require 'simplecov'
SimpleCov.start do
coverage_dir 'log/coverage'
#add_group "RSpec", "lib/assay/rspec.rb"
end
end
## Instruction:
:admin: Move config for QED to etc directory.
## Code After:
QED.configure 'cov' do
require 'simplecov'
SimpleCov.command_name 'qed'
SimpleCov.start do
add_filter '/demo/'
coverage_dir 'log/coverage'
#add_group "Label", "lib/qed/directory"
end
end
| - config :qed, :profile=>:cov do
+
+ QED.configure 'cov' do
require 'simplecov'
+ SimpleCov.command_name 'qed'
SimpleCov.start do
+ add_filter '/demo/'
coverage_dir 'log/coverage'
- #add_group "RSpec", "lib/assay/rspec.rb"
+ #add_group "Label", "lib/qed/directory"
end
end
| 7 | 0.875 | 5 | 2 |
735f393b62554d9449b7e98c0fa7afde29e58681 | pt-BR/reminders.php | pt-BR/reminders.php | <?php
return array(
/*
|--------------------------------------------------------------------------
| Password Reminder Language Lines
|--------------------------------------------------------------------------
|
| The following language lines are the default lines which match reasons
| that are given by the password broker for a password update attempt
| has failed, such as for an invalid token or invalid new password.
|
*/
"password" => "A senha deverá conter pelo menos seis carateres e ser igual à confirmação.",
"user" => "Não conseguimos encontrar nenhum usuário com o endereço de email indicado.",
"token" => "Este código de recuperação da senha é inválido.",
"sent" => "O lembrete da senha foi enviado!",
"reset" => "Password has been reset!",
);
| <?php
return array(
/*
|--------------------------------------------------------------------------
| Password Reminder Language Lines
|--------------------------------------------------------------------------
|
| The following language lines are the default lines which match reasons
| that are given by the password broker for a password update attempt
| has failed, such as for an invalid token or invalid new password.
|
*/
"password" => "A senha deverá conter pelo menos seis carateres e ser igual à confirmação.",
"user" => "Não conseguimos encontrar nenhum usuário com o endereço de email indicado.",
"token" => "Este código de recuperação da senha é inválido.",
"sent" => "O lembrete da senha foi enviado!",
"reset" => "A senha foi redefinida!",
);
| Reset in Reminders in Portuguese | Reset in Reminders in Portuguese | PHP | mit | idhamperdameian/Laravel-lang,overtrue-forks/Laravel-lang,overtrue-forks/Laravel-lang | php | ## Code Before:
<?php
return array(
/*
|--------------------------------------------------------------------------
| Password Reminder Language Lines
|--------------------------------------------------------------------------
|
| The following language lines are the default lines which match reasons
| that are given by the password broker for a password update attempt
| has failed, such as for an invalid token or invalid new password.
|
*/
"password" => "A senha deverá conter pelo menos seis carateres e ser igual à confirmação.",
"user" => "Não conseguimos encontrar nenhum usuário com o endereço de email indicado.",
"token" => "Este código de recuperação da senha é inválido.",
"sent" => "O lembrete da senha foi enviado!",
"reset" => "Password has been reset!",
);
## Instruction:
Reset in Reminders in Portuguese
## Code After:
<?php
return array(
/*
|--------------------------------------------------------------------------
| Password Reminder Language Lines
|--------------------------------------------------------------------------
|
| The following language lines are the default lines which match reasons
| that are given by the password broker for a password update attempt
| has failed, such as for an invalid token or invalid new password.
|
*/
"password" => "A senha deverá conter pelo menos seis carateres e ser igual à confirmação.",
"user" => "Não conseguimos encontrar nenhum usuário com o endereço de email indicado.",
"token" => "Este código de recuperação da senha é inválido.",
"sent" => "O lembrete da senha foi enviado!",
"reset" => "A senha foi redefinida!",
);
| <?php
return array(
/*
|--------------------------------------------------------------------------
| Password Reminder Language Lines
|--------------------------------------------------------------------------
|
| The following language lines are the default lines which match reasons
| that are given by the password broker for a password update attempt
| has failed, such as for an invalid token or invalid new password.
|
*/
"password" => "A senha deverá conter pelo menos seis carateres e ser igual à confirmação.",
"user" => "Não conseguimos encontrar nenhum usuário com o endereço de email indicado.",
"token" => "Este código de recuperação da senha é inválido.",
"sent" => "O lembrete da senha foi enviado!",
- "reset" => "Password has been reset!",
+ "reset" => "A senha foi redefinida!",
); | 2 | 0.08 | 1 | 1 |
27cab6b010e640308002c01e7ca606f1ab10a993 | packages/deb/jenkins.sh | packages/deb/jenkins.sh |
set -o nounset
set -o errexit
set -o nounset
set -o xtrace
declare -r BUILD_TAG="$(date '+%y%m%d%H%M%S')"
declare -r IMG_NAME="debian-builder:${BUILD_TAG}"
declare -r DEB_RELEASE_BUCKET="gs://k8s-release-dev/debian"
docker build -t "${IMG_NAME}" "$( cd "$( dirname "${BASH_SOURCE[0]}" )" && pwd )"
docker run -it --rm -v "${PWD}/bin:/src/bin" "${IMG_NAME}" $@
gsutil -m cp -nrc bin "${DEB_RELEASE_BUCKET}/${BUILD_TAG}"
printf "%s" "${BUILD_TAG}" | gsutil cp - "${DEB_RELEASE_BUCKET}/latest"
|
set -o nounset
set -o errexit
set -o nounset
set -o xtrace
declare -r BUILD_TAG="$(date '+%y%m%d%H%M%S')"
declare -r IMG_NAME="debian-builder:${BUILD_TAG}"
docker build -t "${IMG_NAME}" "$( cd "$( dirname "${BASH_SOURCE[0]}" )" && pwd )"
docker run -it --rm -v "${PWD}/bin:/src/bin" "${IMG_NAME}" $@
printf "%s" "${BUILD_TAG}" | gsutil cp - "${DEB_RELEASE_BUCKET}/latest"
| Stop uploading .deb packages to scratch bucket | Stop uploading .deb packages to scratch bucket
This commit removes the transient upload of .deb files
to a scratch bucket, probably a remnant of an older
process.
Signed-off-by: Adolfo García Veytia (Puerco) <9de42c8261a7cacaae239656a4dd6991d9b9250f@uservers.net>
| Shell | apache-2.0 | kubernetes/release,kubernetes/release | shell | ## Code Before:
set -o nounset
set -o errexit
set -o nounset
set -o xtrace
declare -r BUILD_TAG="$(date '+%y%m%d%H%M%S')"
declare -r IMG_NAME="debian-builder:${BUILD_TAG}"
declare -r DEB_RELEASE_BUCKET="gs://k8s-release-dev/debian"
docker build -t "${IMG_NAME}" "$( cd "$( dirname "${BASH_SOURCE[0]}" )" && pwd )"
docker run -it --rm -v "${PWD}/bin:/src/bin" "${IMG_NAME}" $@
gsutil -m cp -nrc bin "${DEB_RELEASE_BUCKET}/${BUILD_TAG}"
printf "%s" "${BUILD_TAG}" | gsutil cp - "${DEB_RELEASE_BUCKET}/latest"
## Instruction:
Stop uploading .deb packages to scratch bucket
This commit removes the transient upload of .deb files
to a scratch bucket, probably a remnant of an older
process.
Signed-off-by: Adolfo García Veytia (Puerco) <9de42c8261a7cacaae239656a4dd6991d9b9250f@uservers.net>
## Code After:
set -o nounset
set -o errexit
set -o nounset
set -o xtrace
declare -r BUILD_TAG="$(date '+%y%m%d%H%M%S')"
declare -r IMG_NAME="debian-builder:${BUILD_TAG}"
docker build -t "${IMG_NAME}" "$( cd "$( dirname "${BASH_SOURCE[0]}" )" && pwd )"
docker run -it --rm -v "${PWD}/bin:/src/bin" "${IMG_NAME}" $@
printf "%s" "${BUILD_TAG}" | gsutil cp - "${DEB_RELEASE_BUCKET}/latest"
|
set -o nounset
set -o errexit
set -o nounset
set -o xtrace
declare -r BUILD_TAG="$(date '+%y%m%d%H%M%S')"
declare -r IMG_NAME="debian-builder:${BUILD_TAG}"
- declare -r DEB_RELEASE_BUCKET="gs://k8s-release-dev/debian"
docker build -t "${IMG_NAME}" "$( cd "$( dirname "${BASH_SOURCE[0]}" )" && pwd )"
docker run -it --rm -v "${PWD}/bin:/src/bin" "${IMG_NAME}" $@
- gsutil -m cp -nrc bin "${DEB_RELEASE_BUCKET}/${BUILD_TAG}"
printf "%s" "${BUILD_TAG}" | gsutil cp - "${DEB_RELEASE_BUCKET}/latest" | 2 | 0.125 | 0 | 2 |
bdd7242ab8ec452bfa7ef03561d5c42997144097 | .travis.yml | .travis.yml | language: node_js
node_js:
- stable
git:
submodules: false
before_install:
- export DISPLAY=:99.0
- Xvfb :99.0 -extension RANDR -ac -screen 0 1280x1024x16 > /dev/null &
before_script:
- npm run compile
| sudo: true # needed for google-chrome-stable install workaround :(
language: node_js
node_js:
- stable
# Handle git submodules yourself
git:
submodules: false
# Current broken on Travis 15/March/2016
# addons:
# apt:
# sources:
# - google-chrome
# packages:
# - google-chrome-stable
before_install:
# Chrome Stable
- wget -q -O - https://dl-ssl.google.com/linux/linux_signing_key.pub | sudo apt-key add -
- echo "deb [arch=amd64] http://dl.google.com/linux/chrome/deb/ stable main" | sudo tee -a /etc/apt/sources.list
- sudo apt-get update -qq
- sudo apt-get install -y google-chrome-stable
- export CHROME_BIN=google-chrome-stable
- export DISPLAY=:99.0
- Xvfb :99.0 -extension RANDR -ac -screen 0 1280x1024x16 > /dev/null &
- google-chrome-stable --version
before_script:
- npm run compile
| Revert "[tests] Try current version of chrome as well" | Revert "[tests] Try current version of chrome as well"
This reverts commit a62c5e4140af23453b24f1f3aa052527c6493723.
| YAML | mit | toxicFork/react-three-renderer | yaml | ## Code Before:
language: node_js
node_js:
- stable
git:
submodules: false
before_install:
- export DISPLAY=:99.0
- Xvfb :99.0 -extension RANDR -ac -screen 0 1280x1024x16 > /dev/null &
before_script:
- npm run compile
## Instruction:
Revert "[tests] Try current version of chrome as well"
This reverts commit a62c5e4140af23453b24f1f3aa052527c6493723.
## Code After:
sudo: true # needed for google-chrome-stable install workaround :(
language: node_js
node_js:
- stable
# Handle git submodules yourself
git:
submodules: false
# Current broken on Travis 15/March/2016
# addons:
# apt:
# sources:
# - google-chrome
# packages:
# - google-chrome-stable
before_install:
# Chrome Stable
- wget -q -O - https://dl-ssl.google.com/linux/linux_signing_key.pub | sudo apt-key add -
- echo "deb [arch=amd64] http://dl.google.com/linux/chrome/deb/ stable main" | sudo tee -a /etc/apt/sources.list
- sudo apt-get update -qq
- sudo apt-get install -y google-chrome-stable
- export CHROME_BIN=google-chrome-stable
- export DISPLAY=:99.0
- Xvfb :99.0 -extension RANDR -ac -screen 0 1280x1024x16 > /dev/null &
- google-chrome-stable --version
before_script:
- npm run compile
| + sudo: true # needed for google-chrome-stable install workaround :(
language: node_js
node_js:
- stable
+ # Handle git submodules yourself
git:
submodules: false
+ # Current broken on Travis 15/March/2016
+ # addons:
+ # apt:
+ # sources:
+ # - google-chrome
+ # packages:
+ # - google-chrome-stable
before_install:
+ # Chrome Stable
+ - wget -q -O - https://dl-ssl.google.com/linux/linux_signing_key.pub | sudo apt-key add -
+ - echo "deb [arch=amd64] http://dl.google.com/linux/chrome/deb/ stable main" | sudo tee -a /etc/apt/sources.list
+ - sudo apt-get update -qq
+ - sudo apt-get install -y google-chrome-stable
+ - export CHROME_BIN=google-chrome-stable
- export DISPLAY=:99.0
- Xvfb :99.0 -extension RANDR -ac -screen 0 1280x1024x16 > /dev/null &
+ - google-chrome-stable --version
before_script:
- npm run compile | 16 | 1.6 | 16 | 0 |
0aa532213cbc29e5c707b249a8700dd96b64e654 | CHANGELOG.md | CHANGELOG.md | All notable changes to this project will be documented in this file.
## Unreleased
### Added
### Changed
### Fixed
### Removed
## v0.2.3 - 2015-06-29
### Fixed
- Speedruncom autocompleter
## v0.2.2 - 2015-06-24
### Added
- Decrease visibility of world records already browsed
- Ability to browse older world records pages
### Fixed
- Speedruncom mis-autocompleting platform
- `-` character in youtube video identifiers
## v0.2.1 - 2015-06-23
### Added
- Docker configuration file
## v0.2.0 - 2015-06-10
### Added
- Paginate world records on the homepage and sort them by creation date
- Account registration with a confirmation email
- Command to promote a user as moderator
- Bootstrap theme
- Login/disconnect link in the navigation bar
- Submission refusal for moderators
- Thumbnail for runs having a Youtube video
- Configurable Piwik tracking
## v0.1.0 - 2015-05-06
### Added
- Changelog
- Page to add records submissions
- List of all world records on the homepage
- Page to validate pending submissions
- Autocompletion of the submission form by providing a speedrun.com link
| All notable changes to this project will be documented in this file.
## Unreleased
### Added
- Missing default thumbnail
### Changed
- Improve readability of the homepage
### Fixed
### Removed
## v0.2.3 - 2015-06-29
### Fixed
- Speedruncom autocompleter
## v0.2.2 - 2015-06-24
### Added
- Decrease visibility of world records already browsed
- Ability to browse older world records pages
### Fixed
- Speedruncom mis-autocompleting platform
- `-` character in youtube video identifiers
## v0.2.1 - 2015-06-23
### Added
- Docker configuration file
## v0.2.0 - 2015-06-10
### Added
- Paginate world records on the homepage and sort them by creation date
- Account registration with a confirmation email
- Command to promote a user as moderator
- Bootstrap theme
- Login/disconnect link in the navigation bar
- Submission refusal for moderators
- Thumbnail for runs having a Youtube video
- Configurable Piwik tracking
## v0.1.0 - 2015-05-06
### Added
- Changelog
- Page to add records submissions
- List of all world records on the homepage
- Page to validate pending submissions
- Autocompletion of the submission form by providing a speedrun.com link
| Update the changelog for the v0.2.4 release | Update the changelog for the v0.2.4 release
| Markdown | mit | re7/world-records,re7/world-records,re7/world-records,re7/world-records | markdown | ## Code Before:
All notable changes to this project will be documented in this file.
## Unreleased
### Added
### Changed
### Fixed
### Removed
## v0.2.3 - 2015-06-29
### Fixed
- Speedruncom autocompleter
## v0.2.2 - 2015-06-24
### Added
- Decrease visibility of world records already browsed
- Ability to browse older world records pages
### Fixed
- Speedruncom mis-autocompleting platform
- `-` character in youtube video identifiers
## v0.2.1 - 2015-06-23
### Added
- Docker configuration file
## v0.2.0 - 2015-06-10
### Added
- Paginate world records on the homepage and sort them by creation date
- Account registration with a confirmation email
- Command to promote a user as moderator
- Bootstrap theme
- Login/disconnect link in the navigation bar
- Submission refusal for moderators
- Thumbnail for runs having a Youtube video
- Configurable Piwik tracking
## v0.1.0 - 2015-05-06
### Added
- Changelog
- Page to add records submissions
- List of all world records on the homepage
- Page to validate pending submissions
- Autocompletion of the submission form by providing a speedrun.com link
## Instruction:
Update the changelog for the v0.2.4 release
## Code After:
All notable changes to this project will be documented in this file.
## Unreleased
### Added
- Missing default thumbnail
### Changed
- Improve readability of the homepage
### Fixed
### Removed
## v0.2.3 - 2015-06-29
### Fixed
- Speedruncom autocompleter
## v0.2.2 - 2015-06-24
### Added
- Decrease visibility of world records already browsed
- Ability to browse older world records pages
### Fixed
- Speedruncom mis-autocompleting platform
- `-` character in youtube video identifiers
## v0.2.1 - 2015-06-23
### Added
- Docker configuration file
## v0.2.0 - 2015-06-10
### Added
- Paginate world records on the homepage and sort them by creation date
- Account registration with a confirmation email
- Command to promote a user as moderator
- Bootstrap theme
- Login/disconnect link in the navigation bar
- Submission refusal for moderators
- Thumbnail for runs having a Youtube video
- Configurable Piwik tracking
## v0.1.0 - 2015-05-06
### Added
- Changelog
- Page to add records submissions
- List of all world records on the homepage
- Page to validate pending submissions
- Autocompletion of the submission form by providing a speedrun.com link
| All notable changes to this project will be documented in this file.
## Unreleased
### Added
+ - Missing default thumbnail
### Changed
+ - Improve readability of the homepage
### Fixed
### Removed
## v0.2.3 - 2015-06-29
### Fixed
- Speedruncom autocompleter
## v0.2.2 - 2015-06-24
### Added
- Decrease visibility of world records already browsed
- Ability to browse older world records pages
### Fixed
- Speedruncom mis-autocompleting platform
- `-` character in youtube video identifiers
## v0.2.1 - 2015-06-23
### Added
- Docker configuration file
## v0.2.0 - 2015-06-10
### Added
- Paginate world records on the homepage and sort them by creation date
- Account registration with a confirmation email
- Command to promote a user as moderator
- Bootstrap theme
- Login/disconnect link in the navigation bar
- Submission refusal for moderators
- Thumbnail for runs having a Youtube video
- Configurable Piwik tracking
## v0.1.0 - 2015-05-06
### Added
- Changelog
- Page to add records submissions
- List of all world records on the homepage
- Page to validate pending submissions
- Autocompletion of the submission form by providing a speedrun.com link | 2 | 0.043478 | 2 | 0 |
e7dc4af3a45ae68bb38a6f1cab175a18d49a922a | .lgtm.yml | .lgtm.yml | extraction:
java:
index:
gradle:
version: 5.2.1
path_classifiers:
docs:
- "LICENSE*"
- "*.md"
- "*.html"
ci:
- "appveyor.yml"
- ".codecov.yml"
- ".travis.yml"
- ".circleci"
| extraction:
java:
index:
gradle:
version: 5.2.1
path_classifiers:
docs:
- "LICENSE*"
- "*.md"
- "*.html"
ci:
- "appveyor.yml"
- ".codecov.yml"
- ".travis.yml"
- ".circleci"
- ".dependabot"
| Add dependabot as a ci tag | Add dependabot as a ci tag
| YAML | apache-2.0 | adaptris/interlok | yaml | ## Code Before:
extraction:
java:
index:
gradle:
version: 5.2.1
path_classifiers:
docs:
- "LICENSE*"
- "*.md"
- "*.html"
ci:
- "appveyor.yml"
- ".codecov.yml"
- ".travis.yml"
- ".circleci"
## Instruction:
Add dependabot as a ci tag
## Code After:
extraction:
java:
index:
gradle:
version: 5.2.1
path_classifiers:
docs:
- "LICENSE*"
- "*.md"
- "*.html"
ci:
- "appveyor.yml"
- ".codecov.yml"
- ".travis.yml"
- ".circleci"
- ".dependabot"
| extraction:
java:
index:
gradle:
version: 5.2.1
path_classifiers:
docs:
- "LICENSE*"
- "*.md"
- "*.html"
ci:
- "appveyor.yml"
- ".codecov.yml"
- ".travis.yml"
- ".circleci"
+ - ".dependabot" | 1 | 0.0625 | 1 | 0 |
ba830ed1ae892b1be58b55583c90260ffe80cbf2 | README.md | README.md | Dimagi Utils [](https://travis-ci.org/dimagi/dimagi-utils)
============
Misc utility code. Caution! This _is_ shared between a few projects and cannot
be changed arbitrarily.
Note that some utilities have soft requirements - in other words they are not
automatically pulled in by `pip install dimagi-utils` but are required
for various pieces of functionality.
| Dimagi Utils [](https://travis-ci.org/dimagi/dimagi-utils)
============
Misc utility code. Caution! This _is_ shared between a few projects and cannot
be changed arbitrarily.
Note that some utilities have soft requirements - in other words they are not
automatically pulled in by `pip install dimagi-utils` but are required
for various pieces of functionality.
| Fix image to be the right one | Fix image to be the right one
| Markdown | bsd-3-clause | dimagi/commcare-hq,qedsoftware/commcare-hq,qedsoftware/commcare-hq,dimagi/commcare-hq,qedsoftware/commcare-hq,qedsoftware/commcare-hq,dimagi/commcare-hq,puttarajubr/commcare-hq,puttarajubr/commcare-hq,puttarajubr/commcare-hq,dimagi/commcare-hq,puttarajubr/commcare-hq,qedsoftware/commcare-hq,dimagi/commcare-hq | markdown | ## Code Before:
Dimagi Utils [](https://travis-ci.org/dimagi/dimagi-utils)
============
Misc utility code. Caution! This _is_ shared between a few projects and cannot
be changed arbitrarily.
Note that some utilities have soft requirements - in other words they are not
automatically pulled in by `pip install dimagi-utils` but are required
for various pieces of functionality.
## Instruction:
Fix image to be the right one
## Code After:
Dimagi Utils [](https://travis-ci.org/dimagi/dimagi-utils)
============
Misc utility code. Caution! This _is_ shared between a few projects and cannot
be changed arbitrarily.
Note that some utilities have soft requirements - in other words they are not
automatically pulled in by `pip install dimagi-utils` but are required
for various pieces of functionality.
| - Dimagi Utils [](https://travis-ci.org/dimagi/dimagi-utils)
? ^^ -- ^^ ^^
+ Dimagi Utils [](https://travis-ci.org/dimagi/dimagi-utils)
? ^^ ^^ ^^^^^
============
Misc utility code. Caution! This _is_ shared between a few projects and cannot
be changed arbitrarily.
Note that some utilities have soft requirements - in other words they are not
automatically pulled in by `pip install dimagi-utils` but are required
for various pieces of functionality. | 2 | 0.222222 | 1 | 1 |
ea7b245b0630a0cff7972cc3f88911fb947fc574 | paths/bin.zsh | paths/bin.zsh | PATH=~/bin:~/.local/bin:$PATH
| if [ -d ~/bin ]; then
PATH=$PATH:~/bin
fi
if [ -d ~/.local/bin ]; then
PATH=$PATH:~/.local/bin
fi
| Add paths to PATH only if they exist | fix: Add paths to PATH only if they exist
| Shell | mit | scheleaap/zsh-config | shell | ## Code Before:
PATH=~/bin:~/.local/bin:$PATH
## Instruction:
fix: Add paths to PATH only if they exist
## Code After:
if [ -d ~/bin ]; then
PATH=$PATH:~/bin
fi
if [ -d ~/.local/bin ]; then
PATH=$PATH:~/.local/bin
fi
| - PATH=~/bin:~/.local/bin:$PATH
+ if [ -d ~/bin ]; then
+ PATH=$PATH:~/bin
+ fi
+ if [ -d ~/.local/bin ]; then
+ PATH=$PATH:~/.local/bin
+ fi | 7 | 7 | 6 | 1 |
528813b437d2121b2c52f51884746e4cbeeeacf8 | src/main/java/leetcode/Problem714.java | src/main/java/leetcode/Problem714.java | package leetcode;
/**
* https://leetcode.com/problems/best-time-to-buy-and-sell-stock-with-transaction-fee/
*/
public class Problem714 {
public int maxProfit(int[] prices, int fee) {
// TODO
return 0;
}
public static void main(String[] args) {
Problem714 prob = new Problem714();
System.out.println(prob.maxProfit(new int[]{1, 3, 2, 8, 4, 9}, 2)); // 8
System.out.println(prob.maxProfit(new int[]{1, 3, 2, 6, 8, 4, 9}, 2)); // 8
System.out.println(prob.maxProfit(new int[]{1, 4, 2, 1, 4, 8, 4, 9}, 2)); // 9
}
}
| package leetcode;
/**
* https://leetcode.com/problems/best-time-to-buy-and-sell-stock-with-transaction-fee/
*/
public class Problem714 {
public int maxProfit(int[] prices, int fee) {
int cash = 0;
int hold = -prices[0];
for (int i = 1; i < prices.length; i++) {
cash = Math.max(cash, hold + prices[i] - fee);
hold = Math.max(hold, cash - prices[i]);
}
return cash;
}
}
| Solve problem 714 (not really) | Solve problem 714 (not really)
| Java | mit | fredyw/leetcode,fredyw/leetcode,fredyw/leetcode,fredyw/leetcode | java | ## Code Before:
package leetcode;
/**
* https://leetcode.com/problems/best-time-to-buy-and-sell-stock-with-transaction-fee/
*/
public class Problem714 {
public int maxProfit(int[] prices, int fee) {
// TODO
return 0;
}
public static void main(String[] args) {
Problem714 prob = new Problem714();
System.out.println(prob.maxProfit(new int[]{1, 3, 2, 8, 4, 9}, 2)); // 8
System.out.println(prob.maxProfit(new int[]{1, 3, 2, 6, 8, 4, 9}, 2)); // 8
System.out.println(prob.maxProfit(new int[]{1, 4, 2, 1, 4, 8, 4, 9}, 2)); // 9
}
}
## Instruction:
Solve problem 714 (not really)
## Code After:
package leetcode;
/**
* https://leetcode.com/problems/best-time-to-buy-and-sell-stock-with-transaction-fee/
*/
public class Problem714 {
public int maxProfit(int[] prices, int fee) {
int cash = 0;
int hold = -prices[0];
for (int i = 1; i < prices.length; i++) {
cash = Math.max(cash, hold + prices[i] - fee);
hold = Math.max(hold, cash - prices[i]);
}
return cash;
}
}
| package leetcode;
/**
* https://leetcode.com/problems/best-time-to-buy-and-sell-stock-with-transaction-fee/
*/
public class Problem714 {
public int maxProfit(int[] prices, int fee) {
- // TODO
+ int cash = 0;
+ int hold = -prices[0];
+ for (int i = 1; i < prices.length; i++) {
+ cash = Math.max(cash, hold + prices[i] - fee);
+ hold = Math.max(hold, cash - prices[i]);
+ }
- return 0;
? ^
+ return cash;
? ^^^^
- }
-
- public static void main(String[] args) {
- Problem714 prob = new Problem714();
- System.out.println(prob.maxProfit(new int[]{1, 3, 2, 8, 4, 9}, 2)); // 8
- System.out.println(prob.maxProfit(new int[]{1, 3, 2, 6, 8, 4, 9}, 2)); // 8
- System.out.println(prob.maxProfit(new int[]{1, 4, 2, 1, 4, 8, 4, 9}, 2)); // 9
}
} | 16 | 0.888889 | 7 | 9 |
837cce799c953d54dbc2e358ccef0213db2715ad | Marshal.podspec | Marshal.podspec | Pod::Spec.new do |s|
s.name = "Marshal"
s.version = "1.2.1"
s.summary = "Marshal is a simple, lightweight framework for safely extracting values from [String: AnyObject]"
s.description = <<-DESC
In Swift, we all deal with JSON, plists, and various forms of [String: Any]. Marshal believes you don't need a Ph.D. in monads or magic mirrors to deal with these in an expressive and type safe way. Marshal will help you write declarative, performant, error handled code using the power of Protocol Oriented Programming™.
DESC
s.homepage = "https://github.com/utahiosmac/Marshal"
s.license = "MIT"
s.author = "Utah iOS & Mac"
s.osx.deployment_target = "10.9"
s.ios.deployment_target = "8.0"
s.tvos.deployment_target = "9.0"
s.watchos.deployment_target = "2.0"
s.source = { :git => "https://github.com/utahiosmac/Marshal.git",
:tag => "v" + s.version.to_s }
s.source_files = "Marshal/**/*.swift", "Sources/**/*.swift"
s.requires_arc = true
s.module_name = "Marshal"
end
| Pod::Spec.new do |s|
s.name = "Marshal"
s.version = "1.2.3"
s.summary = "Marshal is a simple, lightweight framework for safely extracting values from [String: AnyObject]"
s.description = <<-DESC
In Swift, we all deal with JSON, plists, and various forms of [String: Any]. Marshal believes you don't need a Ph.D. in monads or magic mirrors to deal with these in an expressive and type safe way. Marshal will help you write declarative, performant, error handled code using the power of Protocol Oriented Programming™.
DESC
s.homepage = "https://github.com/utahiosmac/Marshal"
s.license = "MIT"
s.author = "Utah iOS & Mac"
s.osx.deployment_target = "10.9"
s.ios.deployment_target = "8.0"
s.tvos.deployment_target = "9.0"
s.watchos.deployment_target = "2.0"
s.source = { :git => "https://github.com/utahiosmac/Marshal.git",
:tag => "#{s.version}" }
s.source_files = "Marshal/**/*.swift", "Sources/**/*.swift"
s.requires_arc = true
s.module_name = "Marshal"
end
| Update podspec to newest release | Update podspec to newest release
Updates podspec to newest release, and also fixes which tag it looks for in source. | Ruby | mit | utahiosmac/Marshal,utahiosmac/Marshal | ruby | ## Code Before:
Pod::Spec.new do |s|
s.name = "Marshal"
s.version = "1.2.1"
s.summary = "Marshal is a simple, lightweight framework for safely extracting values from [String: AnyObject]"
s.description = <<-DESC
In Swift, we all deal with JSON, plists, and various forms of [String: Any]. Marshal believes you don't need a Ph.D. in monads or magic mirrors to deal with these in an expressive and type safe way. Marshal will help you write declarative, performant, error handled code using the power of Protocol Oriented Programming™.
DESC
s.homepage = "https://github.com/utahiosmac/Marshal"
s.license = "MIT"
s.author = "Utah iOS & Mac"
s.osx.deployment_target = "10.9"
s.ios.deployment_target = "8.0"
s.tvos.deployment_target = "9.0"
s.watchos.deployment_target = "2.0"
s.source = { :git => "https://github.com/utahiosmac/Marshal.git",
:tag => "v" + s.version.to_s }
s.source_files = "Marshal/**/*.swift", "Sources/**/*.swift"
s.requires_arc = true
s.module_name = "Marshal"
end
## Instruction:
Update podspec to newest release
Updates podspec to newest release, and also fixes which tag it looks for in source.
## Code After:
Pod::Spec.new do |s|
s.name = "Marshal"
s.version = "1.2.3"
s.summary = "Marshal is a simple, lightweight framework for safely extracting values from [String: AnyObject]"
s.description = <<-DESC
In Swift, we all deal with JSON, plists, and various forms of [String: Any]. Marshal believes you don't need a Ph.D. in monads or magic mirrors to deal with these in an expressive and type safe way. Marshal will help you write declarative, performant, error handled code using the power of Protocol Oriented Programming™.
DESC
s.homepage = "https://github.com/utahiosmac/Marshal"
s.license = "MIT"
s.author = "Utah iOS & Mac"
s.osx.deployment_target = "10.9"
s.ios.deployment_target = "8.0"
s.tvos.deployment_target = "9.0"
s.watchos.deployment_target = "2.0"
s.source = { :git => "https://github.com/utahiosmac/Marshal.git",
:tag => "#{s.version}" }
s.source_files = "Marshal/**/*.swift", "Sources/**/*.swift"
s.requires_arc = true
s.module_name = "Marshal"
end
| Pod::Spec.new do |s|
s.name = "Marshal"
- s.version = "1.2.1"
? ^
+ s.version = "1.2.3"
? ^
s.summary = "Marshal is a simple, lightweight framework for safely extracting values from [String: AnyObject]"
s.description = <<-DESC
In Swift, we all deal with JSON, plists, and various forms of [String: Any]. Marshal believes you don't need a Ph.D. in monads or magic mirrors to deal with these in an expressive and type safe way. Marshal will help you write declarative, performant, error handled code using the power of Protocol Oriented Programming™.
DESC
s.homepage = "https://github.com/utahiosmac/Marshal"
s.license = "MIT"
s.author = "Utah iOS & Mac"
s.osx.deployment_target = "10.9"
s.ios.deployment_target = "8.0"
s.tvos.deployment_target = "9.0"
s.watchos.deployment_target = "2.0"
s.source = { :git => "https://github.com/utahiosmac/Marshal.git",
- :tag => "v" + s.version.to_s }
? ^^^^^ ^^^^^
+ :tag => "#{s.version}" }
? ^^ ^^
s.source_files = "Marshal/**/*.swift", "Sources/**/*.swift"
s.requires_arc = true
s.module_name = "Marshal"
end | 4 | 0.2 | 2 | 2 |
3ddeb68bdae6142d811cef4b074e06144eb4aa9f | test/functional/data/construct-javascript-function.js | test/functional/data/construct-javascript-function.js | 'use strict';
var assert = require('assert');
function testHandler(actual) {
var expected = testHandler.getExpectedData();
assert.strictEqual(actual.length, expected.length);
assert.strictEqual(
actual[0](),
expected[0]());
assert.strictEqual(
actual[1](10, 20),
expected[1](10, 20));
assert.deepEqual(
actual[2]('book'),
expected[2]('book'));
}
testHandler.getExpectedData = function getExpectedData() {
return [
function () {
return 42;
},
function (x, y) {
return x + y;
},
function (foo) {
var result = 'There is my ' + foo + ' at the table.';
return {
first: 42,
second: 'sum',
third: result
};
}
];
};
module.exports = testHandler;
| 'use strict';
var assert = require('assert');
function testHandler(actual) {
var expected = testHandler.expected;
assert.strictEqual(actual.length, expected.length);
assert.strictEqual(
actual[0](),
expected[0]());
assert.strictEqual(
actual[1](10, 20),
expected[1](10, 20));
assert.deepEqual(
actual[2]('book'),
expected[2]('book'));
}
testHandler.expected = [
function () {
return 42;
},
function (x, y) {
return x + y;
},
function (foo) {
var result = 'There is my ' + foo + ' at the table.';
return {
first: 42,
second: 'sum',
third: result
};
}
];
module.exports = testHandler;
| Use plain `expected` for self-testing samples | Use plain `expected` for self-testing samples
| JavaScript | mit | isaacs/js-yaml,nodeca/js-yaml,bjlxj2008/js-yaml,jonnor/js-yaml,cesarmarinhorj/js-yaml,prose/js-yaml,isaacs/js-yaml,djchie/js-yaml,crissdev/js-yaml,SmartBear/js-yaml,vogelsgesang/js-yaml,isaacs/js-yaml,nodeca/js-yaml,jonnor/js-yaml,cesarmarinhorj/js-yaml,vogelsgesang/js-yaml,rjmunro/js-yaml,minj/js-yaml,denji/js-yaml,pombredanne/js-yaml,crissdev/js-yaml,bjlxj2008/js-yaml,pombredanne/js-yaml,doowb/js-yaml,jonnor/js-yaml,bjlxj2008/js-yaml,doowb/js-yaml,deltreey/js-yaml,pombredanne/js-yaml,rjmunro/js-yaml,minj/js-yaml,denji/js-yaml,djchie/js-yaml,cesarmarinhorj/js-yaml,joshball/js-yaml,vogelsgesang/js-yaml,doowb/js-yaml,SmartBear/js-yaml,denji/js-yaml,rjmunro/js-yaml,deltreey/js-yaml,joshball/js-yaml,SmartBear/js-yaml,djchie/js-yaml,deltreey/js-yaml,nodeca/js-yaml,crissdev/js-yaml,minj/js-yaml,joshball/js-yaml | javascript | ## Code Before:
'use strict';
var assert = require('assert');
function testHandler(actual) {
var expected = testHandler.getExpectedData();
assert.strictEqual(actual.length, expected.length);
assert.strictEqual(
actual[0](),
expected[0]());
assert.strictEqual(
actual[1](10, 20),
expected[1](10, 20));
assert.deepEqual(
actual[2]('book'),
expected[2]('book'));
}
testHandler.getExpectedData = function getExpectedData() {
return [
function () {
return 42;
},
function (x, y) {
return x + y;
},
function (foo) {
var result = 'There is my ' + foo + ' at the table.';
return {
first: 42,
second: 'sum',
third: result
};
}
];
};
module.exports = testHandler;
## Instruction:
Use plain `expected` for self-testing samples
## Code After:
'use strict';
var assert = require('assert');
function testHandler(actual) {
var expected = testHandler.expected;
assert.strictEqual(actual.length, expected.length);
assert.strictEqual(
actual[0](),
expected[0]());
assert.strictEqual(
actual[1](10, 20),
expected[1](10, 20));
assert.deepEqual(
actual[2]('book'),
expected[2]('book'));
}
testHandler.expected = [
function () {
return 42;
},
function (x, y) {
return x + y;
},
function (foo) {
var result = 'There is my ' + foo + ' at the table.';
return {
first: 42,
second: 'sum',
third: result
};
}
];
module.exports = testHandler;
| 'use strict';
var assert = require('assert');
function testHandler(actual) {
- var expected = testHandler.getExpectedData();
? - -- ------
+ var expected = testHandler.expected;
assert.strictEqual(actual.length, expected.length);
assert.strictEqual(
actual[0](),
expected[0]());
assert.strictEqual(
actual[1](10, 20),
expected[1](10, 20));
assert.deepEqual(
actual[2]('book'),
expected[2]('book'));
}
+ testHandler.expected = [
- testHandler.getExpectedData = function getExpectedData() {
- return [
- function () {
? --
+ function () {
- return 42;
? --
+ return 42;
- },
? --
+ },
- function (x, y) {
? --
+ function (x, y) {
- return x + y;
? --
+ return x + y;
- },
? --
+ },
- function (foo) {
? --
+ function (foo) {
- var result = 'There is my ' + foo + ' at the table.';
? --
+ var result = 'There is my ' + foo + ' at the table.';
- return {
? --
+ return {
- first: 42,
? --
+ first: 42,
- second: 'sum',
? --
+ second: 'sum',
- third: result
? --
+ third: result
- };
- }
+ };
? +
- ];
- };
+ }
+ ];
module.exports = testHandler; | 36 | 0.782609 | 17 | 19 |
d27e672ecb775be8beef2c69f6d04cf64a67a5b1 | requirements/production.txt | requirements/production.txt | -r base.txt
# WSGI Handler
# ------------------------------------------------
gevent==1.1.1
gunicorn==19.6.0
# Static and Media Storage
# ------------------------------------------------
boto==2.41.0
django-storages-redux==1.3.2
# Email backends for Mailgun, Postmark, SendGrid and more
# -------------------------------------------------------
django-anymail==0.4.2
# Raven is the Sentry client
# --------------------------
raven==5.22.0
| -r base.txt
# Security measures
# ------------------------------------------------
django-secure==1.0.1
# WSGI Handler
# ------------------------------------------------
gevent==1.1.1
gunicorn==19.6.0
# Static and Media Storage
# ------------------------------------------------
boto==2.41.0
django-storages-redux==1.3.2
# Email backends for Mailgun, Postmark, SendGrid and more
# -------------------------------------------------------
django-anymail==0.4.2
# Raven is the Sentry client
# --------------------------
raven==5.22.0
| Add django secure package to deployment | Add django secure package to deployment
For redirecting to HTTPS | Text | mit | MazeFX/cookiecutter_website_project,MazeFX/cookiecutter_website_project,MazeFX/cookiecutter_website_project,MazeFX/cookiecutter_website_project | text | ## Code Before:
-r base.txt
# WSGI Handler
# ------------------------------------------------
gevent==1.1.1
gunicorn==19.6.0
# Static and Media Storage
# ------------------------------------------------
boto==2.41.0
django-storages-redux==1.3.2
# Email backends for Mailgun, Postmark, SendGrid and more
# -------------------------------------------------------
django-anymail==0.4.2
# Raven is the Sentry client
# --------------------------
raven==5.22.0
## Instruction:
Add django secure package to deployment
For redirecting to HTTPS
## Code After:
-r base.txt
# Security measures
# ------------------------------------------------
django-secure==1.0.1
# WSGI Handler
# ------------------------------------------------
gevent==1.1.1
gunicorn==19.6.0
# Static and Media Storage
# ------------------------------------------------
boto==2.41.0
django-storages-redux==1.3.2
# Email backends for Mailgun, Postmark, SendGrid and more
# -------------------------------------------------------
django-anymail==0.4.2
# Raven is the Sentry client
# --------------------------
raven==5.22.0
| -r base.txt
-
+ # Security measures
+ # ------------------------------------------------
+ django-secure==1.0.1
# WSGI Handler
# ------------------------------------------------
gevent==1.1.1
gunicorn==19.6.0
# Static and Media Storage
# ------------------------------------------------
boto==2.41.0
django-storages-redux==1.3.2
# Email backends for Mailgun, Postmark, SendGrid and more
# -------------------------------------------------------
django-anymail==0.4.2
# Raven is the Sentry client
# --------------------------
raven==5.22.0
| 4 | 0.166667 | 3 | 1 |
25f3e26af42a4c0366c740ffd5212c9d041a5433 | config/application.rb | config/application.rb | require File.expand_path('../boot', __FILE__)
require 'rails/all'
# Require the gems listed in Gemfile, including any gems
# you've limited to :test, :development, or :production.
Bundler.require(*Rails.groups)
module Gemcutter
class Application < Rails::Application
def config_for(name, env = Rails.env)
YAML.load_file(Rails.root.join("config/#{name}.yml"))[env]
end
config.rubygems = Application.config_for :rubygems
config.time_zone = "UTC"
config.encoding = "utf-8"
config.middleware.use "Redirector"
config.active_record.include_root_in_json = false
config.active_record.raise_in_transactional_callbacks = true
config.after_initialize do
RubygemFs.s3! ENV['S3_PROXY'] if ENV['S3_PROXY']
end
config.plugins = [:dynamic_form]
config.autoload_paths << Rails.root.join('lib')
end
def self.config
Rails.application.config.rubygems
end
HOST = config['host']
end
| require File.expand_path('../boot', __FILE__)
require 'rails/all'
# Require the gems listed in Gemfile, including any gems
# you've limited to :test, :development, or :production.
Bundler.require(*Rails.groups)
module Gemcutter
class Application < Rails::Application
def config_for(name, env = Rails.env)
YAML.load_file(Rails.root.join("config/#{name}.yml"))[env]
end
config.rubygems = Application.config_for :rubygems
config.time_zone = "UTC"
config.encoding = "utf-8"
config.middleware.use "Redirector" unless Rails.env.development?
config.active_record.include_root_in_json = false
config.active_record.raise_in_transactional_callbacks = true
config.after_initialize do
RubygemFs.s3! ENV['S3_PROXY'] if ENV['S3_PROXY']
end
config.plugins = [:dynamic_form]
config.autoload_paths << Rails.root.join('lib')
end
def self.config
Rails.application.config.rubygems
end
HOST = config['host']
end
| Make pow work again by not redirecting in dev | Make pow work again by not redirecting in dev
| Ruby | mit | algolia/rubygems.org,Exeia/rubygems.org,Elffers/rubygems.org,polamjag/rubygems.org,huacnlee/rubygems.org,kbrock/rubygems.org,Elffers/rubygems.org,kbrock/rubygems.org,huacnlee/rubygems.org,olivierlacan/rubygems.org,fotanus/rubygems.org,krainboltgreene/rubygems.org,krainboltgreene/rubygems.org,knappe/rubygems.org,wallin/rubygems.org,Elffers/rubygems.org,kbrock/rubygems.org,JuanitoFatas/rubygems.org,knappe/rubygems.org,krainboltgreene/rubygems.org,huacnlee/rubygems.org,jamelablack/rubygems.org,hrs113355/rubygems.org,spk/rubygems.org,iSC-Labs/rubygems.org,iSC-Labs/rubygems.org,rubygems/rubygems.org,spk/rubygems.org,knappe/rubygems.org,wallin/rubygems.org,hrs113355/rubygems.org,rubygems/rubygems.org,iSC-Labs/rubygems.org,olivierlacan/rubygems.org,polamjag/rubygems.org,rubygems/rubygems.org,krainboltgreene/rubygems.org,fotanus/rubygems.org,fotanus/rubygems.org,Exeia/rubygems.org,olivierlacan/rubygems.org,kbrock/rubygems.org,sonalkr132/rubygems.org,JuanitoFatas/rubygems.org,olivierlacan/rubygems.org,huacnlee/rubygems.org,iSC-Labs/rubygems.org,polamjag/rubygems.org,Exeia/rubygems.org,wallin/rubygems.org,jamelablack/rubygems.org,JuanitoFatas/rubygems.org,sonalkr132/rubygems.org,jamelablack/rubygems.org,knappe/rubygems.org,rubygems/rubygems.org,algolia/rubygems.org,farukaydin/rubygems.org,spk/rubygems.org,farukaydin/rubygems.org,farukaydin/rubygems.org,polamjag/rubygems.org,spk/rubygems.org,algolia/rubygems.org,fotanus/rubygems.org,sonalkr132/rubygems.org,jamelablack/rubygems.org,sonalkr132/rubygems.org,hrs113355/rubygems.org,Exeia/rubygems.org,Elffers/rubygems.org,farukaydin/rubygems.org,hrs113355/rubygems.org,algolia/rubygems.org,JuanitoFatas/rubygems.org,wallin/rubygems.org | ruby | ## Code Before:
require File.expand_path('../boot', __FILE__)
require 'rails/all'
# Require the gems listed in Gemfile, including any gems
# you've limited to :test, :development, or :production.
Bundler.require(*Rails.groups)
module Gemcutter
class Application < Rails::Application
def config_for(name, env = Rails.env)
YAML.load_file(Rails.root.join("config/#{name}.yml"))[env]
end
config.rubygems = Application.config_for :rubygems
config.time_zone = "UTC"
config.encoding = "utf-8"
config.middleware.use "Redirector"
config.active_record.include_root_in_json = false
config.active_record.raise_in_transactional_callbacks = true
config.after_initialize do
RubygemFs.s3! ENV['S3_PROXY'] if ENV['S3_PROXY']
end
config.plugins = [:dynamic_form]
config.autoload_paths << Rails.root.join('lib')
end
def self.config
Rails.application.config.rubygems
end
HOST = config['host']
end
## Instruction:
Make pow work again by not redirecting in dev
## Code After:
require File.expand_path('../boot', __FILE__)
require 'rails/all'
# Require the gems listed in Gemfile, including any gems
# you've limited to :test, :development, or :production.
Bundler.require(*Rails.groups)
module Gemcutter
class Application < Rails::Application
def config_for(name, env = Rails.env)
YAML.load_file(Rails.root.join("config/#{name}.yml"))[env]
end
config.rubygems = Application.config_for :rubygems
config.time_zone = "UTC"
config.encoding = "utf-8"
config.middleware.use "Redirector" unless Rails.env.development?
config.active_record.include_root_in_json = false
config.active_record.raise_in_transactional_callbacks = true
config.after_initialize do
RubygemFs.s3! ENV['S3_PROXY'] if ENV['S3_PROXY']
end
config.plugins = [:dynamic_form]
config.autoload_paths << Rails.root.join('lib')
end
def self.config
Rails.application.config.rubygems
end
HOST = config['host']
end
| require File.expand_path('../boot', __FILE__)
require 'rails/all'
# Require the gems listed in Gemfile, including any gems
# you've limited to :test, :development, or :production.
Bundler.require(*Rails.groups)
module Gemcutter
class Application < Rails::Application
def config_for(name, env = Rails.env)
YAML.load_file(Rails.root.join("config/#{name}.yml"))[env]
end
config.rubygems = Application.config_for :rubygems
config.time_zone = "UTC"
config.encoding = "utf-8"
- config.middleware.use "Redirector"
+ config.middleware.use "Redirector" unless Rails.env.development?
config.active_record.include_root_in_json = false
config.active_record.raise_in_transactional_callbacks = true
config.after_initialize do
RubygemFs.s3! ENV['S3_PROXY'] if ENV['S3_PROXY']
end
config.plugins = [:dynamic_form]
config.autoload_paths << Rails.root.join('lib')
end
def self.config
Rails.application.config.rubygems
end
HOST = config['host']
end | 2 | 0.052632 | 1 | 1 |
567f74f06e7969dc27ad3d7050a1910bdbf82d11 | readme.txt | readme.txt | === Better GitHub Widget ===
Contributors: fracek, gavD, jowilki
Tags: github, project
Author URI: http://francesco-cek.com
Author: Francesco Ceccon
Requires at least: 3.0
Tested up to: 3.3.1
Stable tag: 0.5.3
Better GitHub Widget allows you to display your GitHub projects on your site.
== Description ==
= A beautiful widget for your site =
Better GitHub Widget is the best looking widget to display your
GitHub projects, it simply blends with your theme.
= Client side magic =
All of the magic of this widget happens on the user's browser, so it does
not stress your server.
== Installation ==
1. Upload the contents of the zip (including the folder) to the
'/wp-content/plugins/' directory
2. Activate the plugin through the 'Plugins' menu in WordPress
3. Under Appearance > Widgets you will see the "Better Github Widget". Drag it to your sidebar and place it where you want it to appear.
4. Add your Github username and configure the default settings (the number of projects to display and whether to show forked repositories)
== Screenshots ==
1. The widget in the sidebar
2. The widget options
== Changelog ==
= 0.5.3 =
* Add option to show forked repos
= 0.5.2 =
* Add option to modify the title of the widget
= 0.5.1 =
* Add alt attribute to the octocat image
* Fix the a tag
= 0.5 =
* First release
| === Better GitHub Widget ===
Contributors: fracek, gavD, jowilki, cfoellmann
Tags: github, project
Author URI: http://francesco-cek.com
Author: Francesco Ceccon
Requires at least: 3.0
Tested up to: 3.3.1
Stable tag: 0.5.3
Better GitHub Widget allows you to display your GitHub projects on your site.
== Description ==
= A beautiful widget for your site =
Better GitHub Widget is the best looking widget to display your
GitHub projects, it simply blends with your theme.
= Client side magic =
All of the magic of this widget happens on the user's browser, so it does
not stress your server.
== Installation ==
1. Upload the contents of the zip (including the folder) to the
'/wp-content/plugins/' directory
2. Activate the plugin through the 'Plugins' menu in WordPress
3. Under Appearance > Widgets you will see the "Better Github Widget". Drag it to your sidebar and place it where you want it to appear.
4. Add your Github username and configure the default settings (the number of projects to display and whether to show forked repositories)
== Screenshots ==
1. The widget in the sidebar
2. The widget options
== Changelog ==
= 0.5.3 =
* Add option to show forked repos
= 0.5.2 =
* Add option to modify the title of the widget
= 0.5.1 =
* Add alt attribute to the octocat image
* Fix the a tag
= 0.5 =
* First release
| Add Christian Foellmann (cfoellmann) to contributors | Add Christian Foellmann (cfoellmann) to contributors
| Text | bsd-2-clause | fracek/better-github-widget,fracek/better-github-widget | text | ## Code Before:
=== Better GitHub Widget ===
Contributors: fracek, gavD, jowilki
Tags: github, project
Author URI: http://francesco-cek.com
Author: Francesco Ceccon
Requires at least: 3.0
Tested up to: 3.3.1
Stable tag: 0.5.3
Better GitHub Widget allows you to display your GitHub projects on your site.
== Description ==
= A beautiful widget for your site =
Better GitHub Widget is the best looking widget to display your
GitHub projects, it simply blends with your theme.
= Client side magic =
All of the magic of this widget happens on the user's browser, so it does
not stress your server.
== Installation ==
1. Upload the contents of the zip (including the folder) to the
'/wp-content/plugins/' directory
2. Activate the plugin through the 'Plugins' menu in WordPress
3. Under Appearance > Widgets you will see the "Better Github Widget". Drag it to your sidebar and place it where you want it to appear.
4. Add your Github username and configure the default settings (the number of projects to display and whether to show forked repositories)
== Screenshots ==
1. The widget in the sidebar
2. The widget options
== Changelog ==
= 0.5.3 =
* Add option to show forked repos
= 0.5.2 =
* Add option to modify the title of the widget
= 0.5.1 =
* Add alt attribute to the octocat image
* Fix the a tag
= 0.5 =
* First release
## Instruction:
Add Christian Foellmann (cfoellmann) to contributors
## Code After:
=== Better GitHub Widget ===
Contributors: fracek, gavD, jowilki, cfoellmann
Tags: github, project
Author URI: http://francesco-cek.com
Author: Francesco Ceccon
Requires at least: 3.0
Tested up to: 3.3.1
Stable tag: 0.5.3
Better GitHub Widget allows you to display your GitHub projects on your site.
== Description ==
= A beautiful widget for your site =
Better GitHub Widget is the best looking widget to display your
GitHub projects, it simply blends with your theme.
= Client side magic =
All of the magic of this widget happens on the user's browser, so it does
not stress your server.
== Installation ==
1. Upload the contents of the zip (including the folder) to the
'/wp-content/plugins/' directory
2. Activate the plugin through the 'Plugins' menu in WordPress
3. Under Appearance > Widgets you will see the "Better Github Widget". Drag it to your sidebar and place it where you want it to appear.
4. Add your Github username and configure the default settings (the number of projects to display and whether to show forked repositories)
== Screenshots ==
1. The widget in the sidebar
2. The widget options
== Changelog ==
= 0.5.3 =
* Add option to show forked repos
= 0.5.2 =
* Add option to modify the title of the widget
= 0.5.1 =
* Add alt attribute to the octocat image
* Fix the a tag
= 0.5 =
* First release
| === Better GitHub Widget ===
- Contributors: fracek, gavD, jowilki
+ Contributors: fracek, gavD, jowilki, cfoellmann
? ++++++++++++
Tags: github, project
Author URI: http://francesco-cek.com
Author: Francesco Ceccon
Requires at least: 3.0
Tested up to: 3.3.1
Stable tag: 0.5.3
Better GitHub Widget allows you to display your GitHub projects on your site.
== Description ==
= A beautiful widget for your site =
Better GitHub Widget is the best looking widget to display your
GitHub projects, it simply blends with your theme.
= Client side magic =
All of the magic of this widget happens on the user's browser, so it does
not stress your server.
== Installation ==
1. Upload the contents of the zip (including the folder) to the
'/wp-content/plugins/' directory
2. Activate the plugin through the 'Plugins' menu in WordPress
3. Under Appearance > Widgets you will see the "Better Github Widget". Drag it to your sidebar and place it where you want it to appear.
4. Add your Github username and configure the default settings (the number of projects to display and whether to show forked repositories)
== Screenshots ==
1. The widget in the sidebar
2. The widget options
== Changelog ==
= 0.5.3 =
* Add option to show forked repos
= 0.5.2 =
* Add option to modify the title of the widget
= 0.5.1 =
* Add alt attribute to the octocat image
* Fix the a tag
= 0.5 =
* First release
| 2 | 0.04 | 1 | 1 |
d452253ebdd186ea76cd0bce6734e977356edc2e | app.css | app.css | .card {
display: inline-block;
border: thin solid #000;
overflow: hidden;
white-space: nowrap;
margin: 0.5ex;
}
.card.sheep > .body {
display: inline-block;
width: 5.6ex;
height: 3.8ex;
padding: 0.5ex;
border-top: 2ex solid #cff;
border-bottom: 2ex solid #3c3;
background: #fff;
text-align: center;
}
.card.event > .body {
display: inline-block;
width: 5.1ex;
padding: 0.5ex 0.5ex 0.5ex 0;
background: #ec8;
text-align: center;
line-height: 7.8ex;
border-left: 0.5ex solid #db7;
}
| .card {
display: inline-block;
border: thin solid #000;
overflow: hidden;
white-space: nowrap;
margin: 0.5ex;
}
.card.sheep > .body {
display: inline-block;
width: 5.6ex;
height: 3.8ex;
padding: 0.5ex;
border-top: 2ex solid #cff;
border-bottom: 2ex solid #3c3;
background: #fff;
text-align: center;
}
.card.sheep.rank1 > .body {border-bottom-color: #ce0;}
.card.sheep.rank3 > .body {border-bottom-color: #9e0;}
.card.sheep.rank10 > .body {border-bottom-color: #6e0;}
.card.sheep.rank30 > .body {border-bottom-color: #0d0;}
.card.sheep.rank100 > .body {border-bottom-color: #0a0;}
.card.sheep.rank300 > .body {border-bottom-color: #080;}
.card.sheep.rank1000 > .body {border-bottom-color: #060;}
.card.event > .body {
display: inline-block;
width: 5.1ex;
padding: 0.5ex 0.5ex 0.5ex 0;
background: #ec8;
text-align: center;
line-height: 7.8ex;
border-left: 0.5ex solid #db7;
}
| Adjust border colors of sheep cards by their ranks | Adjust border colors of sheep cards by their ranks
| CSS | mit | kana/shephy-js,kana/shephy-js | css | ## Code Before:
.card {
display: inline-block;
border: thin solid #000;
overflow: hidden;
white-space: nowrap;
margin: 0.5ex;
}
.card.sheep > .body {
display: inline-block;
width: 5.6ex;
height: 3.8ex;
padding: 0.5ex;
border-top: 2ex solid #cff;
border-bottom: 2ex solid #3c3;
background: #fff;
text-align: center;
}
.card.event > .body {
display: inline-block;
width: 5.1ex;
padding: 0.5ex 0.5ex 0.5ex 0;
background: #ec8;
text-align: center;
line-height: 7.8ex;
border-left: 0.5ex solid #db7;
}
## Instruction:
Adjust border colors of sheep cards by their ranks
## Code After:
.card {
display: inline-block;
border: thin solid #000;
overflow: hidden;
white-space: nowrap;
margin: 0.5ex;
}
.card.sheep > .body {
display: inline-block;
width: 5.6ex;
height: 3.8ex;
padding: 0.5ex;
border-top: 2ex solid #cff;
border-bottom: 2ex solid #3c3;
background: #fff;
text-align: center;
}
.card.sheep.rank1 > .body {border-bottom-color: #ce0;}
.card.sheep.rank3 > .body {border-bottom-color: #9e0;}
.card.sheep.rank10 > .body {border-bottom-color: #6e0;}
.card.sheep.rank30 > .body {border-bottom-color: #0d0;}
.card.sheep.rank100 > .body {border-bottom-color: #0a0;}
.card.sheep.rank300 > .body {border-bottom-color: #080;}
.card.sheep.rank1000 > .body {border-bottom-color: #060;}
.card.event > .body {
display: inline-block;
width: 5.1ex;
padding: 0.5ex 0.5ex 0.5ex 0;
background: #ec8;
text-align: center;
line-height: 7.8ex;
border-left: 0.5ex solid #db7;
}
| .card {
display: inline-block;
border: thin solid #000;
overflow: hidden;
white-space: nowrap;
margin: 0.5ex;
}
.card.sheep > .body {
display: inline-block;
width: 5.6ex;
height: 3.8ex;
padding: 0.5ex;
border-top: 2ex solid #cff;
border-bottom: 2ex solid #3c3;
background: #fff;
text-align: center;
}
+ .card.sheep.rank1 > .body {border-bottom-color: #ce0;}
+ .card.sheep.rank3 > .body {border-bottom-color: #9e0;}
+ .card.sheep.rank10 > .body {border-bottom-color: #6e0;}
+ .card.sheep.rank30 > .body {border-bottom-color: #0d0;}
+ .card.sheep.rank100 > .body {border-bottom-color: #0a0;}
+ .card.sheep.rank300 > .body {border-bottom-color: #080;}
+ .card.sheep.rank1000 > .body {border-bottom-color: #060;}
+
.card.event > .body {
display: inline-block;
width: 5.1ex;
padding: 0.5ex 0.5ex 0.5ex 0;
background: #ec8;
text-align: center;
line-height: 7.8ex;
border-left: 0.5ex solid #db7;
} | 8 | 0.285714 | 8 | 0 |
6543f00333dcb906ac9a2f92c82acd22c9332126 | lib/jammit-s3.rb | lib/jammit-s3.rb | require 'jammit/command_line'
require 'jammit/s3_command_line'
require 'jammit/s3_uploader'
module Jammit
def self.upload_to_s3!(options = {})
S3Uploader.new(options).upload
end
end
if defined?(Rails)
module Jammit
class JammitRailtie < Rails::Railtie
initializer "set asset host and asset id" do
config.after_initialize do
if Jammit.configuration[:use_cloudfront] && Jammit.configuration[:cloudfront_domain].present?
asset_hostname = "http://#{Jammit.configuration[:cloudfront_domain]}"
else
asset_hostname = "http://#{Jammit.configuration[:s3_bucket]}.s3.amazonaws.com"
end
if Jammit.package_assets and asset_hostname.present?
ActionController::Base.asset_host = asset_hostname
end
end
end
end
end
end
| require 'jammit/command_line'
require 'jammit/s3_command_line'
require 'jammit/s3_uploader'
module Jammit
def self.upload_to_s3!(options = {})
S3Uploader.new(options).upload
end
end
if defined?(Rails)
module Jammit
class JammitRailtie < Rails::Railtie
initializer "set asset host and asset id" do
config.after_initialize do
protocol = Jammit.configuration[:ssl] ? "https" : "http"
if Jammit.configuration[:use_cloudfront] && Jammit.configuration[:cloudfront_domain].present?
asset_hostname = "#{protocol}://#{Jammit.configuration[:cloudfront_domain]}"
else
asset_hostname = "#{protocol}://#{Jammit.configuration[:s3_bucket]}.s3.amazonaws.com"
end
if Jammit.package_assets and asset_hostname.present?
ActionController::Base.asset_host = asset_hostname
end
end
end
end
end
end
| Use https protocol when SSL is enabled. | Use https protocol when SSL is enabled.
| Ruby | mit | railsjedi/jammit-s3,jacquescrocker/jammit-s3,pancakelabs/jammit-s3,mark-ellul/jammit-s3 | ruby | ## Code Before:
require 'jammit/command_line'
require 'jammit/s3_command_line'
require 'jammit/s3_uploader'
module Jammit
def self.upload_to_s3!(options = {})
S3Uploader.new(options).upload
end
end
if defined?(Rails)
module Jammit
class JammitRailtie < Rails::Railtie
initializer "set asset host and asset id" do
config.after_initialize do
if Jammit.configuration[:use_cloudfront] && Jammit.configuration[:cloudfront_domain].present?
asset_hostname = "http://#{Jammit.configuration[:cloudfront_domain]}"
else
asset_hostname = "http://#{Jammit.configuration[:s3_bucket]}.s3.amazonaws.com"
end
if Jammit.package_assets and asset_hostname.present?
ActionController::Base.asset_host = asset_hostname
end
end
end
end
end
end
## Instruction:
Use https protocol when SSL is enabled.
## Code After:
require 'jammit/command_line'
require 'jammit/s3_command_line'
require 'jammit/s3_uploader'
module Jammit
def self.upload_to_s3!(options = {})
S3Uploader.new(options).upload
end
end
if defined?(Rails)
module Jammit
class JammitRailtie < Rails::Railtie
initializer "set asset host and asset id" do
config.after_initialize do
protocol = Jammit.configuration[:ssl] ? "https" : "http"
if Jammit.configuration[:use_cloudfront] && Jammit.configuration[:cloudfront_domain].present?
asset_hostname = "#{protocol}://#{Jammit.configuration[:cloudfront_domain]}"
else
asset_hostname = "#{protocol}://#{Jammit.configuration[:s3_bucket]}.s3.amazonaws.com"
end
if Jammit.package_assets and asset_hostname.present?
ActionController::Base.asset_host = asset_hostname
end
end
end
end
end
end
| require 'jammit/command_line'
require 'jammit/s3_command_line'
require 'jammit/s3_uploader'
module Jammit
def self.upload_to_s3!(options = {})
S3Uploader.new(options).upload
end
end
if defined?(Rails)
module Jammit
class JammitRailtie < Rails::Railtie
initializer "set asset host and asset id" do
config.after_initialize do
+ protocol = Jammit.configuration[:ssl] ? "https" : "http"
if Jammit.configuration[:use_cloudfront] && Jammit.configuration[:cloudfront_domain].present?
- asset_hostname = "http://#{Jammit.configuration[:cloudfront_domain]}"
? ^ ^^ -
+ asset_hostname = "#{protocol}://#{Jammit.configuration[:cloudfront_domain]}"
? ^^^^^ ^^^^^
else
- asset_hostname = "http://#{Jammit.configuration[:s3_bucket]}.s3.amazonaws.com"
? ^ ^^
+ asset_hostname = "#{protocol}://#{Jammit.configuration[:s3_bucket]}.s3.amazonaws.com"
? ^^^^^ ^^^^^
end
if Jammit.package_assets and asset_hostname.present?
ActionController::Base.asset_host = asset_hostname
end
end
end
end
end
end | 5 | 0.172414 | 3 | 2 |
e1092caacec94bd016283b8452ad4399dba0f231 | cogs/gaming_tasks.py | cogs/gaming_tasks.py | from discord.ext import commands
from .utils import checks
import asyncio
import discord
import web.wsgi
from django.utils import timezone
from django.db import models
from django.utils import timezone
from gaming.models import DiscordUser, Game, GameUser, Server, Role, GameSearch, Channel
class GamingTasks:
def __init__(self, bot, task):
self.bot = bot
self.task = task
def __unload(self):
self.task.cancel()
@asyncio.coroutine
def run_tasks(bot):
while True:
yield from asyncio.sleep(15)
channels = Channel.objects.filter(private=False, expire_date__lte=timezone.now(), deleted=False)
if channels.count() >= 1:
for channel in channels:
try:
c = bot.get_channel(channel.channel_id)
yield from bot.delete_channel(c)
except Exception as e:
# Channel no longer exists on server
print(e)
finally:
channel.deleted = True
channel.save()
yield from asyncio.sleep(1)
else:
# No channels found to be deleted
pass
def setup(bot):
loop = asyncio.get_event_loop()
task = loop.create_task(GamingTasks.run_tasks(bot))
bot.add_cog(GamingTasks(bot, task))
| from discord.ext import commands
from .utils import checks
import asyncio
import discord
import web.wsgi
from django.utils import timezone
from django.db import models
from django.utils import timezone
from gaming.models import DiscordUser, Game, GameUser, Server, Role, GameSearch, Channel
class GamingTasks:
def __init__(self, bot, task):
self.bot = bot
self.task = task
def __unload(self):
self.task.cancel()
@asyncio.coroutine
def run_tasks(bot):
while True:
yield from asyncio.sleep(15)
channels = Channel.objects.filter(private=False, expire_date__lte=timezone.now(), deleted=False)
if channels.count() >= 1:
for channel in channels:
if channel is None:
continue
try:
c = bot.get_channel(channel.channel_id)
if c is not None:
yield from bot.delete_channel(c)
except Exception as e:
# Channel no longer exists on server
print(e)
finally:
channel.deleted = True
channel.save()
yield from asyncio.sleep(1)
else:
# No channels found to be deleted
pass
def setup(bot):
loop = asyncio.get_event_loop()
task = loop.create_task(GamingTasks.run_tasks(bot))
bot.add_cog(GamingTasks(bot, task))
| Check if it is None before continuing | Check if it is None before continuing
| Python | mit | bsquidwrd/Squid-Bot,bsquidwrd/Squid-Bot | python | ## Code Before:
from discord.ext import commands
from .utils import checks
import asyncio
import discord
import web.wsgi
from django.utils import timezone
from django.db import models
from django.utils import timezone
from gaming.models import DiscordUser, Game, GameUser, Server, Role, GameSearch, Channel
class GamingTasks:
def __init__(self, bot, task):
self.bot = bot
self.task = task
def __unload(self):
self.task.cancel()
@asyncio.coroutine
def run_tasks(bot):
while True:
yield from asyncio.sleep(15)
channels = Channel.objects.filter(private=False, expire_date__lte=timezone.now(), deleted=False)
if channels.count() >= 1:
for channel in channels:
try:
c = bot.get_channel(channel.channel_id)
yield from bot.delete_channel(c)
except Exception as e:
# Channel no longer exists on server
print(e)
finally:
channel.deleted = True
channel.save()
yield from asyncio.sleep(1)
else:
# No channels found to be deleted
pass
def setup(bot):
loop = asyncio.get_event_loop()
task = loop.create_task(GamingTasks.run_tasks(bot))
bot.add_cog(GamingTasks(bot, task))
## Instruction:
Check if it is None before continuing
## Code After:
from discord.ext import commands
from .utils import checks
import asyncio
import discord
import web.wsgi
from django.utils import timezone
from django.db import models
from django.utils import timezone
from gaming.models import DiscordUser, Game, GameUser, Server, Role, GameSearch, Channel
class GamingTasks:
def __init__(self, bot, task):
self.bot = bot
self.task = task
def __unload(self):
self.task.cancel()
@asyncio.coroutine
def run_tasks(bot):
while True:
yield from asyncio.sleep(15)
channels = Channel.objects.filter(private=False, expire_date__lte=timezone.now(), deleted=False)
if channels.count() >= 1:
for channel in channels:
if channel is None:
continue
try:
c = bot.get_channel(channel.channel_id)
if c is not None:
yield from bot.delete_channel(c)
except Exception as e:
# Channel no longer exists on server
print(e)
finally:
channel.deleted = True
channel.save()
yield from asyncio.sleep(1)
else:
# No channels found to be deleted
pass
def setup(bot):
loop = asyncio.get_event_loop()
task = loop.create_task(GamingTasks.run_tasks(bot))
bot.add_cog(GamingTasks(bot, task))
| from discord.ext import commands
from .utils import checks
import asyncio
import discord
import web.wsgi
from django.utils import timezone
from django.db import models
from django.utils import timezone
from gaming.models import DiscordUser, Game, GameUser, Server, Role, GameSearch, Channel
class GamingTasks:
def __init__(self, bot, task):
self.bot = bot
self.task = task
def __unload(self):
self.task.cancel()
@asyncio.coroutine
def run_tasks(bot):
while True:
yield from asyncio.sleep(15)
channels = Channel.objects.filter(private=False, expire_date__lte=timezone.now(), deleted=False)
if channels.count() >= 1:
for channel in channels:
+ if channel is None:
+ continue
try:
c = bot.get_channel(channel.channel_id)
+ if c is not None:
- yield from bot.delete_channel(c)
+ yield from bot.delete_channel(c)
? ++++
except Exception as e:
# Channel no longer exists on server
print(e)
finally:
channel.deleted = True
channel.save()
yield from asyncio.sleep(1)
else:
# No channels found to be deleted
pass
def setup(bot):
loop = asyncio.get_event_loop()
task = loop.create_task(GamingTasks.run_tasks(bot))
bot.add_cog(GamingTasks(bot, task)) | 5 | 0.111111 | 4 | 1 |
4f53cceb5abf057ba1ea1e8d6aafd3b2831cf172 | spec/integration/default/default_spec.rb | spec/integration/default/default_spec.rb | require 'chef_bones/integration_spec_helper'
describe 'The recipe sharness::default' do
let (:prefix) { '/usr/local' }
it 'installs sharness' do
expect(file "#{prefix}/share/sharness/sharness.sh").to be_file
expect(file "#{prefix}/share/sharness/aggregate-results.sh").to be_file
end
it 'installs documentation' do
expect(file "#{prefix}/share/doc/sharness/README.md").to be_file
expect(file "#{prefix}/share/doc/sharness/COPYING").to be_file
end
it 'installs working examples' do
expect(file "#{prefix}/share/doc/sharness/examples/Makefile").to be_file
expect(command "sudo make -C #{prefix}/share/doc/sharness/examples").to return_exit_status 0
end
end
| require 'chef_bones/integration_spec_helper'
describe 'sharness::default' do
let (:prefix) { '/usr/local' }
it 'installs sharness library' do
expect(file "#{prefix}/share/sharness/sharness.sh").to be_file
end
it 'installs aggregate script' do
expect(file "#{prefix}/share/sharness/aggregate-results.sh").to be_file
end
it 'installs documentation' do
expect(file "#{prefix}/share/doc/sharness/README.md").to be_file
end
context 'test examples' do
let (:examples) { "#{prefix}/share/doc/sharness/examples" }
it 'are installed' do
expect(file "#{examples}/Makefile").to be_file
end
it 'succeed when run with make test' do
expect(command "sudo make test -C #{examples}").to return_exit_status 0
end
it 'succeed when run with make prove' do
expect(command "sudo make prove -C #{examples}").to return_exit_status 0
end
end
end
| Add context 'test examples' to integration test | Add context 'test examples' to integration test
| Ruby | apache-2.0 | mlafeldt/sharness-cookbook | ruby | ## Code Before:
require 'chef_bones/integration_spec_helper'
describe 'The recipe sharness::default' do
let (:prefix) { '/usr/local' }
it 'installs sharness' do
expect(file "#{prefix}/share/sharness/sharness.sh").to be_file
expect(file "#{prefix}/share/sharness/aggregate-results.sh").to be_file
end
it 'installs documentation' do
expect(file "#{prefix}/share/doc/sharness/README.md").to be_file
expect(file "#{prefix}/share/doc/sharness/COPYING").to be_file
end
it 'installs working examples' do
expect(file "#{prefix}/share/doc/sharness/examples/Makefile").to be_file
expect(command "sudo make -C #{prefix}/share/doc/sharness/examples").to return_exit_status 0
end
end
## Instruction:
Add context 'test examples' to integration test
## Code After:
require 'chef_bones/integration_spec_helper'
describe 'sharness::default' do
let (:prefix) { '/usr/local' }
it 'installs sharness library' do
expect(file "#{prefix}/share/sharness/sharness.sh").to be_file
end
it 'installs aggregate script' do
expect(file "#{prefix}/share/sharness/aggregate-results.sh").to be_file
end
it 'installs documentation' do
expect(file "#{prefix}/share/doc/sharness/README.md").to be_file
end
context 'test examples' do
let (:examples) { "#{prefix}/share/doc/sharness/examples" }
it 'are installed' do
expect(file "#{examples}/Makefile").to be_file
end
it 'succeed when run with make test' do
expect(command "sudo make test -C #{examples}").to return_exit_status 0
end
it 'succeed when run with make prove' do
expect(command "sudo make prove -C #{examples}").to return_exit_status 0
end
end
end
| require 'chef_bones/integration_spec_helper'
- describe 'The recipe sharness::default' do
? -----------
+ describe 'sharness::default' do
let (:prefix) { '/usr/local' }
- it 'installs sharness' do
+ it 'installs sharness library' do
? ++++++++
expect(file "#{prefix}/share/sharness/sharness.sh").to be_file
+ end
+
+ it 'installs aggregate script' do
expect(file "#{prefix}/share/sharness/aggregate-results.sh").to be_file
end
it 'installs documentation' do
expect(file "#{prefix}/share/doc/sharness/README.md").to be_file
- expect(file "#{prefix}/share/doc/sharness/COPYING").to be_file
end
- it 'installs working examples' do
+ context 'test examples' do
+ let (:examples) { "#{prefix}/share/doc/sharness/examples" }
+
+ it 'are installed' do
- expect(file "#{prefix}/share/doc/sharness/examples/Makefile").to be_file
? ---------------------------
+ expect(file "#{examples}/Makefile").to be_file
? ++ +
+ end
+
+ it 'succeed when run with make test' do
- expect(command "sudo make -C #{prefix}/share/doc/sharness/examples").to return_exit_status 0
? ---------------------------
+ expect(command "sudo make test -C #{examples}").to return_exit_status 0
? ++ +++++ +
+ end
+
+ it 'succeed when run with make prove' do
+ expect(command "sudo make prove -C #{examples}").to return_exit_status 0
+ end
end
end | 25 | 1.25 | 19 | 6 |
f6316d222bb621385ebbb8d6f2b34af306bda098 | build/nixpkgs.nix | build/nixpkgs.nix | with (import <nixpkgs> {});
stdenv.mkDerivation {
name = "cassava-streams";
buildInputs = [
# GHC:
haskell.packages.lts-6_7.ghc
# Non-Haskell Dependencies:
zlib
];
}
| with (import <nixpkgs> {});
stdenv.mkDerivation {
name = "cassava-streams";
buildInputs = [
# GHC:
haskell.packages.lts-6_7.ghc
# Non-Haskell Dependencies:
gnupg # For signing packages.
zlib
];
}
| Add gpg for signing releases | Add gpg for signing releases
| Nix | bsd-3-clause | pjones/cassava-streams | nix | ## Code Before:
with (import <nixpkgs> {});
stdenv.mkDerivation {
name = "cassava-streams";
buildInputs = [
# GHC:
haskell.packages.lts-6_7.ghc
# Non-Haskell Dependencies:
zlib
];
}
## Instruction:
Add gpg for signing releases
## Code After:
with (import <nixpkgs> {});
stdenv.mkDerivation {
name = "cassava-streams";
buildInputs = [
# GHC:
haskell.packages.lts-6_7.ghc
# Non-Haskell Dependencies:
gnupg # For signing packages.
zlib
];
}
| with (import <nixpkgs> {});
stdenv.mkDerivation {
name = "cassava-streams";
buildInputs = [
# GHC:
haskell.packages.lts-6_7.ghc
# Non-Haskell Dependencies:
+ gnupg # For signing packages.
zlib
];
} | 1 | 0.076923 | 1 | 0 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.