commit stringlengths 40 40 | old_file stringlengths 4 184 | new_file stringlengths 4 184 | old_contents stringlengths 1 3.6k | new_contents stringlengths 5 3.38k | subject stringlengths 15 778 | message stringlengths 16 6.74k | lang stringclasses 201 values | license stringclasses 13 values | repos stringlengths 6 116k | config stringclasses 201 values | content stringlengths 137 7.24k | diff stringlengths 26 5.55k | diff_length int64 1 123 | relative_diff_length float64 0.01 89 | n_lines_added int64 0 108 | n_lines_deleted int64 0 106 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
beef8e07bda93ff95162a9f5d9b7ec1668e20d2b | lib/tasks/deployment.rake | lib/tasks/deployment.rake |
desc 'Deploy, migrate and restart processes'
task :deploy => 'knight:deploy_and_migrate'
desc 'Simple deploy without migration'
task 'deploy:simple' => 'knight:deploy'
namespace :knight do
desc 'Deploy to Heroku'
task :deploy do
puts 'Deploying to production...'
system 'git push heroku master'
end
desc 'Migrate the database'
task :migrate do
puts 'Migrating Database...'
system 'heroku run rake db:migrate'
end
desc 'Migrate new seeds'
task :seed do
puts 'Migrating new Seeds...'
system 'heroku run rake db:seed'
end
desc 'Restart processes'
task :restart do
puts 'Restarting Processes...'
system 'heroku restart'
end
desc ':deploy, :migrate, :restart'
task :deploy_and_migrate => [:deploy, :migrate, :restart]
desc 'Perform serial Knight Update'
task :update_knight do
puts 'Processing update Knight in production...'
system 'heroku run rails runner KnightUpdater.update_knight_serial'
end
desc 'Bust Caches'
task :bust_caches do
puts 'Busting Caches...'
system 'heroku run rails runner Repo.bust_caches'
system 'heroku run rails runner Category.bust_caches'
end
desc 'Clear away empty categories'
task :clean_empty_categories do
puts 'Cleaning Knight of empty categories...'
system 'heroku run rails runner Category.clean'
end
end
|
desc 'Deploy, migrate and restart processes'
task :deploy => 'knight:deploy_and_migrate'
desc 'Simple deploy without migration'
task 'deploy:simple' => 'knight:deploy'
namespace :knight do
desc 'Deploy to Heroku'
task :deploy do
puts 'Pushing to Github...'
system 'git push origin master'
puts 'Deploying to Heroku production...'
system 'git push heroku master'
end
desc 'Migrate the database'
task :migrate do
puts 'Migrating Database...'
system 'heroku run rake db:migrate'
end
desc 'Migrate new seeds'
task :seed do
puts 'Migrating new Seeds...'
system 'heroku run rake db:seed'
end
desc 'Restart processes'
task :restart do
puts 'Restarting Processes...'
system 'heroku restart'
end
desc ':deploy, :migrate, :restart'
task :deploy_and_migrate => [:deploy, :migrate, :restart]
desc 'Perform serial Knight Update'
task :update_knight do
puts 'Processing update Knight in production...'
system 'heroku run rails runner KnightUpdater.update_knight_serial'
end
desc 'Bust Caches'
task :bust_caches do
puts 'Busting Caches...'
system 'heroku run rails runner Repo.bust_caches'
system 'heroku run rails runner Category.bust_caches'
end
desc 'Clear away empty categories'
task :clean_empty_categories do
puts 'Cleaning Knight of empty categories...'
system 'heroku run rails runner Category.clean'
end
end
| Deploy task pushed to github, too | Rake: Deploy task pushed to github, too
| Ruby | mit | thomasklemm/pluginGeek | ruby | ## Code Before:
desc 'Deploy, migrate and restart processes'
task :deploy => 'knight:deploy_and_migrate'
desc 'Simple deploy without migration'
task 'deploy:simple' => 'knight:deploy'
namespace :knight do
desc 'Deploy to Heroku'
task :deploy do
puts 'Deploying to production...'
system 'git push heroku master'
end
desc 'Migrate the database'
task :migrate do
puts 'Migrating Database...'
system 'heroku run rake db:migrate'
end
desc 'Migrate new seeds'
task :seed do
puts 'Migrating new Seeds...'
system 'heroku run rake db:seed'
end
desc 'Restart processes'
task :restart do
puts 'Restarting Processes...'
system 'heroku restart'
end
desc ':deploy, :migrate, :restart'
task :deploy_and_migrate => [:deploy, :migrate, :restart]
desc 'Perform serial Knight Update'
task :update_knight do
puts 'Processing update Knight in production...'
system 'heroku run rails runner KnightUpdater.update_knight_serial'
end
desc 'Bust Caches'
task :bust_caches do
puts 'Busting Caches...'
system 'heroku run rails runner Repo.bust_caches'
system 'heroku run rails runner Category.bust_caches'
end
desc 'Clear away empty categories'
task :clean_empty_categories do
puts 'Cleaning Knight of empty categories...'
system 'heroku run rails runner Category.clean'
end
end
## Instruction:
Rake: Deploy task pushed to github, too
## Code After:
desc 'Deploy, migrate and restart processes'
task :deploy => 'knight:deploy_and_migrate'
desc 'Simple deploy without migration'
task 'deploy:simple' => 'knight:deploy'
namespace :knight do
desc 'Deploy to Heroku'
task :deploy do
puts 'Pushing to Github...'
system 'git push origin master'
puts 'Deploying to Heroku production...'
system 'git push heroku master'
end
desc 'Migrate the database'
task :migrate do
puts 'Migrating Database...'
system 'heroku run rake db:migrate'
end
desc 'Migrate new seeds'
task :seed do
puts 'Migrating new Seeds...'
system 'heroku run rake db:seed'
end
desc 'Restart processes'
task :restart do
puts 'Restarting Processes...'
system 'heroku restart'
end
desc ':deploy, :migrate, :restart'
task :deploy_and_migrate => [:deploy, :migrate, :restart]
desc 'Perform serial Knight Update'
task :update_knight do
puts 'Processing update Knight in production...'
system 'heroku run rails runner KnightUpdater.update_knight_serial'
end
desc 'Bust Caches'
task :bust_caches do
puts 'Busting Caches...'
system 'heroku run rails runner Repo.bust_caches'
system 'heroku run rails runner Category.bust_caches'
end
desc 'Clear away empty categories'
task :clean_empty_categories do
puts 'Cleaning Knight of empty categories...'
system 'heroku run rails runner Category.clean'
end
end
|
desc 'Deploy, migrate and restart processes'
task :deploy => 'knight:deploy_and_migrate'
desc 'Simple deploy without migration'
task 'deploy:simple' => 'knight:deploy'
namespace :knight do
desc 'Deploy to Heroku'
task :deploy do
+ puts 'Pushing to Github...'
+ system 'git push origin master'
- puts 'Deploying to production...'
+ puts 'Deploying to Heroku production...'
? +++++++
system 'git push heroku master'
end
desc 'Migrate the database'
task :migrate do
puts 'Migrating Database...'
system 'heroku run rake db:migrate'
end
desc 'Migrate new seeds'
task :seed do
puts 'Migrating new Seeds...'
system 'heroku run rake db:seed'
end
desc 'Restart processes'
task :restart do
puts 'Restarting Processes...'
system 'heroku restart'
end
desc ':deploy, :migrate, :restart'
task :deploy_and_migrate => [:deploy, :migrate, :restart]
desc 'Perform serial Knight Update'
task :update_knight do
puts 'Processing update Knight in production...'
system 'heroku run rails runner KnightUpdater.update_knight_serial'
end
desc 'Bust Caches'
task :bust_caches do
puts 'Busting Caches...'
system 'heroku run rails runner Repo.bust_caches'
system 'heroku run rails runner Category.bust_caches'
end
desc 'Clear away empty categories'
task :clean_empty_categories do
puts 'Cleaning Knight of empty categories...'
system 'heroku run rails runner Category.clean'
end
end | 4 | 0.074074 | 3 | 1 |
9e7379f863ab7bb8b4ae2fd39530692a427df5a1 | test/loader.js | test/loader.js | /*globals mocha*/
require.config({
baseUrl: '/',
paths: {
'tabler': '../',
'jquery': 'test/lib/jquery-1.7.1',
'underscore': 'test/lib/underscore'
}
});
mocha.setup({
ui: 'bdd',
globals: ['jQuery']
});
require(['jquery',
'test/tabler.tests',
'test/tabler.aggregator.tests',
'test/tabler.columnGrouper.tests',
'test/tabler.jumpToPage.tests',
'test/tabler.pageSize.tests',
'test/tabler.pager.tests',
'test/tabler.toggleColumns.tests',
'test/tabler.sortable.tests',
'test/tabler.columnGrouper.tests'
], function(){
'use strict';
mocha.run();
});
| /*globals mocha*/
require.config({
baseUrl: '/',
paths: {
'tabler': '../',
'jquery': 'test/lib/jquery-1.7.1',
'underscore': 'test/lib/underscore'
}
});
mocha.setup({
ui: 'bdd',
globals: ['jQuery']
});
require(['jquery',
'test/tabler.tests',
'test/tabler.aggregator.tests',
'test/tabler.columnGrouper.tests',
'test/tabler.jumpToPage.tests',
'test/tabler.pageSize.tests',
'test/tabler.pager.tests',
'test/tabler.toggleColumns.tests',
'test/tabler.removeColumns.tests',
'test/tabler.sortable.tests',
'test/tabler.columnGrouper.tests'
], function(){
'use strict';
mocha.run();
});
| Add removeColumns plugin to the test suite | Add removeColumns plugin to the test suite
| JavaScript | mit | BrandwatchLtd/tabler,BrandwatchLtd/tabler,BrandwatchLtd/tabler | javascript | ## Code Before:
/*globals mocha*/
require.config({
baseUrl: '/',
paths: {
'tabler': '../',
'jquery': 'test/lib/jquery-1.7.1',
'underscore': 'test/lib/underscore'
}
});
mocha.setup({
ui: 'bdd',
globals: ['jQuery']
});
require(['jquery',
'test/tabler.tests',
'test/tabler.aggregator.tests',
'test/tabler.columnGrouper.tests',
'test/tabler.jumpToPage.tests',
'test/tabler.pageSize.tests',
'test/tabler.pager.tests',
'test/tabler.toggleColumns.tests',
'test/tabler.sortable.tests',
'test/tabler.columnGrouper.tests'
], function(){
'use strict';
mocha.run();
});
## Instruction:
Add removeColumns plugin to the test suite
## Code After:
/*globals mocha*/
require.config({
baseUrl: '/',
paths: {
'tabler': '../',
'jquery': 'test/lib/jquery-1.7.1',
'underscore': 'test/lib/underscore'
}
});
mocha.setup({
ui: 'bdd',
globals: ['jQuery']
});
require(['jquery',
'test/tabler.tests',
'test/tabler.aggregator.tests',
'test/tabler.columnGrouper.tests',
'test/tabler.jumpToPage.tests',
'test/tabler.pageSize.tests',
'test/tabler.pager.tests',
'test/tabler.toggleColumns.tests',
'test/tabler.removeColumns.tests',
'test/tabler.sortable.tests',
'test/tabler.columnGrouper.tests'
], function(){
'use strict';
mocha.run();
});
| /*globals mocha*/
require.config({
baseUrl: '/',
paths: {
'tabler': '../',
'jquery': 'test/lib/jquery-1.7.1',
'underscore': 'test/lib/underscore'
}
});
mocha.setup({
ui: 'bdd',
globals: ['jQuery']
});
require(['jquery',
'test/tabler.tests',
'test/tabler.aggregator.tests',
'test/tabler.columnGrouper.tests',
'test/tabler.jumpToPage.tests',
'test/tabler.pageSize.tests',
'test/tabler.pager.tests',
'test/tabler.toggleColumns.tests',
+ 'test/tabler.removeColumns.tests',
'test/tabler.sortable.tests',
'test/tabler.columnGrouper.tests'
], function(){
'use strict';
mocha.run();
}); | 1 | 0.037037 | 1 | 0 |
8e98701a349105d2f306f872107690fa6a012080 | ccvpn2_web/templates/config.ovpn.mako | ccvpn2_web/templates/config.ovpn.mako |
verb 3
client
tls-client
dev tun0
tun-ipv6
auth-user-pass
# you can use this and put username/password, one per line, in a file
# auth-user-pass cred.txt
% if android:
# de.blinkt.openvpn does not support <connection>
remote ${remotes[0]} 443 tcp
% else:
% for remote in remotes:
remote ${remote} 1194 udp
remote ${remote} 443 tcp
% endfor
% endif
resolv-retry infinite
nobind
persist-key
persist-tun
comp-lzo
redirect-gateway def1
<ca>
${ca_content}</ca>
|
verb 3
client
tls-client
dev tun0
tun-ipv6
auth-user-pass
# you can use this and put username/password, one per line, in a file
# auth-user-pass cred.txt
% if android:
# de.blinkt.openvpn does not support <connection>
remote ${remotes[0]} 443 tcp
% else:
% for remote in remotes:
remote ${remote} 1194 udp
remote ${remote} 443 tcp
% endfor
% endif
resolv-retry infinite
nobind
persist-key
persist-tun
comp-lzo
redirect-gateway def1
<ca>
${ca_content}</ca>
| Remove spaces in config. Network-manager, fuck you. | Remove spaces in config. Network-manager, fuck you.
| Mako | mit | bspavel/ccvpn,bspavel/ccvpn,bspavel/ccvpn | mako | ## Code Before:
verb 3
client
tls-client
dev tun0
tun-ipv6
auth-user-pass
# you can use this and put username/password, one per line, in a file
# auth-user-pass cred.txt
% if android:
# de.blinkt.openvpn does not support <connection>
remote ${remotes[0]} 443 tcp
% else:
% for remote in remotes:
remote ${remote} 1194 udp
remote ${remote} 443 tcp
% endfor
% endif
resolv-retry infinite
nobind
persist-key
persist-tun
comp-lzo
redirect-gateway def1
<ca>
${ca_content}</ca>
## Instruction:
Remove spaces in config. Network-manager, fuck you.
## Code After:
verb 3
client
tls-client
dev tun0
tun-ipv6
auth-user-pass
# you can use this and put username/password, one per line, in a file
# auth-user-pass cred.txt
% if android:
# de.blinkt.openvpn does not support <connection>
remote ${remotes[0]} 443 tcp
% else:
% for remote in remotes:
remote ${remote} 1194 udp
remote ${remote} 443 tcp
% endfor
% endif
resolv-retry infinite
nobind
persist-key
persist-tun
comp-lzo
redirect-gateway def1
<ca>
${ca_content}</ca>
|
verb 3
client
tls-client
dev tun0
tun-ipv6
auth-user-pass
# you can use this and put username/password, one per line, in a file
# auth-user-pass cred.txt
% if android:
- # de.blinkt.openvpn does not support <connection>
? ----
+ # de.blinkt.openvpn does not support <connection>
- remote ${remotes[0]} 443 tcp
? ----
+ remote ${remotes[0]} 443 tcp
% else:
- % for remote in remotes:
? ----
+ % for remote in remotes:
- remote ${remote} 1194 udp
? ----
+ remote ${remote} 1194 udp
- remote ${remote} 443 tcp
? ----
+ remote ${remote} 443 tcp
- % endfor
? ----
+ % endfor
% endif
resolv-retry infinite
nobind
persist-key
persist-tun
comp-lzo
redirect-gateway def1
<ca>
${ca_content}</ca>
| 12 | 0.387097 | 6 | 6 |
2d2589a8df7bb181f8b23e137578dbabadc1b1f5 | src/components/ImageView.jsx | src/components/ImageView.jsx | // @flow
import React, { Component } from 'react';
import Image from '../domain/Image';
export default class ImageView extends Component {
props: {
image: Image,
showPreview: Function,
remove: Function,
};
render() {
return (
<li className="image-view">
<div className="image-container">
<img alt="approve" style={{ maxHeight: '180px', maxWidth: '180px' }} src={this.props.image.src} />
</div>
<div
className="remove-button-container"
onClick={() => this.props.showPreview(this.props.image.src)}
>
<button
type="button"
className="remove-button" aria-label="Close"
onClick={() => this.props.remove(this.props.image)}
>
<i className="fa fa-trash" aria-hidden="true" />
</button>
</div>
</li>
);
}
}
| // @flow
import React, { Component } from 'react';
import Image from '../domain/Image';
export default class ImageView extends Component {
props: {
image: Image,
showPreview: Function,
remove: Function,
};
render() {
return (
<li className="image-view">
<div className="image-container">
<img alt="approve" style={{ maxHeight: '180px', maxWidth: '180px' }} src={this.props.image.src} />
</div>
<a
tabIndex="-1"
className="remove-button-container"
onClick={() => { this.props.showPreview(this.props.image.src); }}
>
<button
type="button"
className="remove-button"
aria-label="Close"
onClick={(ev) => {
ev.stopPropagation();
this.props.remove(this.props.image);
}}
>
<i className="fa fa-trash" aria-hidden="true" />
</button>
</a>
</li>
);
}
}
| Fix remove button event bubling | Fix remove button event bubling
| JSX | mit | mizoguche/approve-pr-with-image,mizoguche/approve-pr-with-image | jsx | ## Code Before:
// @flow
import React, { Component } from 'react';
import Image from '../domain/Image';
export default class ImageView extends Component {
props: {
image: Image,
showPreview: Function,
remove: Function,
};
render() {
return (
<li className="image-view">
<div className="image-container">
<img alt="approve" style={{ maxHeight: '180px', maxWidth: '180px' }} src={this.props.image.src} />
</div>
<div
className="remove-button-container"
onClick={() => this.props.showPreview(this.props.image.src)}
>
<button
type="button"
className="remove-button" aria-label="Close"
onClick={() => this.props.remove(this.props.image)}
>
<i className="fa fa-trash" aria-hidden="true" />
</button>
</div>
</li>
);
}
}
## Instruction:
Fix remove button event bubling
## Code After:
// @flow
import React, { Component } from 'react';
import Image from '../domain/Image';
export default class ImageView extends Component {
props: {
image: Image,
showPreview: Function,
remove: Function,
};
render() {
return (
<li className="image-view">
<div className="image-container">
<img alt="approve" style={{ maxHeight: '180px', maxWidth: '180px' }} src={this.props.image.src} />
</div>
<a
tabIndex="-1"
className="remove-button-container"
onClick={() => { this.props.showPreview(this.props.image.src); }}
>
<button
type="button"
className="remove-button"
aria-label="Close"
onClick={(ev) => {
ev.stopPropagation();
this.props.remove(this.props.image);
}}
>
<i className="fa fa-trash" aria-hidden="true" />
</button>
</a>
</li>
);
}
}
| // @flow
import React, { Component } from 'react';
import Image from '../domain/Image';
export default class ImageView extends Component {
props: {
image: Image,
showPreview: Function,
remove: Function,
};
render() {
return (
<li className="image-view">
<div className="image-container">
<img alt="approve" style={{ maxHeight: '180px', maxWidth: '180px' }} src={this.props.image.src} />
</div>
- <div
? ^^^
+ <a
? ^
+ tabIndex="-1"
className="remove-button-container"
- onClick={() => this.props.showPreview(this.props.image.src)}
+ onClick={() => { this.props.showPreview(this.props.image.src); }}
? ++ ++ +
>
<button
type="button"
- className="remove-button" aria-label="Close"
? -------------------
+ className="remove-button"
+ aria-label="Close"
+ onClick={(ev) => {
+ ev.stopPropagation();
- onClick={() => this.props.remove(this.props.image)}
? ----------- -- ^
+ this.props.remove(this.props.image);
? ^
+ }}
>
<i className="fa fa-trash" aria-hidden="true" />
</button>
- </div>
? ^^^
+ </a>
? ^
</li>
);
}
} | 15 | 0.454545 | 10 | 5 |
955f58d9de5e0ddc5b3c661b5cbc51d9330e4f47 | README.md | README.md | GMLRVA - Generic Multiple Layout Recycler View Adapter
===============
[  ](https://bintray.com/simdea/GMLRVA/gmlrva.lib/_latestVersion)
Generic RecyclerView Adapter that supports multiple layout implementations.
Build status
------------
| Master | [](https://travis-ci.org/Simdea/gmlrva) |
|----------|-------------|
| **Dev** | [](https://travis-ci.org/Simdea/gmlrva) |
Usage
-----
Please check the sample app present on this repository.
Configuration
-------------
### Install
##### Add to app build.gradle:
```gradle
dependencies {
...
compile 'pt.simdea:gmlrva.lib:(last version)'
...
}
```
Developed By
============
This library is owned and maintained by [Simdea][1].
License
=======
See [LICENSE.txt][2].
Original work licensed under [MIT license][3].
[1]: http://simdea.pt/
[2]: LICENSE.txt
[3]: https://github.com/noveogroup/android-check/blob/master/LICENSE.txt
| GMLRVA - Generic Multiple Layout Recycler View Adapter
===============
[  ](https://bintray.com/simdea/GMLRVA/gmlrva.lib/_latestVersion)
Generic RecyclerView Adapter that supports multiple layout implementations.
Build status
------------
| Master | [](https://travis-ci.org/Simdea/gmlrva) |
|----------|-------------|
| **Dev** | [](https://travis-ci.org/Simdea/gmlrva) |
Usage
-----
Please check the sample app present on this repository.
Configuration
-------------
### Install
##### Add to app build.gradle:
```gradle
dependencies {
...
compile 'pt.simdea:gmlrva.lib:(last version)'
...
}
```
Developed By
============
This library is owned and maintained by [Simdea][1].
License
=======
See [LICENSE.txt][2].
Original work licensed under [MIT license][3].
[1]: http://simdea.pt/
[2]: LICENSE.txt
[3]: https://github.com/noveogroup/android-check/blob/master/LICENSE.txt
| Update to the Readme.md file. | [Fix] Update to the Readme.md file.
| Markdown | mit | Simdea/gmlrva | markdown | ## Code Before:
GMLRVA - Generic Multiple Layout Recycler View Adapter
===============
[  ](https://bintray.com/simdea/GMLRVA/gmlrva.lib/_latestVersion)
Generic RecyclerView Adapter that supports multiple layout implementations.
Build status
------------
| Master | [](https://travis-ci.org/Simdea/gmlrva) |
|----------|-------------|
| **Dev** | [](https://travis-ci.org/Simdea/gmlrva) |
Usage
-----
Please check the sample app present on this repository.
Configuration
-------------
### Install
##### Add to app build.gradle:
```gradle
dependencies {
...
compile 'pt.simdea:gmlrva.lib:(last version)'
...
}
```
Developed By
============
This library is owned and maintained by [Simdea][1].
License
=======
See [LICENSE.txt][2].
Original work licensed under [MIT license][3].
[1]: http://simdea.pt/
[2]: LICENSE.txt
[3]: https://github.com/noveogroup/android-check/blob/master/LICENSE.txt
## Instruction:
[Fix] Update to the Readme.md file.
## Code After:
GMLRVA - Generic Multiple Layout Recycler View Adapter
===============
[  ](https://bintray.com/simdea/GMLRVA/gmlrva.lib/_latestVersion)
Generic RecyclerView Adapter that supports multiple layout implementations.
Build status
------------
| Master | [](https://travis-ci.org/Simdea/gmlrva) |
|----------|-------------|
| **Dev** | [](https://travis-ci.org/Simdea/gmlrva) |
Usage
-----
Please check the sample app present on this repository.
Configuration
-------------
### Install
##### Add to app build.gradle:
```gradle
dependencies {
...
compile 'pt.simdea:gmlrva.lib:(last version)'
...
}
```
Developed By
============
This library is owned and maintained by [Simdea][1].
License
=======
See [LICENSE.txt][2].
Original work licensed under [MIT license][3].
[1]: http://simdea.pt/
[2]: LICENSE.txt
[3]: https://github.com/noveogroup/android-check/blob/master/LICENSE.txt
| GMLRVA - Generic Multiple Layout Recycler View Adapter
===============
[  ](https://bintray.com/simdea/GMLRVA/gmlrva.lib/_latestVersion)
Generic RecyclerView Adapter that supports multiple layout implementations.
Build status
------------
| Master | [](https://travis-ci.org/Simdea/gmlrva) |
|----------|-------------|
- | **Dev** | [](https://travis-ci.org/Simdea/gmlrva) |
? ^^
+ | **Dev** | [](https://travis-ci.org/Simdea/gmlrva) |
? + ++++ ^^^^^^^ ++++
Usage
-----
Please check the sample app present on this repository.
Configuration
-------------
### Install
##### Add to app build.gradle:
```gradle
dependencies {
...
compile 'pt.simdea:gmlrva.lib:(last version)'
...
}
```
Developed By
============
This library is owned and maintained by [Simdea][1].
License
=======
See [LICENSE.txt][2].
Original work licensed under [MIT license][3].
[1]: http://simdea.pt/
[2]: LICENSE.txt
[3]: https://github.com/noveogroup/android-check/blob/master/LICENSE.txt | 2 | 0.043478 | 1 | 1 |
36013acb135246818a76ac249576984e8d3249bb | appveyor.yml | appveyor.yml | version: "{build}"
platform:
- x86
- x64
clone_folder: c:\gopath\src\github.com\vlifesystems\rulehuntersrv
environment:
GOPATH: c:\gopath
GOROOT: c:\go-x86
install:
- IF "%PLATFORM%" == "x86" set PATH=%GOROOT%\bin;C:\MinGW\bin;%PATH%
- IF "%PLATFORM%" == "x64" set PATH=C:\msys64\mingw64\bin;%PATH%
- echo %PATH%
- go version
- go env
- go get -v -t ./...
build_script:
- go install -v ./...
test_script:
- go test -v ./...
| version: "{build}"
platform:
- x86
- x64
clone_folder: c:\gopath\src\github.com\vlifesystems\rulehuntersrv
environment:
GOPATH: c:\gopath
install:
- IF "%PLATFORM%" == "x86" (
set GOROOT=c:\go-x86
set PATH=%GOROOT%\bin;C:\MinGW\bin;%PATH%
)
- IF "%PLATFORM%" == "x64" set PATH=C:\msys64\mingw64\bin;%PATH%
- echo %PATH%
- go version
- go env
- go get -v -t ./...
build_script:
- go install -v ./...
test_script:
- go test -v ./...
| Set GOROOT for x86 only | Set GOROOT for x86 only
| YAML | unknown | vlifesystems/rulehuntersrv,vlifesystems/rulehuntersrv,vlifesystems/rulehunter,vlifesystems/rulehunter | yaml | ## Code Before:
version: "{build}"
platform:
- x86
- x64
clone_folder: c:\gopath\src\github.com\vlifesystems\rulehuntersrv
environment:
GOPATH: c:\gopath
GOROOT: c:\go-x86
install:
- IF "%PLATFORM%" == "x86" set PATH=%GOROOT%\bin;C:\MinGW\bin;%PATH%
- IF "%PLATFORM%" == "x64" set PATH=C:\msys64\mingw64\bin;%PATH%
- echo %PATH%
- go version
- go env
- go get -v -t ./...
build_script:
- go install -v ./...
test_script:
- go test -v ./...
## Instruction:
Set GOROOT for x86 only
## Code After:
version: "{build}"
platform:
- x86
- x64
clone_folder: c:\gopath\src\github.com\vlifesystems\rulehuntersrv
environment:
GOPATH: c:\gopath
install:
- IF "%PLATFORM%" == "x86" (
set GOROOT=c:\go-x86
set PATH=%GOROOT%\bin;C:\MinGW\bin;%PATH%
)
- IF "%PLATFORM%" == "x64" set PATH=C:\msys64\mingw64\bin;%PATH%
- echo %PATH%
- go version
- go env
- go get -v -t ./...
build_script:
- go install -v ./...
test_script:
- go test -v ./...
| version: "{build}"
platform:
- x86
- x64
clone_folder: c:\gopath\src\github.com\vlifesystems\rulehuntersrv
environment:
GOPATH: c:\gopath
- GOROOT: c:\go-x86
install:
+ - IF "%PLATFORM%" == "x86" (
+ set GOROOT=c:\go-x86
- - IF "%PLATFORM%" == "x86" set PATH=%GOROOT%\bin;C:\MinGW\bin;%PATH%
? - -- ------------ --------
+ set PATH=%GOROOT%\bin;C:\MinGW\bin;%PATH%
+ )
- IF "%PLATFORM%" == "x64" set PATH=C:\msys64\mingw64\bin;%PATH%
- echo %PATH%
- go version
- go env
- go get -v -t ./...
build_script:
- go install -v ./...
test_script:
- go test -v ./... | 6 | 0.24 | 4 | 2 |
76fe16b4198f89b665798866b6b599c319c4e4e6 | solr/solr.xml | solr/solr.xml | <?xml version="1.0" encoding="UTF-8" ?>
<solr persistent="true">
<cores adminPath="/admin/cores">
<core name="UserContent" instanceDir="./" schema="schema_user.xml"/>
<core name="18thConnectLocalContent" instanceDir="./" schema="schema_user.xml"/>
<core name="NINESLocalContent" instanceDir="./" schema="schema_user.xml"/>
<core name="test18thConnectLocalContent" instanceDir="./" schema="schema_user.xml"/>
<core name="testNINESLocalContent" instanceDir="./" schema="schema_user.xml"/>
<core name="testResources" instanceDir="./"/>
<core name="resources" instanceDir="./"/>
<core name="archive_rossetti" instanceDir="./"/>
<core name="archive_cali" instanceDir="./"/>
<core name="archive_bancroft" instanceDir="./"/>
<core name="archive_ufldloc" instanceDir="./"/>
</cores>
</solr>
| <?xml version="1.0" encoding="UTF-8" ?>
<solr persistent="true">
<cores adminPath="/admin/cores">
<core name="UserContent" instanceDir="./" schema="schema_user.xml"/>
<core name="18thConnectLocalContent" instanceDir="./" schema="schema_user.xml"/>
<core name="NINESLocalContent" instanceDir="./" schema="schema_user.xml"/>
<core name="MESALocalContent" instanceDir="./" schema="schema_user.xml"/>
<core name="test18thConnectLocalContent" instanceDir="./" schema="schema_user.xml"/>
<core name="testNINESLocalContent" instanceDir="./" schema="schema_user.xml"/>
<core name="testMESALocalContent" instanceDir="./" schema="schema_user.xml"/>
<core name="testResources" instanceDir="./"/>
<core name="resources" instanceDir="./"/>
</cores>
</solr>
| Add MESA to the cores | Add MESA to the cores
| XML | apache-2.0 | treydock/solr,collex/solr,treydock/solr,treydock/solr | xml | ## Code Before:
<?xml version="1.0" encoding="UTF-8" ?>
<solr persistent="true">
<cores adminPath="/admin/cores">
<core name="UserContent" instanceDir="./" schema="schema_user.xml"/>
<core name="18thConnectLocalContent" instanceDir="./" schema="schema_user.xml"/>
<core name="NINESLocalContent" instanceDir="./" schema="schema_user.xml"/>
<core name="test18thConnectLocalContent" instanceDir="./" schema="schema_user.xml"/>
<core name="testNINESLocalContent" instanceDir="./" schema="schema_user.xml"/>
<core name="testResources" instanceDir="./"/>
<core name="resources" instanceDir="./"/>
<core name="archive_rossetti" instanceDir="./"/>
<core name="archive_cali" instanceDir="./"/>
<core name="archive_bancroft" instanceDir="./"/>
<core name="archive_ufldloc" instanceDir="./"/>
</cores>
</solr>
## Instruction:
Add MESA to the cores
## Code After:
<?xml version="1.0" encoding="UTF-8" ?>
<solr persistent="true">
<cores adminPath="/admin/cores">
<core name="UserContent" instanceDir="./" schema="schema_user.xml"/>
<core name="18thConnectLocalContent" instanceDir="./" schema="schema_user.xml"/>
<core name="NINESLocalContent" instanceDir="./" schema="schema_user.xml"/>
<core name="MESALocalContent" instanceDir="./" schema="schema_user.xml"/>
<core name="test18thConnectLocalContent" instanceDir="./" schema="schema_user.xml"/>
<core name="testNINESLocalContent" instanceDir="./" schema="schema_user.xml"/>
<core name="testMESALocalContent" instanceDir="./" schema="schema_user.xml"/>
<core name="testResources" instanceDir="./"/>
<core name="resources" instanceDir="./"/>
</cores>
</solr>
| <?xml version="1.0" encoding="UTF-8" ?>
<solr persistent="true">
<cores adminPath="/admin/cores">
<core name="UserContent" instanceDir="./" schema="schema_user.xml"/>
<core name="18thConnectLocalContent" instanceDir="./" schema="schema_user.xml"/>
<core name="NINESLocalContent" instanceDir="./" schema="schema_user.xml"/>
+ <core name="MESALocalContent" instanceDir="./" schema="schema_user.xml"/>
<core name="test18thConnectLocalContent" instanceDir="./" schema="schema_user.xml"/>
<core name="testNINESLocalContent" instanceDir="./" schema="schema_user.xml"/>
+ <core name="testMESALocalContent" instanceDir="./" schema="schema_user.xml"/>
<core name="testResources" instanceDir="./"/>
<core name="resources" instanceDir="./"/>
- <core name="archive_rossetti" instanceDir="./"/>
- <core name="archive_cali" instanceDir="./"/>
- <core name="archive_bancroft" instanceDir="./"/>
- <core name="archive_ufldloc" instanceDir="./"/>
</cores>
</solr> | 6 | 0.375 | 2 | 4 |
2973b664e8c9cf551d5d7277ab4995125be5fad0 | python/reference.py | python/reference.py | import os
# Current directory
# If you call this from the current directory without abspath,
# then it will not work since __file__ is a relative path
os.path.dirname(os.path.abspath(__file__))
| import os
# Current directory
# If you call this from the current directory without abspath,
# then it will not work since __file__ is a relative path
os.path.dirname(os.path.abspath(__file__))
# Get all files in a directory
# Never use os.walk again
def all_sub_files(root):
for path, subdirs, files in os.walk(root):
for name in files:
yield os.path.join(path, name)
| Add util for not using os.walk | Add util for not using os.walk
| Python | mit | brycepg/how-to | python | ## Code Before:
import os
# Current directory
# If you call this from the current directory without abspath,
# then it will not work since __file__ is a relative path
os.path.dirname(os.path.abspath(__file__))
## Instruction:
Add util for not using os.walk
## Code After:
import os
# Current directory
# If you call this from the current directory without abspath,
# then it will not work since __file__ is a relative path
os.path.dirname(os.path.abspath(__file__))
# Get all files in a directory
# Never use os.walk again
def all_sub_files(root):
for path, subdirs, files in os.walk(root):
for name in files:
yield os.path.join(path, name)
| import os
# Current directory
# If you call this from the current directory without abspath,
# then it will not work since __file__ is a relative path
os.path.dirname(os.path.abspath(__file__))
+
+ # Get all files in a directory
+ # Never use os.walk again
+ def all_sub_files(root):
+ for path, subdirs, files in os.walk(root):
+ for name in files:
+ yield os.path.join(path, name) | 7 | 1.4 | 7 | 0 |
edbd794b27206b3dab6ae8bb22df466cf7d009e0 | .travis.yml | .travis.yml | osx_image: xcode9.3
language: objective-c
# Travis is currently experiencing issues with Xcode builds: https://github.com/travis-ci/travis-ci/issues/6675#issuecomment-257964767
# When this is resolved, remove the entire `before_install` block, as well as the `travis_retry` command below.
before_install:
- export IOS_SIMULATOR_UDID=`instruments -s devices | grep "iPhone 7 (11.3" | awk -F '[ ]' '{print $4}' | awk -F '[\[]' '{print $2}' | sed 's/.$//'`
- echo $IOS_SIMULATOR_UDID
- open -a "simulator" --args -CurrentDeviceUDID $IOS_SIMULATOR_UDID
install:
- gem install xcpretty
script:
- set -o pipefail && travis_retry xcodebuild -project Bond.xcodeproj -scheme Bond-iOS -sdk iphonesimulator -destination 'platform=iOS Simulator,name=iPhone X,OS=11.3' test | xcpretty
- set -o pipefail && travis_retry xcodebuild -project Bond.xcodeproj -scheme Bond-macOS test | xcpretty
- swift test
after_success:
- bash <(curl -s https://codecov.io/bash)
| osx_image: xcode9.3
language: objective-c
# Travis is currently experiencing issues with Xcode builds: https://github.com/travis-ci/travis-ci/issues/6675#issuecomment-257964767
# When this is resolved, remove the entire `before_install` block, as well as the `travis_retry` command below.
before_install:
- export IOS_SIMULATOR_UDID=`instruments -s devices | grep "iPhone 7 (11.3" | awk -F '[ ]' '{print $4}' | awk -F '[\[]' '{print $2}' | sed 's/.$//'`
- echo $IOS_SIMULATOR_UDID
- open -a "simulator" --args -CurrentDeviceUDID $IOS_SIMULATOR_UDID
install:
- gem install xcpretty
script:
- set -o pipefail && travis_retry xcodebuild -project Bond.xcodeproj -scheme Bond-iOS -sdk iphonesimulator -destination 'platform=iOS Simulator,name=iPhone X,OS=11.3' test | xcpretty
- swift test
after_success:
- bash <(curl -s https://codecov.io/bash)
| Disable testing the macOS target on TravisCI | Disable testing the macOS target on TravisCI
| YAML | mit | SwiftBond/Bond,ReactiveKit/Bond,tonyarnold/Bond,SwiftBond/Bond,ReactiveKit/Bond,tonyarnold/Bond | yaml | ## Code Before:
osx_image: xcode9.3
language: objective-c
# Travis is currently experiencing issues with Xcode builds: https://github.com/travis-ci/travis-ci/issues/6675#issuecomment-257964767
# When this is resolved, remove the entire `before_install` block, as well as the `travis_retry` command below.
before_install:
- export IOS_SIMULATOR_UDID=`instruments -s devices | grep "iPhone 7 (11.3" | awk -F '[ ]' '{print $4}' | awk -F '[\[]' '{print $2}' | sed 's/.$//'`
- echo $IOS_SIMULATOR_UDID
- open -a "simulator" --args -CurrentDeviceUDID $IOS_SIMULATOR_UDID
install:
- gem install xcpretty
script:
- set -o pipefail && travis_retry xcodebuild -project Bond.xcodeproj -scheme Bond-iOS -sdk iphonesimulator -destination 'platform=iOS Simulator,name=iPhone X,OS=11.3' test | xcpretty
- set -o pipefail && travis_retry xcodebuild -project Bond.xcodeproj -scheme Bond-macOS test | xcpretty
- swift test
after_success:
- bash <(curl -s https://codecov.io/bash)
## Instruction:
Disable testing the macOS target on TravisCI
## Code After:
osx_image: xcode9.3
language: objective-c
# Travis is currently experiencing issues with Xcode builds: https://github.com/travis-ci/travis-ci/issues/6675#issuecomment-257964767
# When this is resolved, remove the entire `before_install` block, as well as the `travis_retry` command below.
before_install:
- export IOS_SIMULATOR_UDID=`instruments -s devices | grep "iPhone 7 (11.3" | awk -F '[ ]' '{print $4}' | awk -F '[\[]' '{print $2}' | sed 's/.$//'`
- echo $IOS_SIMULATOR_UDID
- open -a "simulator" --args -CurrentDeviceUDID $IOS_SIMULATOR_UDID
install:
- gem install xcpretty
script:
- set -o pipefail && travis_retry xcodebuild -project Bond.xcodeproj -scheme Bond-iOS -sdk iphonesimulator -destination 'platform=iOS Simulator,name=iPhone X,OS=11.3' test | xcpretty
- swift test
after_success:
- bash <(curl -s https://codecov.io/bash)
| osx_image: xcode9.3
language: objective-c
# Travis is currently experiencing issues with Xcode builds: https://github.com/travis-ci/travis-ci/issues/6675#issuecomment-257964767
# When this is resolved, remove the entire `before_install` block, as well as the `travis_retry` command below.
before_install:
- export IOS_SIMULATOR_UDID=`instruments -s devices | grep "iPhone 7 (11.3" | awk -F '[ ]' '{print $4}' | awk -F '[\[]' '{print $2}' | sed 's/.$//'`
- echo $IOS_SIMULATOR_UDID
- open -a "simulator" --args -CurrentDeviceUDID $IOS_SIMULATOR_UDID
install:
- gem install xcpretty
script:
- set -o pipefail && travis_retry xcodebuild -project Bond.xcodeproj -scheme Bond-iOS -sdk iphonesimulator -destination 'platform=iOS Simulator,name=iPhone X,OS=11.3' test | xcpretty
- - set -o pipefail && travis_retry xcodebuild -project Bond.xcodeproj -scheme Bond-macOS test | xcpretty
- swift test
after_success:
- bash <(curl -s https://codecov.io/bash) | 1 | 0.05 | 0 | 1 |
6801fd493432edb53da771730f28591d921e7d97 | mrbgem.rake | mrbgem.rake | MRuby::Gem::Specification.new('mruby-maxminddb') do |spec|
spec.license = 'MIT'
spec.authors = 'Kenichi Mitsugi'
spec.add_test_dependency 'mruby-io', :mgem => 'mruby-io'
def maxminddb_origin_path
if File.exist?("../test/fixtures/GeoLite2-City.mmdb.gz")
"#{build.build_dir}/../../../test/fixtures/GeoLite2-City.mmdb"
else
"#{build.build_dir}/../mrbgems/mruby-maxminddb/test/fixtures/GeoLite2-City.mmdb"
end
end
maxminddb_ci_path = "#{build.build_dir}/../maxminddb-fixtures/GeoLite2-City.mmdb"
sh "mkdir -p #{File.dirname maxminddb_ci_path}"
sh "gzip -cd #{maxminddb_origin_path}.gz > #{maxminddb_ci_path}" unless File.exist?(maxminddb_ci_path)
end
| MRuby::Gem::Specification.new('mruby-maxminddb') do |spec|
spec.license = 'MIT'
spec.authors = 'Kenichi Mitsugi'
spec.add_test_dependency 'mruby-io', :mgem => 'mruby-io'
def maxminddb_origin_path
"#{dir}/test/fixtures/GeoLite2-City.mmdb"
end
maxminddb_ci_path = "#{build.build_dir}/../maxminddb-fixtures/GeoLite2-City.mmdb"
sh "mkdir -p #{File.dirname maxminddb_ci_path}"
sh "gzip -cd #{maxminddb_origin_path}.gz > #{maxminddb_ci_path}" unless File.exist?(maxminddb_ci_path)
end
| Use `dir` method intead to get fixture path. | Use `dir` method intead to get fixture path.
| Ruby | mit | happysiro/mruby-maxminddb,happysiro/mruby-maxminddb | ruby | ## Code Before:
MRuby::Gem::Specification.new('mruby-maxminddb') do |spec|
spec.license = 'MIT'
spec.authors = 'Kenichi Mitsugi'
spec.add_test_dependency 'mruby-io', :mgem => 'mruby-io'
def maxminddb_origin_path
if File.exist?("../test/fixtures/GeoLite2-City.mmdb.gz")
"#{build.build_dir}/../../../test/fixtures/GeoLite2-City.mmdb"
else
"#{build.build_dir}/../mrbgems/mruby-maxminddb/test/fixtures/GeoLite2-City.mmdb"
end
end
maxminddb_ci_path = "#{build.build_dir}/../maxminddb-fixtures/GeoLite2-City.mmdb"
sh "mkdir -p #{File.dirname maxminddb_ci_path}"
sh "gzip -cd #{maxminddb_origin_path}.gz > #{maxminddb_ci_path}" unless File.exist?(maxminddb_ci_path)
end
## Instruction:
Use `dir` method intead to get fixture path.
## Code After:
MRuby::Gem::Specification.new('mruby-maxminddb') do |spec|
spec.license = 'MIT'
spec.authors = 'Kenichi Mitsugi'
spec.add_test_dependency 'mruby-io', :mgem => 'mruby-io'
def maxminddb_origin_path
"#{dir}/test/fixtures/GeoLite2-City.mmdb"
end
maxminddb_ci_path = "#{build.build_dir}/../maxminddb-fixtures/GeoLite2-City.mmdb"
sh "mkdir -p #{File.dirname maxminddb_ci_path}"
sh "gzip -cd #{maxminddb_origin_path}.gz > #{maxminddb_ci_path}" unless File.exist?(maxminddb_ci_path)
end
| MRuby::Gem::Specification.new('mruby-maxminddb') do |spec|
spec.license = 'MIT'
spec.authors = 'Kenichi Mitsugi'
spec.add_test_dependency 'mruby-io', :mgem => 'mruby-io'
def maxminddb_origin_path
- if File.exist?("../test/fixtures/GeoLite2-City.mmdb.gz")
- "#{build.build_dir}/../../../test/fixtures/GeoLite2-City.mmdb"
? -- ------------ ---------
+ "#{dir}/test/fixtures/GeoLite2-City.mmdb"
- else
- "#{build.build_dir}/../mrbgems/mruby-maxminddb/test/fixtures/GeoLite2-City.mmdb"
- end
end
maxminddb_ci_path = "#{build.build_dir}/../maxminddb-fixtures/GeoLite2-City.mmdb"
sh "mkdir -p #{File.dirname maxminddb_ci_path}"
sh "gzip -cd #{maxminddb_origin_path}.gz > #{maxminddb_ci_path}" unless File.exist?(maxminddb_ci_path)
end | 6 | 0.333333 | 1 | 5 |
75b92b91ddbbdec4a67b2f7019001a31e250cb8a | resources/css/nephrowiki.less | resources/css/nephrowiki.less | @import "../bootstrap/less/bootstrap.less";
@import "../bootstrap/less/theme.less";
@import "mw-overrides.less";
// Override Bootstrap variables
@font-family-sans-serif: "Segoe UI", "Droid Sans", "Helvetica Neue", Helvetica, Arial, sans-serif;
@font-family-serif: Georgia, "Droid Serif", "Times New Roman", Times, serif;
@font-family-monospace: Menlo, Monaco, Consolas, "Droid Mono", "Courier New", monospace;
@font-size-base: 16px;
@font-size-h1: floor((@font-size-base * 2.0));
@font-size-h2: floor((@font-size-base * 1.4));
@font-size-h3: floor((@font-size-base * 1.2));
@font-size-h4: floor((@font-size-base * 1.0));
@heading-color: @brand-primary;
// Additional style definitions
body {
margin-top: 50px; // accomodate fixed-top navbar
}
.footer {
margin-top: @padding-large-vertical * 2;
}
| @import "../bootstrap/less/bootstrap.less";
@import "../bootstrap/less/theme.less";
@import "mw-overrides.less";
// Override Bootstrap variables
@font-family-sans-serif: "Segoe UI", "Droid Sans", "Helvetica Neue", Helvetica, Arial, sans-serif;
@font-family-serif: Georgia, "Droid Serif", "Times New Roman", Times, serif;
@font-family-monospace: Menlo, Monaco, Consolas, "Droid Mono", "Courier New", monospace;
@font-size-base: 16px;
@font-size-h1: floor((@font-size-base * 2.0));
@font-size-h2: floor((@font-size-base * 1.4));
@font-size-h3: floor((@font-size-base * 1.2));
@font-size-h4: floor((@font-size-base * 1.0));
@headings-color: @brand-primary;
@line-height-base: 1.5;
// Additional style definitions
body {
margin-top: 50px; // accomodate fixed-top navbar
}
.footer {
margin-top: @padding-large-vertical * 2;
}
| Fix typo, increase line height. | Fix typo, increase line height.
| Less | mit | bovender/NephroWikiSkin,bovender/NephroWikiSkin | less | ## Code Before:
@import "../bootstrap/less/bootstrap.less";
@import "../bootstrap/less/theme.less";
@import "mw-overrides.less";
// Override Bootstrap variables
@font-family-sans-serif: "Segoe UI", "Droid Sans", "Helvetica Neue", Helvetica, Arial, sans-serif;
@font-family-serif: Georgia, "Droid Serif", "Times New Roman", Times, serif;
@font-family-monospace: Menlo, Monaco, Consolas, "Droid Mono", "Courier New", monospace;
@font-size-base: 16px;
@font-size-h1: floor((@font-size-base * 2.0));
@font-size-h2: floor((@font-size-base * 1.4));
@font-size-h3: floor((@font-size-base * 1.2));
@font-size-h4: floor((@font-size-base * 1.0));
@heading-color: @brand-primary;
// Additional style definitions
body {
margin-top: 50px; // accomodate fixed-top navbar
}
.footer {
margin-top: @padding-large-vertical * 2;
}
## Instruction:
Fix typo, increase line height.
## Code After:
@import "../bootstrap/less/bootstrap.less";
@import "../bootstrap/less/theme.less";
@import "mw-overrides.less";
// Override Bootstrap variables
@font-family-sans-serif: "Segoe UI", "Droid Sans", "Helvetica Neue", Helvetica, Arial, sans-serif;
@font-family-serif: Georgia, "Droid Serif", "Times New Roman", Times, serif;
@font-family-monospace: Menlo, Monaco, Consolas, "Droid Mono", "Courier New", monospace;
@font-size-base: 16px;
@font-size-h1: floor((@font-size-base * 2.0));
@font-size-h2: floor((@font-size-base * 1.4));
@font-size-h3: floor((@font-size-base * 1.2));
@font-size-h4: floor((@font-size-base * 1.0));
@headings-color: @brand-primary;
@line-height-base: 1.5;
// Additional style definitions
body {
margin-top: 50px; // accomodate fixed-top navbar
}
.footer {
margin-top: @padding-large-vertical * 2;
}
| @import "../bootstrap/less/bootstrap.less";
@import "../bootstrap/less/theme.less";
@import "mw-overrides.less";
// Override Bootstrap variables
@font-family-sans-serif: "Segoe UI", "Droid Sans", "Helvetica Neue", Helvetica, Arial, sans-serif;
@font-family-serif: Georgia, "Droid Serif", "Times New Roman", Times, serif;
@font-family-monospace: Menlo, Monaco, Consolas, "Droid Mono", "Courier New", monospace;
@font-size-base: 16px;
@font-size-h1: floor((@font-size-base * 2.0));
@font-size-h2: floor((@font-size-base * 1.4));
@font-size-h3: floor((@font-size-base * 1.2));
@font-size-h4: floor((@font-size-base * 1.0));
- @heading-color: @brand-primary;
+ @headings-color: @brand-primary;
? +
+ @line-height-base: 1.5;
// Additional style definitions
body {
margin-top: 50px; // accomodate fixed-top navbar
}
.footer {
margin-top: @padding-large-vertical * 2;
} | 3 | 0.111111 | 2 | 1 |
4bb4ce2f61143a702d14c317d62066b3af57fafd | www/templates/link_top.html | www/templates/link_top.html | <div id="toplink" class="col-sm-12">
<div class="col-lg-6 imgbox">
<img src='{{ self.thumb_url }}' alt="{{ self.title }}"/>
</div>
<div class="col-lg-6 textbox">
<div class="title-header"><a href="{{ globals.base_url }}top_active">{% trans _('Destacada') %}</a></div>
<h1><a href="{% exec self.get_permalink %}">{{ self.title }}</a></h1>
{% if not globals.mobile %}
{% if self.thumb_url %}
<p class="news-content">{% exec text_to_summary self.content 600 %}</p>
{% else %}
<p class="news-content">{% exec text_to_summary self.content 200 %}</p>
{% endif %}
{% endif %}
</div>
</div> <!-- toplink -->
<div class="clearfix visible-sm-block"></div>
| <div id="toplink" class="col-sm-12">
<div class="col-lg-6 imgbox">
<img src='{{ self.thumb_url }}' alt="{{ self.title }}"/>
</div>
<div class="col-lg-6 textbox">
<div class="title-header"><a href="{{ globals.base_url }}top_active">{% trans _('Destacada') %}</a></div>
<h1><a href="{% exec self.get_permalink %}">{{ self.title }}</a></h1>
{% if not globals.mobile %}
{% if self.thumb_url %}
<a href="{{ self.relative_permalink }}" class="content-link" title="más información"><p class="news-content">{% exec text_to_summary self.content 600 %}</p></a>
{% else %}
<a href="{{ self.relative_permalink }}" class="content-link" title="más información"><p class="news-content">{% exec text_to_summary self.content 200 %}</p></a>
{% endif %}
{% endif %}
</div>
</div> <!-- toplink -->
<div class="clearfix visible-sm-block"></div>
| Add link in story text in top link | Add link in story text in top link
| HTML | agpl-3.0 | consultiacl/copuchalo,consultiacl/copuchalo,consultiaio/mediatize,consultiaio/mediatize,consultiaio/mediatize,consultiacl/copuchalo,consultiaio/mediatize,consultiacl/copuchalo,consultiacl/copuchalo,consultiaio/mediatize,consultiaio/mediatize,consultiacl/copuchalo | html | ## Code Before:
<div id="toplink" class="col-sm-12">
<div class="col-lg-6 imgbox">
<img src='{{ self.thumb_url }}' alt="{{ self.title }}"/>
</div>
<div class="col-lg-6 textbox">
<div class="title-header"><a href="{{ globals.base_url }}top_active">{% trans _('Destacada') %}</a></div>
<h1><a href="{% exec self.get_permalink %}">{{ self.title }}</a></h1>
{% if not globals.mobile %}
{% if self.thumb_url %}
<p class="news-content">{% exec text_to_summary self.content 600 %}</p>
{% else %}
<p class="news-content">{% exec text_to_summary self.content 200 %}</p>
{% endif %}
{% endif %}
</div>
</div> <!-- toplink -->
<div class="clearfix visible-sm-block"></div>
## Instruction:
Add link in story text in top link
## Code After:
<div id="toplink" class="col-sm-12">
<div class="col-lg-6 imgbox">
<img src='{{ self.thumb_url }}' alt="{{ self.title }}"/>
</div>
<div class="col-lg-6 textbox">
<div class="title-header"><a href="{{ globals.base_url }}top_active">{% trans _('Destacada') %}</a></div>
<h1><a href="{% exec self.get_permalink %}">{{ self.title }}</a></h1>
{% if not globals.mobile %}
{% if self.thumb_url %}
<a href="{{ self.relative_permalink }}" class="content-link" title="más información"><p class="news-content">{% exec text_to_summary self.content 600 %}</p></a>
{% else %}
<a href="{{ self.relative_permalink }}" class="content-link" title="más información"><p class="news-content">{% exec text_to_summary self.content 200 %}</p></a>
{% endif %}
{% endif %}
</div>
</div> <!-- toplink -->
<div class="clearfix visible-sm-block"></div>
| <div id="toplink" class="col-sm-12">
<div class="col-lg-6 imgbox">
<img src='{{ self.thumb_url }}' alt="{{ self.title }}"/>
</div>
<div class="col-lg-6 textbox">
<div class="title-header"><a href="{{ globals.base_url }}top_active">{% trans _('Destacada') %}</a></div>
<h1><a href="{% exec self.get_permalink %}">{{ self.title }}</a></h1>
{% if not globals.mobile %}
{% if self.thumb_url %}
- <p class="news-content">{% exec text_to_summary self.content 600 %}</p>
+ <a href="{{ self.relative_permalink }}" class="content-link" title="más información"><p class="news-content">{% exec text_to_summary self.content 600 %}</p></a>
{% else %}
- <p class="news-content">{% exec text_to_summary self.content 200 %}</p>
+ <a href="{{ self.relative_permalink }}" class="content-link" title="más información"><p class="news-content">{% exec text_to_summary self.content 200 %}</p></a>
{% endif %}
{% endif %}
</div>
</div> <!-- toplink -->
<div class="clearfix visible-sm-block"></div>
| 4 | 0.166667 | 2 | 2 |
f81e9472c05197da73a9f4caae7635e57aa460e2 | spec/features/main_spec.rb | spec/features/main_spec.rb | describe 'main process', type: :feature, js: true do
it "html includes content 'Listing buckets' and 'my-bucket'" do
Aws.config[:s3] = {
stub_responses: {
list_buckets: { buckets: [{ name: 'my-bucket' }] }
}
}
visit '/buckets'
expect(page).to have_content 'Listing buckets'
expect(page).to have_content 'my-bucket'
click_on 'my-bucket'
click_on 'Back'
end
end
| describe 'main process', type: :feature, js: true do
before do
FakeS3Server.restart
s3 = Aws::S3::Client.new
s3.create_bucket(bucket: 'my-bucket')
Tempfile.open('file') do |file|
file.puts 'body'
file.flush
s3.put_object(
bucket: 'my-bucket',
key: 'my-folder/my-file',
body: file
)
end
end
it "html includes content 'Listing buckets' and 'my-bucket'" do
visit '/buckets'
expect(page).to have_content 'Listing buckets'
expect(page).to have_content 'my-bucket'
click_on 'my-bucket'
click_on 'Back'
end
end
| Modify feature spec to use FakeS3Server | Modify feature spec to use FakeS3Server
| Ruby | mit | hoshinotsuyoshi/s3_explorer,hoshinotsuyoshi/s3_explorer,hoshinotsuyoshi/s3_explorer | ruby | ## Code Before:
describe 'main process', type: :feature, js: true do
it "html includes content 'Listing buckets' and 'my-bucket'" do
Aws.config[:s3] = {
stub_responses: {
list_buckets: { buckets: [{ name: 'my-bucket' }] }
}
}
visit '/buckets'
expect(page).to have_content 'Listing buckets'
expect(page).to have_content 'my-bucket'
click_on 'my-bucket'
click_on 'Back'
end
end
## Instruction:
Modify feature spec to use FakeS3Server
## Code After:
describe 'main process', type: :feature, js: true do
before do
FakeS3Server.restart
s3 = Aws::S3::Client.new
s3.create_bucket(bucket: 'my-bucket')
Tempfile.open('file') do |file|
file.puts 'body'
file.flush
s3.put_object(
bucket: 'my-bucket',
key: 'my-folder/my-file',
body: file
)
end
end
it "html includes content 'Listing buckets' and 'my-bucket'" do
visit '/buckets'
expect(page).to have_content 'Listing buckets'
expect(page).to have_content 'my-bucket'
click_on 'my-bucket'
click_on 'Back'
end
end
| describe 'main process', type: :feature, js: true do
+ before do
+ FakeS3Server.restart
+ s3 = Aws::S3::Client.new
+ s3.create_bucket(bucket: 'my-bucket')
+ Tempfile.open('file') do |file|
+ file.puts 'body'
+ file.flush
+ s3.put_object(
+ bucket: 'my-bucket',
+ key: 'my-folder/my-file',
+ body: file
+ )
+ end
+ end
+
it "html includes content 'Listing buckets' and 'my-bucket'" do
- Aws.config[:s3] = {
- stub_responses: {
- list_buckets: { buckets: [{ name: 'my-bucket' }] }
- }
- }
visit '/buckets'
expect(page).to have_content 'Listing buckets'
expect(page).to have_content 'my-bucket'
click_on 'my-bucket'
click_on 'Back'
end
end | 20 | 1.25 | 15 | 5 |
4dc72607700a27037fe013c7f15856d9e4ff5d62 | README.md | README.md | Esta es una aplicacion que muestra las casillas que amenazan las piezas del caballo y la reina en un tablero de ajedrez
| Esta es una aplicacion que muestra las casillas que amenazan las piezas del caballo y la reina en un tablero de ajedrez.
Imagen de fondo para el tablero tomada de OpenGameArt.org:
http://opengameart.org/sites/default/files/styles/medium/public/wood_pre.png
| Add source of image to Readme | Add source of image to Readme | Markdown | mit | hguilla/casillas-amenazadas,hguilla/casillas-amenazadas | markdown | ## Code Before:
Esta es una aplicacion que muestra las casillas que amenazan las piezas del caballo y la reina en un tablero de ajedrez
## Instruction:
Add source of image to Readme
## Code After:
Esta es una aplicacion que muestra las casillas que amenazan las piezas del caballo y la reina en un tablero de ajedrez.
Imagen de fondo para el tablero tomada de OpenGameArt.org:
http://opengameart.org/sites/default/files/styles/medium/public/wood_pre.png
| - Esta es una aplicacion que muestra las casillas que amenazan las piezas del caballo y la reina en un tablero de ajedrez
+ Esta es una aplicacion que muestra las casillas que amenazan las piezas del caballo y la reina en un tablero de ajedrez.
? +
+
+ Imagen de fondo para el tablero tomada de OpenGameArt.org:
+ http://opengameart.org/sites/default/files/styles/medium/public/wood_pre.png | 5 | 5 | 4 | 1 |
62956a39fcbd1ae047451457b786b8fc3bfd0462 | test/unit/test_transparent_pdfs.rb | test/unit/test_transparent_pdfs.rb | here = File.expand_path(File.dirname(__FILE__))
require File.join(here, '..', 'test_helper')
require 'tmpdir'
class TransparentPDFsTest < Test::Unit::TestCase
def setup
@klass = Class.new
@klass.send(:include, Docsplit::TransparentPDFs)
@detector = @klass.new
end
def test_files_with_pdf_extension_are_always_considered_a_pdf
pdfs = Dir.glob('test/fixtures/with_pdf_extension/*.pdf').select { |path| File.file?(path) }
assert pdfs.any?, 'ensure pdfs with extensions are available to test with'
pdfs.each do |pdf|
assert @detector.is_pdf?(pdf), "#{pdf} with '.pdf' extension is identified as a PDF (regardless of its file contents)"
end
end
def test_pdfs_without_the_pdf_file_extension_is_considerd_a_pdf
pdfs = Dir.glob('test/fixtures/without_pdf_extension/*/*').select { |path| File.file?(path) }
assert pdfs.any?, 'ensure pdfs without extensions are available to test with'
pdfs.each do |pdf|
assert @detector.is_pdf?(pdf), "#{pdf} with '.pdf' extension is identified as a PDF"
end
end
end
| here = File.expand_path(File.dirname(__FILE__))
require File.join(here, '..', 'test_helper')
require 'tmpdir'
class TransparentPDFsTest < Minitest::Test
def setup
@klass = Class.new
@klass.send(:include, Docsplit::TransparentPDFs)
@detector = @klass.new
end
def test_files_with_pdf_extension_are_always_considered_a_pdf
pdfs = Dir.glob('test/fixtures/with_pdf_extension/*.pdf').select { |path| File.file?(path) }
assert pdfs.any?, 'ensure pdfs with extensions are available to test with'
pdfs.each do |pdf|
assert @detector.is_pdf?(pdf), "#{pdf} with '.pdf' extension is identified as a PDF (regardless of its file contents)"
end
end
def test_pdfs_without_the_pdf_file_extension_is_considerd_a_pdf
pdfs = Dir.glob('test/fixtures/without_pdf_extension/*/*').select { |path| File.file?(path) }
assert pdfs.any?, 'ensure pdfs without extensions are available to test with'
pdfs.each do |pdf|
assert @detector.is_pdf?(pdf), "#{pdf} with '.pdf' extension is identified as a PDF"
end
end
end
| Update merged test to minitest | Update merged test to minitest | Ruby | mit | jurek7/docsplit,jurek7/docsplit | ruby | ## Code Before:
here = File.expand_path(File.dirname(__FILE__))
require File.join(here, '..', 'test_helper')
require 'tmpdir'
class TransparentPDFsTest < Test::Unit::TestCase
def setup
@klass = Class.new
@klass.send(:include, Docsplit::TransparentPDFs)
@detector = @klass.new
end
def test_files_with_pdf_extension_are_always_considered_a_pdf
pdfs = Dir.glob('test/fixtures/with_pdf_extension/*.pdf').select { |path| File.file?(path) }
assert pdfs.any?, 'ensure pdfs with extensions are available to test with'
pdfs.each do |pdf|
assert @detector.is_pdf?(pdf), "#{pdf} with '.pdf' extension is identified as a PDF (regardless of its file contents)"
end
end
def test_pdfs_without_the_pdf_file_extension_is_considerd_a_pdf
pdfs = Dir.glob('test/fixtures/without_pdf_extension/*/*').select { |path| File.file?(path) }
assert pdfs.any?, 'ensure pdfs without extensions are available to test with'
pdfs.each do |pdf|
assert @detector.is_pdf?(pdf), "#{pdf} with '.pdf' extension is identified as a PDF"
end
end
end
## Instruction:
Update merged test to minitest
## Code After:
here = File.expand_path(File.dirname(__FILE__))
require File.join(here, '..', 'test_helper')
require 'tmpdir'
class TransparentPDFsTest < Minitest::Test
def setup
@klass = Class.new
@klass.send(:include, Docsplit::TransparentPDFs)
@detector = @klass.new
end
def test_files_with_pdf_extension_are_always_considered_a_pdf
pdfs = Dir.glob('test/fixtures/with_pdf_extension/*.pdf').select { |path| File.file?(path) }
assert pdfs.any?, 'ensure pdfs with extensions are available to test with'
pdfs.each do |pdf|
assert @detector.is_pdf?(pdf), "#{pdf} with '.pdf' extension is identified as a PDF (regardless of its file contents)"
end
end
def test_pdfs_without_the_pdf_file_extension_is_considerd_a_pdf
pdfs = Dir.glob('test/fixtures/without_pdf_extension/*/*').select { |path| File.file?(path) }
assert pdfs.any?, 'ensure pdfs without extensions are available to test with'
pdfs.each do |pdf|
assert @detector.is_pdf?(pdf), "#{pdf} with '.pdf' extension is identified as a PDF"
end
end
end
| here = File.expand_path(File.dirname(__FILE__))
require File.join(here, '..', 'test_helper')
require 'tmpdir'
- class TransparentPDFsTest < Test::Unit::TestCase
? ^ ------ ----
+ class TransparentPDFsTest < Minitest::Test
? ^^^^^
def setup
@klass = Class.new
@klass.send(:include, Docsplit::TransparentPDFs)
@detector = @klass.new
end
def test_files_with_pdf_extension_are_always_considered_a_pdf
pdfs = Dir.glob('test/fixtures/with_pdf_extension/*.pdf').select { |path| File.file?(path) }
assert pdfs.any?, 'ensure pdfs with extensions are available to test with'
pdfs.each do |pdf|
assert @detector.is_pdf?(pdf), "#{pdf} with '.pdf' extension is identified as a PDF (regardless of its file contents)"
end
end
def test_pdfs_without_the_pdf_file_extension_is_considerd_a_pdf
pdfs = Dir.glob('test/fixtures/without_pdf_extension/*/*').select { |path| File.file?(path) }
assert pdfs.any?, 'ensure pdfs without extensions are available to test with'
pdfs.each do |pdf|
assert @detector.is_pdf?(pdf), "#{pdf} with '.pdf' extension is identified as a PDF"
end
end
end | 2 | 0.068966 | 1 | 1 |
531e60a00aae26a9f8d135ade309719812787f0c | index.php | index.php | <?php
/**
* The main template file.
*
*/
get_header(); ?>
<div class="page-content">
<div class="wrapper">
<div style="float: left; width: 200px;">
<?php get_sidebar(); ?>
</div>
<div style="float: right; width: 600px; text-align: left;">
<script src="//platform.linkedin.com/in.js" type="text/javascript">
/* <![CDATA[ */
lang: en_US
/* ]]> */
</script>
<!-- THIS IS WHERE THE WORDPRESS CODE TO INCLUDE CONTENT GOES...! -->
<?php
if ( is_page( 'Home' ) ) {
get_template_part( 'latest-solicitations', get_post_format() );
}
?>
<?php
if(have_posts()) :
while(have_posts()) :
?>
<p>
<?php
the_post(); ?>
<div class="alert alert-info">
<?php edit_post_link( 'Edit this page ', '', '' ); ?> · <a href="<?php _e(get_delete_post_link()); ?>">Delete this page</a>
</div>
<?php
the_content();
endwhile;
endif;
?>
</div>
</div>
</div>
| <?php
/**
* The main template file.
*
*/
get_header(); ?>
<div class="page-content">
<div class="wrapper">
<div style="float: left; width: 200px;">
<?php get_sidebar(); ?>
</div>
<div style="float: right; width: 600px; text-align: left;">
<script src="//platform.linkedin.com/in.js" type="text/javascript">
/* <![CDATA[ */
lang: en_US
/* ]]> */
</script>
<!-- THIS IS WHERE THE WORDPRESS CODE TO INCLUDE CONTENT GOES...! -->
<?php
if ( is_page( 'Home' ) ) {
get_template_part( 'latest-solicitations', get_post_format() );
}
?>
<?php
if(have_posts()) :
while(have_posts()) :
?>
<p>
<?php
the_post(); ?>
<div class="alert alert-info">
<?php edit_post_link( 'Edit this page ', '', '' ); ?> · <a href="<?php _e(get_delete_post_link()); ?>">Delete this page</a>
</div>
<?php
the_content();
endwhile;
endif;
?>
</div>
</div>
</div>
<?php do_shortcode('[google-translator]'); get_footer();
| Include shortcode for Google Translate plugin. | Include shortcode for Google Translate plugin.
| PHP | mit | codeforamerica/atlanta-procurement-wp | php | ## Code Before:
<?php
/**
* The main template file.
*
*/
get_header(); ?>
<div class="page-content">
<div class="wrapper">
<div style="float: left; width: 200px;">
<?php get_sidebar(); ?>
</div>
<div style="float: right; width: 600px; text-align: left;">
<script src="//platform.linkedin.com/in.js" type="text/javascript">
/* <![CDATA[ */
lang: en_US
/* ]]> */
</script>
<!-- THIS IS WHERE THE WORDPRESS CODE TO INCLUDE CONTENT GOES...! -->
<?php
if ( is_page( 'Home' ) ) {
get_template_part( 'latest-solicitations', get_post_format() );
}
?>
<?php
if(have_posts()) :
while(have_posts()) :
?>
<p>
<?php
the_post(); ?>
<div class="alert alert-info">
<?php edit_post_link( 'Edit this page ', '', '' ); ?> · <a href="<?php _e(get_delete_post_link()); ?>">Delete this page</a>
</div>
<?php
the_content();
endwhile;
endif;
?>
</div>
</div>
</div>
## Instruction:
Include shortcode for Google Translate plugin.
## Code After:
<?php
/**
* The main template file.
*
*/
get_header(); ?>
<div class="page-content">
<div class="wrapper">
<div style="float: left; width: 200px;">
<?php get_sidebar(); ?>
</div>
<div style="float: right; width: 600px; text-align: left;">
<script src="//platform.linkedin.com/in.js" type="text/javascript">
/* <![CDATA[ */
lang: en_US
/* ]]> */
</script>
<!-- THIS IS WHERE THE WORDPRESS CODE TO INCLUDE CONTENT GOES...! -->
<?php
if ( is_page( 'Home' ) ) {
get_template_part( 'latest-solicitations', get_post_format() );
}
?>
<?php
if(have_posts()) :
while(have_posts()) :
?>
<p>
<?php
the_post(); ?>
<div class="alert alert-info">
<?php edit_post_link( 'Edit this page ', '', '' ); ?> · <a href="<?php _e(get_delete_post_link()); ?>">Delete this page</a>
</div>
<?php
the_content();
endwhile;
endif;
?>
</div>
</div>
</div>
<?php do_shortcode('[google-translator]'); get_footer();
| <?php
/**
* The main template file.
*
*/
get_header(); ?>
<div class="page-content">
<div class="wrapper">
<div style="float: left; width: 200px;">
<?php get_sidebar(); ?>
</div>
<div style="float: right; width: 600px; text-align: left;">
<script src="//platform.linkedin.com/in.js" type="text/javascript">
/* <![CDATA[ */
lang: en_US
/* ]]> */
</script>
<!-- THIS IS WHERE THE WORDPRESS CODE TO INCLUDE CONTENT GOES...! -->
<?php
if ( is_page( 'Home' ) ) {
get_template_part( 'latest-solicitations', get_post_format() );
}
?>
<?php
if(have_posts()) :
while(have_posts()) :
?>
<p>
<?php
the_post(); ?>
<div class="alert alert-info">
<?php edit_post_link( 'Edit this page ', '', '' ); ?> · <a href="<?php _e(get_delete_post_link()); ?>">Delete this page</a>
</div>
<?php
the_content();
endwhile;
endif;
?>
</div>
</div>
</div>
+ <?php do_shortcode('[google-translator]'); get_footer(); | 1 | 0.021277 | 1 | 0 |
7717da9f8251cd021bc30b1fd7f08b1eada42252 | src/main.css | src/main.css | html, body {
touch-action: none;
padding: 0;
margin: 0;
overflow: hidden;
width: 100%;
height: 100%;
}
body {
display: flex;
align-items: center;
justify-content: center;
background-color: #181818;
}
canvas {
background-color: #86C1B9;
}
#assets {
display: none;
}
| html, body {
touch-action: none;
padding: 0;
margin: 0;
overflow: hidden;
width: 100%;
height: 100%;
}
body {
text-align: center;
background-color: #181818;
}
canvas {
background-color: #86C1B9;
}
#assets {
display: none;
}
| Use text-align to center the canvas. | Use text-align to center the canvas.
| CSS | mit | flannelhead/pasi | css | ## Code Before:
html, body {
touch-action: none;
padding: 0;
margin: 0;
overflow: hidden;
width: 100%;
height: 100%;
}
body {
display: flex;
align-items: center;
justify-content: center;
background-color: #181818;
}
canvas {
background-color: #86C1B9;
}
#assets {
display: none;
}
## Instruction:
Use text-align to center the canvas.
## Code After:
html, body {
touch-action: none;
padding: 0;
margin: 0;
overflow: hidden;
width: 100%;
height: 100%;
}
body {
text-align: center;
background-color: #181818;
}
canvas {
background-color: #86C1B9;
}
#assets {
display: none;
}
| html, body {
touch-action: none;
padding: 0;
margin: 0;
overflow: hidden;
width: 100%;
height: 100%;
}
body {
- display: flex;
- align-items: center;
? ------
+ text-align: center;
? +++++
- justify-content: center;
background-color: #181818;
}
canvas {
background-color: #86C1B9;
}
#assets {
display: none;
}
| 4 | 0.166667 | 1 | 3 |
cda4fe0f5f2f47a2d151c752e5fd9c9ae48f2274 | app/helpers/georgia/meta_tags_helper.rb | app/helpers/georgia/meta_tags_helper.rb | module Georgia
module MetaTagsHelper
def meta_title title
site_title = ""
site_title << "#{title} | " if title
site_title << Georgia.title
content_tag :title, site_title
end
def meta_description description
return unless description
tag :meta, name: 'description', content: description
end
def meta_keywords keywords
return unless keywords
tag :meta, name: 'keywords', content: keywords
end
end
end | module Georgia
module MetaTagsHelper
def meta_title title
site_title = ""
site_title << "#{title} | " unless title.blank?
site_title << Georgia.title
content_tag :title, site_title
end
def meta_description description
return unless description
tag :meta, name: 'description', content: description
end
def meta_keywords keywords
return unless keywords
tag :meta, name: 'keywords', content: keywords
end
end
end | Fix | in title when no title | Fix | in title when no title
| Ruby | mit | georgia-cms/georgia,georgia-cms/georgia,georgia-cms/georgia | ruby | ## Code Before:
module Georgia
module MetaTagsHelper
def meta_title title
site_title = ""
site_title << "#{title} | " if title
site_title << Georgia.title
content_tag :title, site_title
end
def meta_description description
return unless description
tag :meta, name: 'description', content: description
end
def meta_keywords keywords
return unless keywords
tag :meta, name: 'keywords', content: keywords
end
end
end
## Instruction:
Fix | in title when no title
## Code After:
module Georgia
module MetaTagsHelper
def meta_title title
site_title = ""
site_title << "#{title} | " unless title.blank?
site_title << Georgia.title
content_tag :title, site_title
end
def meta_description description
return unless description
tag :meta, name: 'description', content: description
end
def meta_keywords keywords
return unless keywords
tag :meta, name: 'keywords', content: keywords
end
end
end | module Georgia
module MetaTagsHelper
def meta_title title
site_title = ""
- site_title << "#{title} | " if title
? ^^
+ site_title << "#{title} | " unless title.blank?
? ^^^^^^ +++++++
site_title << Georgia.title
content_tag :title, site_title
end
def meta_description description
return unless description
tag :meta, name: 'description', content: description
end
def meta_keywords keywords
return unless keywords
tag :meta, name: 'keywords', content: keywords
end
end
end | 2 | 0.090909 | 1 | 1 |
cea2eb7c2afba4fcf47c75d2686826b07fdaef89 | dropwizard/src/main/java/com/example/helloworld/resources/JsonResource.java | dropwizard/src/main/java/com/example/helloworld/resources/JsonResource.java |
package com.example.helloworld.resources;
import java.util.HashMap;
import java.util.Map;
import javax.ws.rs.GET;
import javax.ws.rs.Path;
import javax.ws.rs.Produces;
import javax.ws.rs.core.MediaType;
@Path("/json")
@Produces(MediaType.APPLICATION_JSON)
public class JsonResource
{
private final Map<String, String> MESSAGE = new HashMap<String, String>();
public JsonResource()
{
MESSAGE.put("message", "Hello, world!");
}
@GET
public Map<String, String> sayHello()
{
return MESSAGE;
}
}
|
package com.example.helloworld.resources;
import java.util.HashMap;
import java.util.Map;
import javax.ws.rs.GET;
import javax.ws.rs.Path;
import javax.ws.rs.Produces;
import javax.ws.rs.core.MediaType;
@Path("/json")
@Produces(MediaType.APPLICATION_JSON)
public class JsonResource
{
// Response message class (copied from 'servlet' test)
public final static class HelloMessage {
public final String message;
public HelloMessage(String m) { message = m; }
}
public JsonResource() { }
@GET
public HelloMessage sayHello()
{
return new HelloMessage("Hello, World!");
}
}
| Improve DropWizard test to use JSON generation similar to Servlet test -- more efficient that way | Improve DropWizard test to use JSON generation similar to Servlet test -- more efficient that way
| Java | bsd-3-clause | thousandsofthem/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,doom369/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,testn/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,joshk/FrameworkBenchmarks,Ocramius/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,grob/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,zapov/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,jamming/FrameworkBenchmarks,doom369/FrameworkBenchmarks,torhve/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,doom369/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,actframework/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,sxend/FrameworkBenchmarks,zapov/FrameworkBenchmarks,Verber/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,actframework/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,denkab/FrameworkBenchmarks,zapov/FrameworkBenchmarks,kbrock/FrameworkBenchmarks,julienschmidt/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,Verber/FrameworkBenchmarks,herloct/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,hperadin/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,testn/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,zapov/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,kbrock/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,seem-sky/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,herloct/FrameworkBenchmarks,jamming/FrameworkBenchmarks,joshk/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,Ocramius/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,sxend/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,actframework/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,kellabyte/FrameworkBenchmarks,leafo/FrameworkBenchmarks,sgml/FrameworkBenchmarks,Verber/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,grob/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,zapov/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,herloct/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,herloct/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,Ocramius/FrameworkBenchmarks,ratpack/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,hperadin/FrameworkBenchmarks,kbrock/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,testn/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,torhve/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,zloster/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,Verber/FrameworkBenchmarks,leafo/FrameworkBenchmarks,doom369/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,zloster/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,sxend/FrameworkBenchmarks,Verber/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,valyala/FrameworkBenchmarks,herloct/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,seem-sky/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,khellang/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,hperadin/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,dmacd/FB-try1,sxend/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,hperadin/FrameworkBenchmarks,sxend/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,leafo/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,Verber/FrameworkBenchmarks,zapov/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,denkab/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,methane/FrameworkBenchmarks,herloct/FrameworkBenchmarks,testn/FrameworkBenchmarks,leafo/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,khellang/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,Ocramius/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,sgml/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,Verber/FrameworkBenchmarks,herloct/FrameworkBenchmarks,Verber/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,khellang/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,leafo/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,hperadin/FrameworkBenchmarks,valyala/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,valyala/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,denkab/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,kbrock/FrameworkBenchmarks,kellabyte/FrameworkBenchmarks,jamming/FrameworkBenchmarks,doom369/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,testn/FrameworkBenchmarks,methane/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,herloct/FrameworkBenchmarks,sxend/FrameworkBenchmarks,herloct/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,actframework/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,testn/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,Ocramius/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,julienschmidt/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,khellang/FrameworkBenchmarks,kellabyte/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,doom369/FrameworkBenchmarks,ratpack/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,actframework/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,hperadin/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,julienschmidt/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,torhve/FrameworkBenchmarks,testn/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,kellabyte/FrameworkBenchmarks,hperadin/FrameworkBenchmarks,Verber/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,julienschmidt/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,doom369/FrameworkBenchmarks,actframework/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,sgml/FrameworkBenchmarks,kbrock/FrameworkBenchmarks,joshk/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,sxend/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,ratpack/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,jamming/FrameworkBenchmarks,grob/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,khellang/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,kbrock/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,Verber/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,denkab/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,zloster/FrameworkBenchmarks,jamming/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,torhve/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,torhve/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,dmacd/FB-try1,sgml/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,actframework/FrameworkBenchmarks,ratpack/FrameworkBenchmarks,actframework/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,zloster/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,julienschmidt/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,joshk/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,sxend/FrameworkBenchmarks,julienschmidt/FrameworkBenchmarks,actframework/FrameworkBenchmarks,Ocramius/FrameworkBenchmarks,torhve/FrameworkBenchmarks,kbrock/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,zapov/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,julienschmidt/FrameworkBenchmarks,dmacd/FB-try1,greg-hellings/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,methane/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,khellang/FrameworkBenchmarks,grob/FrameworkBenchmarks,zapov/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,ratpack/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,valyala/FrameworkBenchmarks,ratpack/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,herloct/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,sgml/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,dmacd/FB-try1,grob/FrameworkBenchmarks,zloster/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,seem-sky/FrameworkBenchmarks,dmacd/FB-try1,torhve/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,methane/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,kellabyte/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,grob/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,zapov/FrameworkBenchmarks,sxend/FrameworkBenchmarks,sgml/FrameworkBenchmarks,jamming/FrameworkBenchmarks,leafo/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,khellang/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,sxend/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,leafo/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,methane/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,leafo/FrameworkBenchmarks,dmacd/FB-try1,zane-techempower/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,leafo/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,kbrock/FrameworkBenchmarks,zloster/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,hperadin/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,joshk/FrameworkBenchmarks,khellang/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,kellabyte/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,doom369/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,valyala/FrameworkBenchmarks,julienschmidt/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,denkab/FrameworkBenchmarks,doom369/FrameworkBenchmarks,hperadin/FrameworkBenchmarks,joshk/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,dmacd/FB-try1,nbrady-techempower/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,valyala/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,kellabyte/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,testn/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,ratpack/FrameworkBenchmarks,dmacd/FB-try1,jeevatkm/FrameworkBenchmarks,julienschmidt/FrameworkBenchmarks,khellang/FrameworkBenchmarks,denkab/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,hperadin/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,leafo/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,jamming/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,doom369/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,doom369/FrameworkBenchmarks,actframework/FrameworkBenchmarks,zloster/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,sxend/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,methane/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,denkab/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,doom369/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,ratpack/FrameworkBenchmarks,doom369/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,testn/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,kbrock/FrameworkBenchmarks,testn/FrameworkBenchmarks,julienschmidt/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,sgml/FrameworkBenchmarks,seem-sky/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,dmacd/FB-try1,markkolich/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,actframework/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,grob/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,kbrock/FrameworkBenchmarks,seem-sky/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,valyala/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,herloct/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,zloster/FrameworkBenchmarks,kellabyte/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,methane/FrameworkBenchmarks,ratpack/FrameworkBenchmarks,dmacd/FB-try1,Synchro/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,doom369/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,herloct/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,jamming/FrameworkBenchmarks,ratpack/FrameworkBenchmarks,doom369/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,denkab/FrameworkBenchmarks,methane/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,kellabyte/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,zapov/FrameworkBenchmarks,khellang/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,hperadin/FrameworkBenchmarks,ratpack/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,zloster/FrameworkBenchmarks,doom369/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,kellabyte/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,seem-sky/FrameworkBenchmarks,methane/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,ratpack/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,torhve/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,joshk/FrameworkBenchmarks,kellabyte/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,joshk/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,actframework/FrameworkBenchmarks,sxend/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,zapov/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,leafo/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,Verber/FrameworkBenchmarks,joshk/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,sgml/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,zloster/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,torhve/FrameworkBenchmarks,denkab/FrameworkBenchmarks,jamming/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,raziel057/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,Verber/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,joshk/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,sxend/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,julienschmidt/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,sgml/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,methane/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,sgml/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,sxend/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,zapov/FrameworkBenchmarks,zloster/FrameworkBenchmarks,valyala/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,jamming/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,julienschmidt/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,valyala/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,valyala/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,sgml/FrameworkBenchmarks,zloster/FrameworkBenchmarks,zapov/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,testn/FrameworkBenchmarks,xitrum-framework/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,methane/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,actframework/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,Ocramius/FrameworkBenchmarks,sgml/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,khellang/FrameworkBenchmarks,victorbriz/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,testn/FrameworkBenchmarks,Ocramius/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,denkab/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,seem-sky/FrameworkBenchmarks,kellabyte/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,Ocramius/FrameworkBenchmarks,grob/FrameworkBenchmarks,zloster/FrameworkBenchmarks,zloster/FrameworkBenchmarks,joshk/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,zloster/FrameworkBenchmarks,joshk/FrameworkBenchmarks,joshk/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,sgml/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,dmacd/FB-try1,jetty-project/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,Verber/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,actframework/FrameworkBenchmarks,grob/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,jamming/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,hperadin/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,valyala/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,leafo/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,jamming/FrameworkBenchmarks,zloster/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,sgml/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,herloct/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,denkab/FrameworkBenchmarks,grob/FrameworkBenchmarks,doom369/FrameworkBenchmarks,jamming/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,Rydgel/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,sxend/FrameworkBenchmarks,testn/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,sxend/FrameworkBenchmarks,Ocramius/FrameworkBenchmarks,torhve/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,actframework/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,actframework/FrameworkBenchmarks,nkasvosve/FrameworkBenchmarks,zloster/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,kostya-sh/FrameworkBenchmarks,seem-sky/FrameworkBenchmarks,sxend/FrameworkBenchmarks,seem-sky/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,zane-techempower/FrameworkBenchmarks,RockinRoel/FrameworkBenchmarks,zapov/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,grob/FrameworkBenchmarks,seem-sky/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,donovanmuller/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,zapov/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,nathana1/FrameworkBenchmarks,seem-sky/FrameworkBenchmarks,grob/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,grob/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,sanjoydesk/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,kbrock/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,zdanek/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,zloster/FrameworkBenchmarks,knewmanTE/FrameworkBenchmarks,denkab/FrameworkBenchmarks,zhuochenKIDD/FrameworkBenchmarks,jeevatkm/FrameworkBenchmarks,joshk/FrameworkBenchmarks,alubbe/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,MTDdk/FrameworkBenchmarks,testn/FrameworkBenchmarks,jetty-project/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,hperadin/FrameworkBenchmarks,mfirry/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,khellang/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,valyala/FrameworkBenchmarks,dmacd/FB-try1,diablonhn/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,Ocramius/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,jaguililla/FrameworkBenchmarks,Rayne/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,Ocramius/FrameworkBenchmarks,zapov/FrameworkBenchmarks,khellang/FrameworkBenchmarks,marko-asplund/FrameworkBenchmarks,lcp0578/FrameworkBenchmarks,grob/FrameworkBenchmarks,seem-sky/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks,PermeAgility/FrameworkBenchmarks,steveklabnik/FrameworkBenchmarks,valyala/FrameworkBenchmarks,khellang/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,seem-sky/FrameworkBenchmarks,torhve/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,jamming/FrameworkBenchmarks,greg-hellings/FrameworkBenchmarks,sxend/FrameworkBenchmarks,kellabyte/FrameworkBenchmarks,methane/FrameworkBenchmarks,methane/FrameworkBenchmarks,Jesterovskiy/FrameworkBenchmarks,martin-g/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,valyala/FrameworkBenchmarks,kbrock/FrameworkBenchmarks,k-r-g/FrameworkBenchmarks,F3Community/FrameworkBenchmarks,psfblair/FrameworkBenchmarks,denkab/FrameworkBenchmarks,methane/FrameworkBenchmarks,circlespainter/FrameworkBenchmarks,nbrady-techempower/FrameworkBenchmarks,herloct/FrameworkBenchmarks,denkab/FrameworkBenchmarks,s-ludwig/FrameworkBenchmarks,jebbstewart/FrameworkBenchmarks,Verber/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,stefanocasazza/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,yunspace/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,greenlaw110/FrameworkBenchmarks,Eyepea/FrameworkBenchmarks,hamiltont/FrameworkBenchmarks,markkolich/FrameworkBenchmarks,Synchro/FrameworkBenchmarks,ashawnbandy-te-tfb/FrameworkBenchmarks,diablonhn/FrameworkBenchmarks,herloct/FrameworkBenchmarks,thousandsofthem/FrameworkBenchmarks,fabianmurariu/FrameworkBenchmarks,waiteb3/FrameworkBenchmarks,youprofit/FrameworkBenchmarks,torhve/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,saturday06/FrameworkBenchmarks,sagenschneider/FrameworkBenchmarks,Dith3r/FrameworkBenchmarks | java | ## Code Before:
package com.example.helloworld.resources;
import java.util.HashMap;
import java.util.Map;
import javax.ws.rs.GET;
import javax.ws.rs.Path;
import javax.ws.rs.Produces;
import javax.ws.rs.core.MediaType;
@Path("/json")
@Produces(MediaType.APPLICATION_JSON)
public class JsonResource
{
private final Map<String, String> MESSAGE = new HashMap<String, String>();
public JsonResource()
{
MESSAGE.put("message", "Hello, world!");
}
@GET
public Map<String, String> sayHello()
{
return MESSAGE;
}
}
## Instruction:
Improve DropWizard test to use JSON generation similar to Servlet test -- more efficient that way
## Code After:
package com.example.helloworld.resources;
import java.util.HashMap;
import java.util.Map;
import javax.ws.rs.GET;
import javax.ws.rs.Path;
import javax.ws.rs.Produces;
import javax.ws.rs.core.MediaType;
@Path("/json")
@Produces(MediaType.APPLICATION_JSON)
public class JsonResource
{
// Response message class (copied from 'servlet' test)
public final static class HelloMessage {
public final String message;
public HelloMessage(String m) { message = m; }
}
public JsonResource() { }
@GET
public HelloMessage sayHello()
{
return new HelloMessage("Hello, World!");
}
}
|
package com.example.helloworld.resources;
import java.util.HashMap;
import java.util.Map;
import javax.ws.rs.GET;
import javax.ws.rs.Path;
import javax.ws.rs.Produces;
import javax.ws.rs.core.MediaType;
@Path("/json")
@Produces(MediaType.APPLICATION_JSON)
public class JsonResource
{
- private final Map<String, String> MESSAGE = new HashMap<String, String>();
+ // Response message class (copied from 'servlet' test)
+ public final static class HelloMessage {
+ public final String message;
+ public HelloMessage(String m) { message = m; }
- public JsonResource()
- {
- MESSAGE.put("message", "Hello, world!");
}
+ public JsonResource() { }
+
@GET
- public Map<String, String> sayHello()
+ public HelloMessage sayHello()
{
- return MESSAGE;
+ return new HelloMessage("Hello, World!");
}
} | 14 | 0.5 | 8 | 6 |
c33b5c9ef5b597850f66c66e8b68451f15adcfe7 | libyaul/scu/bus/a/cs0/dram-cartridge/dram-cartridge.h | libyaul/scu/bus/a/cs0/dram-cartridge/dram-cartridge.h | /*
* Copyright (c) 2012 Israel Jacques
* See LICENSE for details.
*
* Israel Jacques <mrko@eecs.berkeley.edu>
*/
#ifndef _DRAM_CARTRIDGE_H__
#define _DRAM_CARTRIDGE_H__
extern void dram_init(void);
#endif /* !_DRAM_CARTRIDGE_H_
| /*
* Copyright (c) 2012 Israel Jacques
* See LICENSE for details.
*
* Israel Jacques <mrko@eecs.berkeley.edu>
*/
#ifndef _DRAM_CARTRIDGE_H__
#define _DRAM_CARTRIDGE_H__
extern void dram_cartridge_init(void);
#endif /* !_DRAM_CARTRIDGE_H_
| Change incorrect prototype from 'dram_init(void)' to 'dram_cartridge_init(void)' | Change incorrect prototype from 'dram_init(void)' to 'dram_cartridge_init(void)'
| C | mit | ChillyWillyGuru/libyaul,ChillyWillyGuru/libyaul | c | ## Code Before:
/*
* Copyright (c) 2012 Israel Jacques
* See LICENSE for details.
*
* Israel Jacques <mrko@eecs.berkeley.edu>
*/
#ifndef _DRAM_CARTRIDGE_H__
#define _DRAM_CARTRIDGE_H__
extern void dram_init(void);
#endif /* !_DRAM_CARTRIDGE_H_
## Instruction:
Change incorrect prototype from 'dram_init(void)' to 'dram_cartridge_init(void)'
## Code After:
/*
* Copyright (c) 2012 Israel Jacques
* See LICENSE for details.
*
* Israel Jacques <mrko@eecs.berkeley.edu>
*/
#ifndef _DRAM_CARTRIDGE_H__
#define _DRAM_CARTRIDGE_H__
extern void dram_cartridge_init(void);
#endif /* !_DRAM_CARTRIDGE_H_
| /*
* Copyright (c) 2012 Israel Jacques
* See LICENSE for details.
*
* Israel Jacques <mrko@eecs.berkeley.edu>
*/
#ifndef _DRAM_CARTRIDGE_H__
#define _DRAM_CARTRIDGE_H__
- extern void dram_init(void);
+ extern void dram_cartridge_init(void);
? ++++++++++
#endif /* !_DRAM_CARTRIDGE_H_ | 2 | 0.153846 | 1 | 1 |
878f01e03c4bee5ea32301a9eefee3162317f406 | src/app.yaml | src/app.yaml | application: temp
version: 1
runtime: go
api_version: go1
handlers:
- url: /
static_files: frontend/app/index.html
upload: frontend/app/(.*\.html)
- url: /css
static_dir: frontend/app/css
- url: /img
static_dir: frontend/app/img
- url: /js
static_dir: frontend/app/js
- url: /lib
static_dir: frontend/app/lib
- url: /partials
static_dir: frontend/app/partials
- url: /bower_components
static_dir: frontend/app/bower_components
- url: /.*
script: _go_app
| application: temp
version: 1
runtime: go
api_version: go1
handlers:
- url: /
static_files: frontend/app/index.html
upload: frontend/app/(.*\.html)
- url: /css
static_dir: frontend/app/css
- url: /img
static_dir: frontend/app/img
- url: /js
static_dir: frontend/app/js
- url: /partials
static_dir: frontend/app/partials
- url: /bower_components
static_dir: frontend/app/bower_components
- url: /.*
script: _go_app
| Remove the lib/ static folder. | Remove the lib/ static folder.
| YAML | mit | jimbeaudoin/golang-appengine-boilerplate,jimbeaudoin/golang-appengine-boilerplate | yaml | ## Code Before:
application: temp
version: 1
runtime: go
api_version: go1
handlers:
- url: /
static_files: frontend/app/index.html
upload: frontend/app/(.*\.html)
- url: /css
static_dir: frontend/app/css
- url: /img
static_dir: frontend/app/img
- url: /js
static_dir: frontend/app/js
- url: /lib
static_dir: frontend/app/lib
- url: /partials
static_dir: frontend/app/partials
- url: /bower_components
static_dir: frontend/app/bower_components
- url: /.*
script: _go_app
## Instruction:
Remove the lib/ static folder.
## Code After:
application: temp
version: 1
runtime: go
api_version: go1
handlers:
- url: /
static_files: frontend/app/index.html
upload: frontend/app/(.*\.html)
- url: /css
static_dir: frontend/app/css
- url: /img
static_dir: frontend/app/img
- url: /js
static_dir: frontend/app/js
- url: /partials
static_dir: frontend/app/partials
- url: /bower_components
static_dir: frontend/app/bower_components
- url: /.*
script: _go_app
| application: temp
version: 1
runtime: go
api_version: go1
handlers:
- url: /
static_files: frontend/app/index.html
upload: frontend/app/(.*\.html)
- url: /css
static_dir: frontend/app/css
- url: /img
static_dir: frontend/app/img
- url: /js
static_dir: frontend/app/js
- - url: /lib
- static_dir: frontend/app/lib
-
- url: /partials
static_dir: frontend/app/partials
- url: /bower_components
static_dir: frontend/app/bower_components
- url: /.*
script: _go_app | 3 | 0.1 | 0 | 3 |
d23ad78a65b593b611d1f1885631b3487958effb | shellcheck.sh | shellcheck.sh |
set -eo pipefail
# Paths to ignore.
shellcheck_skip="
./tools/git-churn.sh
./bash_includes/git-completion.sh
./bash_includes/git-prompt.sh
./vim/undo/*
"
# Linting errors to ignore.
export SHELLCHECK_OPTS="-e SC2164"
# Define find_cmd() to exclude files from $shellcheck_skip.
find_cmd() {
cmd="find . -type f -and \( -perm +111 -or -name '*.sh' \) $(find_prunes)"
for path in $shellcheck_skip; do
cmd="$cmd ! -path '$path'"
done
echo "$cmd"
}
# Protect find_cmd() to prevent ./build/build.sh from redefining it.
readonly -f find_cmd
# Run the linter.
source ./build/build.sh
|
set -eo pipefail
# Paths to ignore.
shellcheck_skip="
tools/git-churn.sh
bash_includes/git-completion.sh
bash_includes/git-prompt.sh
vim/undo/
"
# Linting errors to ignore.
export SHELLCHECK_OPTS="-e SC2164"
# Define find_scripts() to exclude files from $shellcheck_skip.
find_scripts() {
cmd="git ls-tree -r HEAD | egrep '^1007|.*\..*sh$' | awk '{print \$4}'"
for path in $shellcheck_skip; do
echo $path
cmd="$cmd | grep -v '^$path'"
done
echo "$cmd"
}
# Protect find_scripts() to prevent ./build/build.sh from redefining it.
readonly -f find_scripts
# Run the linter.
source ./build/build.sh
| Update build script to match changes in framework. | Update build script to match changes in framework.
| Shell | mit | guillaumeaubert/dot-files,guillaumeaubert/dot-files | shell | ## Code Before:
set -eo pipefail
# Paths to ignore.
shellcheck_skip="
./tools/git-churn.sh
./bash_includes/git-completion.sh
./bash_includes/git-prompt.sh
./vim/undo/*
"
# Linting errors to ignore.
export SHELLCHECK_OPTS="-e SC2164"
# Define find_cmd() to exclude files from $shellcheck_skip.
find_cmd() {
cmd="find . -type f -and \( -perm +111 -or -name '*.sh' \) $(find_prunes)"
for path in $shellcheck_skip; do
cmd="$cmd ! -path '$path'"
done
echo "$cmd"
}
# Protect find_cmd() to prevent ./build/build.sh from redefining it.
readonly -f find_cmd
# Run the linter.
source ./build/build.sh
## Instruction:
Update build script to match changes in framework.
## Code After:
set -eo pipefail
# Paths to ignore.
shellcheck_skip="
tools/git-churn.sh
bash_includes/git-completion.sh
bash_includes/git-prompt.sh
vim/undo/
"
# Linting errors to ignore.
export SHELLCHECK_OPTS="-e SC2164"
# Define find_scripts() to exclude files from $shellcheck_skip.
find_scripts() {
cmd="git ls-tree -r HEAD | egrep '^1007|.*\..*sh$' | awk '{print \$4}'"
for path in $shellcheck_skip; do
echo $path
cmd="$cmd | grep -v '^$path'"
done
echo "$cmd"
}
# Protect find_scripts() to prevent ./build/build.sh from redefining it.
readonly -f find_scripts
# Run the linter.
source ./build/build.sh
|
set -eo pipefail
# Paths to ignore.
shellcheck_skip="
- ./tools/git-churn.sh
? --
+ tools/git-churn.sh
- ./bash_includes/git-completion.sh
? --
+ bash_includes/git-completion.sh
- ./bash_includes/git-prompt.sh
? --
+ bash_includes/git-prompt.sh
- ./vim/undo/*
? -- -
+ vim/undo/
"
# Linting errors to ignore.
export SHELLCHECK_OPTS="-e SC2164"
- # Define find_cmd() to exclude files from $shellcheck_skip.
? ^^
+ # Define find_scripts() to exclude files from $shellcheck_skip.
? + ^^^^^
- find_cmd() {
- cmd="find . -type f -and \( -perm +111 -or -name '*.sh' \) $(find_prunes)"
+ find_scripts() {
+ cmd="git ls-tree -r HEAD | egrep '^1007|.*\..*sh$' | awk '{print \$4}'"
for path in $shellcheck_skip; do
+ echo $path
- cmd="$cmd ! -path '$path'"
? ^ ^^^^
+ cmd="$cmd | grep -v '^$path'"
? ^^^^^^ ^ +
done
echo "$cmd"
}
- # Protect find_cmd() to prevent ./build/build.sh from redefining it.
? ^^
+ # Protect find_scripts() to prevent ./build/build.sh from redefining it.
? + ^^^^^
- readonly -f find_cmd
? ^^
+ readonly -f find_scripts
? + ^^^^^
# Run the linter.
source ./build/build.sh | 21 | 0.75 | 11 | 10 |
7a1ad6ec07cd7134bc90b4a8c9caa2aa9a5c22c9 | config-model-api/src/main/java/com/yahoo/config/model/api/Reindexing.java | config-model-api/src/main/java/com/yahoo/config/model/api/Reindexing.java | // Copyright Verizon Media. Licensed under the terms of the Apache 2.0 license. See LICENSE in the project root.
package com.yahoo.config.model.api;
import java.time.Instant;
import java.util.Map;
/**
* Status of reindexing for the documents of an application.
*
* @author jonmv
*/
public interface Reindexing {
/** Reindexing status for a given application, cluster and document type. */
default Status status(String cluster, String documentType) { return () -> Instant.MAX; }
/** Reindexing status of a given document type in a given cluster in a given application. */
interface Status {
/** The instant at which reindexing may begin. */
Instant ready();
}
}
| // Copyright Verizon Media. Licensed under the terms of the Apache 2.0 license. See LICENSE in the project root.
package com.yahoo.config.model.api;
import java.time.Instant;
import java.util.Map;
/**
* Status of reindexing for the documents of an application.
*
* @author jonmv
*/
public interface Reindexing {
/** No reindexing should be done for this document type and cluster. */
Status NO_REINDEXING = () -> Instant.MAX;
/** Reindexing status for a given application, cluster and document type. */
default Status status(String cluster, String documentType) { return NO_REINDEXING; }
/** Reindexing status of a given document type in a given cluster in a given application. */
interface Status {
/** The instant at which reindexing may begin. */
Instant ready();
}
}
| Use more explicit value when reindexing should not be done | Use more explicit value when reindexing should not be done
| Java | apache-2.0 | vespa-engine/vespa,vespa-engine/vespa,vespa-engine/vespa,vespa-engine/vespa,vespa-engine/vespa,vespa-engine/vespa,vespa-engine/vespa,vespa-engine/vespa,vespa-engine/vespa,vespa-engine/vespa | java | ## Code Before:
// Copyright Verizon Media. Licensed under the terms of the Apache 2.0 license. See LICENSE in the project root.
package com.yahoo.config.model.api;
import java.time.Instant;
import java.util.Map;
/**
* Status of reindexing for the documents of an application.
*
* @author jonmv
*/
public interface Reindexing {
/** Reindexing status for a given application, cluster and document type. */
default Status status(String cluster, String documentType) { return () -> Instant.MAX; }
/** Reindexing status of a given document type in a given cluster in a given application. */
interface Status {
/** The instant at which reindexing may begin. */
Instant ready();
}
}
## Instruction:
Use more explicit value when reindexing should not be done
## Code After:
// Copyright Verizon Media. Licensed under the terms of the Apache 2.0 license. See LICENSE in the project root.
package com.yahoo.config.model.api;
import java.time.Instant;
import java.util.Map;
/**
* Status of reindexing for the documents of an application.
*
* @author jonmv
*/
public interface Reindexing {
/** No reindexing should be done for this document type and cluster. */
Status NO_REINDEXING = () -> Instant.MAX;
/** Reindexing status for a given application, cluster and document type. */
default Status status(String cluster, String documentType) { return NO_REINDEXING; }
/** Reindexing status of a given document type in a given cluster in a given application. */
interface Status {
/** The instant at which reindexing may begin. */
Instant ready();
}
}
| // Copyright Verizon Media. Licensed under the terms of the Apache 2.0 license. See LICENSE in the project root.
package com.yahoo.config.model.api;
import java.time.Instant;
import java.util.Map;
/**
* Status of reindexing for the documents of an application.
*
* @author jonmv
*/
public interface Reindexing {
+ /** No reindexing should be done for this document type and cluster. */
+ Status NO_REINDEXING = () -> Instant.MAX;
+
/** Reindexing status for a given application, cluster and document type. */
- default Status status(String cluster, String documentType) { return () -> Instant.MAX; }
? ^^^^^^ ^^^^^^^^^
+ default Status status(String cluster, String documentType) { return NO_REINDEXING; }
? ^^^^^ ^^^ +++
/** Reindexing status of a given document type in a given cluster in a given application. */
interface Status {
/** The instant at which reindexing may begin. */
Instant ready();
}
} | 5 | 0.2 | 4 | 1 |
4bee0753f42c1f598e0e670cd5dc5562f59f5b77 | lineman/app/services/key_event_service.coffee | lineman/app/services/key_event_service.coffee | angular.module('loomioApp').factory 'KeyEventService', ($rootScope) ->
new class KeyEventService
keyboardShortcuts:
73: 'pressedI'
71: 'pressedG'
84: 'pressedT'
27: 'pressedEsc'
13: 'pressedEnter'
191: 'pressedSlash'
38: 'pressedUpArrow'
40: 'pressedDownArrow'
broadcast: (event) ->
key = @keyboardShortcuts[event.which]
if key == 'pressedEnter' or (key and !event.ctrlKey and !event.metaKey)
$rootScope.$broadcast key, event, angular.element(document.activeElement)[0]
registerKeyEvent: (scope, eventCode, execute, shouldExecute) ->
shouldExecute = shouldExecute or @defaultShouldExecute
scope.$on eventCode, (angularEvent, originalEvent, active) ->
if shouldExecute(active, originalEvent)
angularEvent.preventDefault() and originalEvent.preventDefault()
execute(active)
defaultShouldExecute: (active = {}, event = {}) ->
!event.ctrlKey and !event.altKey and !_.contains(['INPUT', 'TEXTAREA', 'SELECT'], active.nodeName)
submitOnEnter: (scope) ->
@registerKeyEvent scope, 'pressedEnter', scope.submit, (active, event) =>
(event.ctrlKey or event.metaKey) and
angular.element(active).scope() == scope and
_.contains(active.classList, 'lmo-primary-form-input')
| angular.module('loomioApp').factory 'KeyEventService', ($rootScope) ->
new class KeyEventService
keyboardShortcuts:
73: 'pressedI'
71: 'pressedG'
84: 'pressedT'
27: 'pressedEsc'
13: 'pressedEnter'
191: 'pressedSlash'
38: 'pressedUpArrow'
40: 'pressedDownArrow'
broadcast: (event) ->
key = @keyboardShortcuts[event.which]
if key == 'pressedEnter' or (key and !event.ctrlKey and !event.metaKey)
$rootScope.$broadcast key, event, angular.element(document.activeElement)[0]
registerKeyEvent: (scope, eventCode, execute, shouldExecute) ->
shouldExecute = shouldExecute or @defaultShouldExecute
scope.$on eventCode, (angularEvent, originalEvent, active) ->
if shouldExecute(active, originalEvent)
angularEvent.preventDefault() and originalEvent.preventDefault()
execute(active)
defaultShouldExecute: (active = {}, event = {}) ->
!event.ctrlKey and !event.altKey and !_.contains(['INPUT', 'TEXTAREA', 'SELECT'], active.nodeName)
submitOnEnter: (scope) ->
@registerKeyEvent scope, 'pressedEnter', scope.submit, (active, event) =>
(event.ctrlKey or event.metaKey) and
angular.element(active).scope().$$isolateBindings == scope.$$isolateBindings and
_.contains(active.classList, 'lmo-primary-form-input')
| Fix cmd + enter submit | Fix cmd + enter submit
| CoffeeScript | agpl-3.0 | mhjb/loomio,loomio/loomio,sicambria/loomio,mhjb/loomio,mhjb/loomio,FSFTN/Loomio,loomio/loomio,piratas-ar/loomio,piratas-ar/loomio,piratas-ar/loomio,FSFTN/Loomio,mhjb/loomio,sicambria/loomio,sicambria/loomio,FSFTN/Loomio,loomio/loomio,piratas-ar/loomio,sicambria/loomio,loomio/loomio,FSFTN/Loomio | coffeescript | ## Code Before:
angular.module('loomioApp').factory 'KeyEventService', ($rootScope) ->
new class KeyEventService
keyboardShortcuts:
73: 'pressedI'
71: 'pressedG'
84: 'pressedT'
27: 'pressedEsc'
13: 'pressedEnter'
191: 'pressedSlash'
38: 'pressedUpArrow'
40: 'pressedDownArrow'
broadcast: (event) ->
key = @keyboardShortcuts[event.which]
if key == 'pressedEnter' or (key and !event.ctrlKey and !event.metaKey)
$rootScope.$broadcast key, event, angular.element(document.activeElement)[0]
registerKeyEvent: (scope, eventCode, execute, shouldExecute) ->
shouldExecute = shouldExecute or @defaultShouldExecute
scope.$on eventCode, (angularEvent, originalEvent, active) ->
if shouldExecute(active, originalEvent)
angularEvent.preventDefault() and originalEvent.preventDefault()
execute(active)
defaultShouldExecute: (active = {}, event = {}) ->
!event.ctrlKey and !event.altKey and !_.contains(['INPUT', 'TEXTAREA', 'SELECT'], active.nodeName)
submitOnEnter: (scope) ->
@registerKeyEvent scope, 'pressedEnter', scope.submit, (active, event) =>
(event.ctrlKey or event.metaKey) and
angular.element(active).scope() == scope and
_.contains(active.classList, 'lmo-primary-form-input')
## Instruction:
Fix cmd + enter submit
## Code After:
angular.module('loomioApp').factory 'KeyEventService', ($rootScope) ->
new class KeyEventService
keyboardShortcuts:
73: 'pressedI'
71: 'pressedG'
84: 'pressedT'
27: 'pressedEsc'
13: 'pressedEnter'
191: 'pressedSlash'
38: 'pressedUpArrow'
40: 'pressedDownArrow'
broadcast: (event) ->
key = @keyboardShortcuts[event.which]
if key == 'pressedEnter' or (key and !event.ctrlKey and !event.metaKey)
$rootScope.$broadcast key, event, angular.element(document.activeElement)[0]
registerKeyEvent: (scope, eventCode, execute, shouldExecute) ->
shouldExecute = shouldExecute or @defaultShouldExecute
scope.$on eventCode, (angularEvent, originalEvent, active) ->
if shouldExecute(active, originalEvent)
angularEvent.preventDefault() and originalEvent.preventDefault()
execute(active)
defaultShouldExecute: (active = {}, event = {}) ->
!event.ctrlKey and !event.altKey and !_.contains(['INPUT', 'TEXTAREA', 'SELECT'], active.nodeName)
submitOnEnter: (scope) ->
@registerKeyEvent scope, 'pressedEnter', scope.submit, (active, event) =>
(event.ctrlKey or event.metaKey) and
angular.element(active).scope().$$isolateBindings == scope.$$isolateBindings and
_.contains(active.classList, 'lmo-primary-form-input')
| angular.module('loomioApp').factory 'KeyEventService', ($rootScope) ->
new class KeyEventService
keyboardShortcuts:
73: 'pressedI'
71: 'pressedG'
84: 'pressedT'
27: 'pressedEsc'
13: 'pressedEnter'
191: 'pressedSlash'
38: 'pressedUpArrow'
40: 'pressedDownArrow'
broadcast: (event) ->
key = @keyboardShortcuts[event.which]
if key == 'pressedEnter' or (key and !event.ctrlKey and !event.metaKey)
$rootScope.$broadcast key, event, angular.element(document.activeElement)[0]
registerKeyEvent: (scope, eventCode, execute, shouldExecute) ->
shouldExecute = shouldExecute or @defaultShouldExecute
scope.$on eventCode, (angularEvent, originalEvent, active) ->
if shouldExecute(active, originalEvent)
angularEvent.preventDefault() and originalEvent.preventDefault()
execute(active)
defaultShouldExecute: (active = {}, event = {}) ->
!event.ctrlKey and !event.altKey and !_.contains(['INPUT', 'TEXTAREA', 'SELECT'], active.nodeName)
submitOnEnter: (scope) ->
@registerKeyEvent scope, 'pressedEnter', scope.submit, (active, event) =>
(event.ctrlKey or event.metaKey) and
- angular.element(active).scope() == scope and
+ angular.element(active).scope().$$isolateBindings == scope.$$isolateBindings and
_.contains(active.classList, 'lmo-primary-form-input') | 2 | 0.060606 | 1 | 1 |
0280f9ad704ab2dfd66ce3be554f43df3d22da39 | tensorflow/tools/ci_build/install/install_deb_packages.sh | tensorflow/tools/ci_build/install/install_deb_packages.sh |
set -e
# Install dependencies from ubuntu deb repository.
apt-get update
apt-get install -y --no-install-recommends \
autoconf \
automake \
build-essential \
cmake \
curl \
ffmpeg \
git \
libcurl4-openssl-dev \
libtool \
openjdk-8-jdk \
openjdk-8-jre-headless \
pkg-config \
python-dev \
python-pip \
python-virtualenv \
python3-dev \
python3-pip \
rsync \
sudo \
swig \
unzip \
wget \
zip \
zlib1g-dev
apt-get clean
rm -rf /var/lib/apt/lists/*
|
set -e
# Install dependencies from ubuntu deb repository.
apt-get update
apt-get install -y --no-install-recommends \
autoconf \
automake \
build-essential \
cmake \
curl \
ffmpeg \
git \
libcurl4-openssl-dev \
libtool \
openjdk-8-jdk \
openjdk-8-jre-headless \
pkg-config \
python-dev \
python-pip \
python-virtualenv \
python3-dev \
python3-pip \
rsync \
sudo \
swig \
unzip \
wget \
zip \
zlib1g-dev
# Install ca-certificates, and update the certificate store.
apt-get install ca-certificates-java
update-ca-certificates -f
apt-get clean
rm -rf /var/lib/apt/lists/*
| Fix certificate issues in bazel for ci build. | Fix certificate issues in bazel for ci build.
(cherry picked from commit 923c999d6e910d7566f10a852d2ff12c4a88fad8)
| Shell | apache-2.0 | girving/tensorflow,memo/tensorflow,benoitsteiner/tensorflow-opencl,mdrumond/tensorflow,kamcpp/tensorflow,benoitsteiner/tensorflow-xsmm,chenjun0210/tensorflow,MycChiu/tensorflow,code-sauce/tensorflow,alheinecke/tensorflow-xsmm,tongwang01/tensorflow,theflofly/tensorflow,nanditav/15712-TensorFlow,caisq/tensorflow,alistairlow/tensorflow,chemelnucfin/tensorflow,ArtsiomCh/tensorflow,tensorflow/tensorflow-pywrap_saved_model,alisidd/tensorflow,pavelchristof/gomoku-ai,aldian/tensorflow,Bulochkin/tensorflow_pack,aam-at/tensorflow,renyi533/tensorflow,snnn/tensorflow,jhseu/tensorflow,Bismarrck/tensorflow,ageron/tensorflow,kobejean/tensorflow,ibmsoe/tensorflow,lukeiwanski/tensorflow,lukeiwanski/tensorflow,a-doumoulakis/tensorflow,with-git/tensorflow,kevin-coder/tensorflow-fork,wangyum/tensorflow,sarvex/tensorflow,eerwitt/tensorflow,tensorflow/tensorflow-pywrap_saved_model,nolanliou/tensorflow,rdipietro/tensorflow,krikru/tensorflow-opencl,tensorflow/tensorflow-experimental_link_static_libraries_once,tensorflow/tensorflow-pywrap_saved_model,hsaputra/tensorflow,alistairlow/tensorflow,alistairlow/tensorflow,tntnatbry/tensorflow,tornadozou/tensorflow,JingJunYin/tensorflow,brchiu/tensorflow,annarev/tensorflow,DavidNorman/tensorflow,admcrae/tensorflow,cancan101/tensorflow,alheinecke/tensorflow-xsmm,HKUST-SING/tensorflow,LUTAN/tensorflow,ychfan/tensorflow,guschmue/tensorflow,ppries/tensorflow,caisq/tensorflow,manipopopo/tensorflow,yanchen036/tensorflow,alivecor/tensorflow,girving/tensorflow,tensorflow/tensorflow-pywrap_tf_optimizer,gojira/tensorflow,aam-at/tensorflow,JVillella/tensorflow,jhseu/tensorflow,anilmuthineni/tensorflow,lakshayg/tensorflow,gautam1858/tensorflow,rdipietro/tensorflow,mengxn/tensorflow,theflofly/tensorflow,anilmuthineni/tensorflow,unsiloai/syntaxnet-ops-hack,paolodedios/tensorflow,markslwong/tensorflow,AnishShah/tensorflow,eerwitt/tensorflow,sjperkins/tensorflow,HKUST-SING/tensorflow,av8ramit/tensorflow,ageron/tensorflow,tensorflow/tensorflow-experimental_link_static_libraries_once,mavenlin/tensorflow,sarvex/tensorflow,Intel-tensorflow/tensorflow,hfp/tensorflow-xsmm,AndreasMadsen/tensorflow,krikru/tensorflow-opencl,karllessard/tensorflow,lukeiwanski/tensorflow,ibmsoe/tensorflow,mavenlin/tensorflow,Mazecreator/tensorflow,maciekcc/tensorflow,karllessard/tensorflow,mortada/tensorflow,DCSaunders/tensorflow,laszlocsomor/tensorflow,Xeralux/tensorflow,drpngx/tensorflow,dongjoon-hyun/tensorflow,gautam1858/tensorflow,meteorcloudy/tensorflow,alsrgv/tensorflow,alsrgv/tensorflow,jalexvig/tensorflow,xodus7/tensorflow,code-sauce/tensorflow,karllessard/tensorflow,karllessard/tensorflow,rabipanda/tensorflow,chenjun0210/tensorflow,alivecor/tensorflow,Intel-Corporation/tensorflow,martinwicke/tensorflow,thesuperzapper/tensorflow,benoitsteiner/tensorflow-opencl,ishay2b/tensorflow,seanli9jan/tensorflow,seanli9jan/tensorflow,ArtsiomCh/tensorflow,jeffzheng1/tensorflow,alistairlow/tensorflow,hehongliang/tensorflow,rdipietro/tensorflow,snnn/tensorflow,sandeepdsouza93/TensorFlow-15712,arborh/tensorflow,taknevski/tensorflow-xsmm,nolanliou/tensorflow,hfp/tensorflow-xsmm,kevin-coder/tensorflow-fork,theflofly/tensorflow,dyoung418/tensorflow,alsrgv/tensorflow,cxxgtxy/tensorflow,renyi533/tensorflow,hehongliang/tensorflow,Mazecreator/tensorflow,nburn42/tensorflow,krikru/tensorflow-opencl,allenlavoie/tensorflow,adit-chandra/tensorflow,seanli9jan/tensorflow,horance-liu/tensorflow,andrewcmyers/tensorflow,benoitsteiner/tensorflow-xsmm,raymondxyang/tensorflow,mavenlin/tensorflow,xodus7/tensorflow,chemelnucfin/tensorflow,chris-chris/tensorflow,nolanliou/tensorflow,karllessard/tensorflow,a-doumoulakis/tensorflow,jendap/tensorflow,Xeralux/tensorflow,aldian/tensorflow,chemelnucfin/tensorflow,jhaux/tensorflow,handroissuazo/tensorflow,adit-chandra/tensorflow,AndreasMadsen/tensorflow,unsiloai/syntaxnet-ops-hack,dyoung418/tensorflow,AnishShah/tensorflow,Carmezim/tensorflow,Bismarrck/tensorflow,tornadozou/tensorflow,benoitsteiner/tensorflow-xsmm,chenjun0210/tensorflow,chris-chris/tensorflow,llhe/tensorflow,jart/tensorflow,JingJunYin/tensorflow,MycChiu/tensorflow,tongwang01/tensorflow,krikru/tensorflow-opencl,jalexvig/tensorflow,codrut3/tensorflow,kchodorow/tensorflow,eaplatanios/tensorflow,Kongsea/tensorflow,strint/tensorflow,JingJunYin/tensorflow,sjperkins/tensorflow,arborh/tensorflow,nburn42/tensorflow,zycdragonball/tensorflow,ghchinoy/tensorflow,benoitsteiner/tensorflow-xsmm,ishay2b/tensorflow,nikste/tensorflow,theflofly/tensorflow,meteorcloudy/tensorflow,HKUST-SING/tensorflow,caisq/tensorflow,arborh/tensorflow,alshedivat/tensorflow,laszlocsomor/tensorflow,haeusser/tensorflow,ppwwyyxx/tensorflow,ychfan/tensorflow,yanchen036/tensorflow,annarev/tensorflow,cancan101/tensorflow,ageron/tensorflow,ppries/tensorflow,ppries/tensorflow,Kongsea/tensorflow,Carmezim/tensorflow,hfp/tensorflow-xsmm,manipopopo/tensorflow,alistairlow/tensorflow,adit-chandra/tensorflow,allenlavoie/tensorflow,chris-chris/tensorflow,zasdfgbnm/tensorflow,alisidd/tensorflow,zycdragonball/tensorflow,sarvex/tensorflow,vrv/tensorflow,zasdfgbnm/tensorflow,johndpope/tensorflow,arborh/tensorflow,code-sauce/tensorflow,manazhao/tf_recsys,eaplatanios/tensorflow,lukeiwanski/tensorflow,ychfan/tensorflow,gautam1858/tensorflow,sarvex/tensorflow,gibiansky/tensorflow,ppwwyyxx/tensorflow,sjperkins/tensorflow,mdrumond/tensorflow,adit-chandra/tensorflow,Moriadry/tensorflow,Bulochkin/tensorflow_pack,bowang/tensorflow,xzturn/tensorflow,aselle/tensorflow,nightjean/Deep-Learning,MycChiu/tensorflow,anand-c-goog/tensorflow,ibmsoe/tensorflow,thjashin/tensorflow,jhseu/tensorflow,AndreasMadsen/tensorflow,Bulochkin/tensorflow_pack,dongjoon-hyun/tensorflow,bowang/tensorflow,apark263/tensorflow,yanchen036/tensorflow,ArtsiomCh/tensorflow,ychfan/tensorflow,kevin-coder/tensorflow-fork,paolodedios/tensorflow,arborh/tensorflow,laszlocsomor/tensorflow,ravindrapanda/tensorflow,alshedivat/tensorflow,alisidd/tensorflow,petewarden/tensorflow,anand-c-goog/tensorflow,pierreg/tensorflow,raymondxyang/tensorflow,vrv/tensorflow,yaroslavvb/tensorflow,Mistobaan/tensorflow,hsaputra/tensorflow,strint/tensorflow,gnieboer/tensorflow,karllessard/tensorflow,vrv/tensorflow,ghchinoy/tensorflow,dancingdan/tensorflow,Bismarrck/tensorflow,zasdfgbnm/tensorflow,adamtiger/tensorflow,guschmue/tensorflow,jhseu/tensorflow,kchodorow/tensorflow,kchodorow/tensorflow,xodus7/tensorflow,haeusser/tensorflow,anilmuthineni/tensorflow,sandeepgupta2k4/tensorflow,suiyuan2009/tensorflow,sarvex/tensorflow,jalexvig/tensorflow,Bismarrck/tensorflow,eaplatanios/tensorflow,ZhangXinNan/tensorflow,aselle/tensorflow,alivecor/tensorflow,MycChiu/tensorflow,wangyum/tensorflow,jwlawson/tensorflow,adit-chandra/tensorflow,chris-chris/tensorflow,sandeepgupta2k4/tensorflow,jhseu/tensorflow,abhitopia/tensorflow,alshedivat/tensorflow,jart/tensorflow,hehongliang/tensorflow,yufengg/tensorflow,jeffzheng1/tensorflow,rdipietro/tensorflow,petewarden/tensorflow,ageron/tensorflow,sjperkins/tensorflow,aam-at/tensorflow,Bismarrck/tensorflow,seaotterman/tensorflow,horance-liu/tensorflow,adit-chandra/tensorflow,DCSaunders/tensorflow,eadgarchen/tensorflow,horance-liu/tensorflow,meteorcloudy/tensorflow,girving/tensorflow,zasdfgbnm/tensorflow,gojira/tensorflow,Mistobaan/tensorflow,Bulochkin/tensorflow_pack,xzturn/tensorflow,xzturn/tensorflow,Carmezim/tensorflow,Carmezim/tensorflow,ibmsoe/tensorflow,RapidApplicationDevelopment/tensorflow,with-git/tensorflow,kobejean/tensorflow,handroissuazo/tensorflow,davidzchen/tensorflow,hsaputra/tensorflow,thjashin/tensorflow,Intel-tensorflow/tensorflow,benoitsteiner/tensorflow-xsmm,thesuperzapper/tensorflow,hfp/tensorflow-xsmm,jart/tensorflow,tomasreimers/tensorflow-emscripten,MoamerEncsConcordiaCa/tensorflow,nolanliou/tensorflow,nightjean/Deep-Learning,aam-at/tensorflow,tensorflow/tensorflow-pywrap_tf_optimizer,chenjun0210/tensorflow,av8ramit/tensorflow,petewarden/tensorflow,mengxn/tensorflow,frreiss/tensorflow-fred,unsiloai/syntaxnet-ops-hack,chemelnucfin/tensorflow,ibmsoe/tensorflow,HKUST-SING/tensorflow,HKUST-SING/tensorflow,bowang/tensorflow,dancingdan/tensorflow,benoitsteiner/tensorflow,Mistobaan/tensorflow,llhe/tensorflow,andrewcmyers/tensorflow,snnn/tensorflow,kevin-coder/tensorflow-fork,asadziach/tensorflow,Carmezim/tensorflow,kevin-coder/tensorflow-fork,Moriadry/tensorflow,dendisuhubdy/tensorflow,mdrumond/tensorflow,caisq/tensorflow,manipopopo/tensorflow,ibmsoe/tensorflow,wangyum/tensorflow,adamtiger/tensorflow,dyoung418/tensorflow,ravindrapanda/tensorflow,kamcpp/tensorflow,chemelnucfin/tensorflow,pavelchristof/gomoku-ai,av8ramit/tensorflow,krikru/tensorflow-opencl,martinwicke/tensorflow,apark263/tensorflow,admcrae/tensorflow,calebfoss/tensorflow,handroissuazo/tensorflow,horance-liu/tensorflow,LUTAN/tensorflow,frreiss/tensorflow-fred,nanditav/15712-TensorFlow,mavenlin/tensorflow,nanditav/15712-TensorFlow,nolanliou/tensorflow,Intel-Corporation/tensorflow,mortada/tensorflow,adamtiger/tensorflow,arborh/tensorflow,manjunaths/tensorflow,kevin-coder/tensorflow-fork,Kongsea/tensorflow,gibiansky/tensorflow,theflofly/tensorflow,arborh/tensorflow,Carmezim/tensorflow,anilmuthineni/tensorflow,nolanliou/tensorflow,ghchinoy/tensorflow,gibiansky/tensorflow,DCSaunders/tensorflow,ishay2b/tensorflow,handroissuazo/tensorflow,Mistobaan/tensorflow,girving/tensorflow,cancan101/tensorflow,martinwicke/tensorflow,Bulochkin/tensorflow_pack,yaroslavvb/tensorflow,mixturemodel-flow/tensorflow,allenlavoie/tensorflow,tongwang01/tensorflow,aam-at/tensorflow,freedomtan/tensorflow,mavenlin/tensorflow,benoitsteiner/tensorflow-xsmm,jalexvig/tensorflow,alisidd/tensorflow,zasdfgbnm/tensorflow,RapidApplicationDevelopment/tensorflow,haeusser/tensorflow,gojira/tensorflow,manipopopo/tensorflow,elingg/tensorflow,martinwicke/tensorflow,tensorflow/tensorflow-pywrap_tf_optimizer,dyoung418/tensorflow,ville-k/tensorflow,benoitsteiner/tensorflow-opencl,eerwitt/tensorflow,jendap/tensorflow,ZhangXinNan/tensorflow,adit-chandra/tensorflow,jart/tensorflow,av8ramit/tensorflow,haeusser/tensorflow,johndpope/tensorflow,jwlawson/tensorflow,hehongliang/tensorflow,seanli9jan/tensorflow,seaotterman/tensorflow,renyi533/tensorflow,Bismarrck/tensorflow,thjashin/tensorflow,alsrgv/tensorflow,lukeiwanski/tensorflow-opencl,ZhangXinNan/tensorflow,jwlawson/tensorflow,ville-k/tensorflow,haeusser/tensorflow,JVillella/tensorflow,tensorflow/tensorflow,aam-at/tensorflow,ran5515/DeepDecision,jendap/tensorflow,manjunaths/tensorflow,mavenlin/tensorflow,rdipietro/tensorflow,Bismarrck/tensorflow,SnakeJenny/TensorFlow,seanli9jan/tensorflow,andrewcmyers/tensorflow,maciekcc/tensorflow,jwlawson/tensorflow,nikste/tensorflow,manjunaths/tensorflow,tiagofrepereira2012/tensorflow,dongjoon-hyun/tensorflow,maciekcc/tensorflow,aldian/tensorflow,admcrae/tensorflow,tensorflow/tensorflow-pywrap_saved_model,RapidApplicationDevelopment/tensorflow,dyoung418/tensorflow,yufengg/tensorflow,benoitsteiner/tensorflow-opencl,pcm17/tensorflow,HKUST-SING/tensorflow,eadgarchen/tensorflow,tensorflow/tensorflow-experimental_link_static_libraries_once,jbedorf/tensorflow,gnieboer/tensorflow,markslwong/tensorflow,rabipanda/tensorflow,aselle/tensorflow,sandeepgupta2k4/tensorflow,laszlocsomor/tensorflow,SnakeJenny/TensorFlow,karllessard/tensorflow,Xeralux/tensorflow,gunan/tensorflow,petewarden/tensorflow,apark263/tensorflow,gojira/tensorflow,vrv/tensorflow,paolodedios/tensorflow,tensorflow/tensorflow,drpngx/tensorflow,pavelchristof/gomoku-ai,Bismarrck/tensorflow,aam-at/tensorflow,mortada/tensorflow,code-sauce/tensorflow,hfp/tensorflow-xsmm,llhe/tensorflow,hfp/tensorflow-xsmm,Intel-Corporation/tensorflow,tomasreimers/tensorflow-emscripten,Xeralux/tensorflow,apark263/tensorflow,lakshayg/tensorflow,davidzchen/tensorflow,admcrae/tensorflow,nburn42/tensorflow,tensorflow/tensorflow-pywrap_tf_optimizer,raymondxyang/tensorflow,vrv/tensorflow,petewarden/tensorflow,ppwwyyxx/tensorflow,sjperkins/tensorflow,laszlocsomor/tensorflow,dancingdan/tensorflow,tensorflow/tensorflow-pywrap_saved_model,nightjean/Deep-Learning,DCSaunders/tensorflow,freedomtan/tensorflow,tensorflow/tensorflow-experimental_link_static_libraries_once,hfp/tensorflow-xsmm,aldian/tensorflow,benoitsteiner/tensorflow-xsmm,yaroslavvb/tensorflow,admcrae/tensorflow,scenarios/tensorflow,martinwicke/tensorflow,ZhangXinNan/tensorflow,yongtang/tensorflow,drpngx/tensorflow,manazhao/tf_recsys,tensorflow/tensorflow-pywrap_tf_optimizer,annarev/tensorflow,hsaputra/tensorflow,handroissuazo/tensorflow,frreiss/tensorflow-fred,unsiloai/syntaxnet-ops-hack,tensorflow/tensorflow-experimental_link_static_libraries_once,alshedivat/tensorflow,drpngx/tensorflow,apark263/tensorflow,lukeiwanski/tensorflow,kevin-coder/tensorflow-fork,alshedivat/tensorflow,mdrumond/tensorflow,whn09/tensorflow,brchiu/tensorflow,krikru/tensorflow-opencl,tensorflow/tensorflow,pierreg/tensorflow,sandeepgupta2k4/tensorflow,AnishShah/tensorflow,ppwwyyxx/tensorflow,aselle/tensorflow,manjunaths/tensorflow,drpngx/tensorflow,raymondxyang/tensorflow,wangyum/tensorflow,jalexvig/tensorflow,eaplatanios/tensorflow,tiagofrepereira2012/tensorflow,aselle/tensorflow,dancingdan/tensorflow,kevin-coder/tensorflow-fork,nightjean/Deep-Learning,davidzchen/tensorflow,brchiu/tensorflow,tensorflow/tensorflow-pywrap_tf_optimizer,yufengg/tensorflow,JVillella/tensorflow,dendisuhubdy/tensorflow,whn09/tensorflow,anand-c-goog/tensorflow,tensorflow/tensorflow,dancingdan/tensorflow,AnishShah/tensorflow,chenjun0210/tensorflow,jhaux/tensorflow,meteorcloudy/tensorflow,jendap/tensorflow,benoitsteiner/tensorflow,yufengg/tensorflow,manazhao/tf_recsys,admcrae/tensorflow,alshedivat/tensorflow,zycdragonball/tensorflow,nightjean/Deep-Learning,tensorflow/tensorflow,elingg/tensorflow,yanchen036/tensorflow,Bismarrck/tensorflow,renyi533/tensorflow,mengxn/tensorflow,jendap/tensorflow,aldian/tensorflow,ghchinoy/tensorflow,eaplatanios/tensorflow,haeusser/tensorflow,tensorflow/tensorflow-pywrap_saved_model,benoitsteiner/tensorflow,Xeralux/tensorflow,ArtsiomCh/tensorflow,jostep/tensorflow,JingJunYin/tensorflow,jendap/tensorflow,petewarden/tensorflow,odejesush/tensorflow,haeusser/tensorflow,ghchinoy/tensorflow,memo/tensorflow,adit-chandra/tensorflow,DavidNorman/tensorflow,gibiansky/tensorflow,MoamerEncsConcordiaCa/tensorflow,benoitsteiner/tensorflow,DavidNorman/tensorflow,yanchen036/tensorflow,Bulochkin/tensorflow_pack,jendap/tensorflow,Bulochkin/tensorflow_pack,ishay2b/tensorflow,mengxn/tensorflow,gunan/tensorflow,gnieboer/tensorflow,Intel-tensorflow/tensorflow,mavenlin/tensorflow,jhaux/tensorflow,RapidApplicationDevelopment/tensorflow,frreiss/tensorflow-fred,hsaputra/tensorflow,ville-k/tensorflow,gnieboer/tensorflow,nolanliou/tensorflow,Mazecreator/tensorflow,gojira/tensorflow,johndpope/tensorflow,HKUST-SING/tensorflow,snnn/tensorflow,jeffzheng1/tensorflow,mdrumond/tensorflow,odejesush/tensorflow,strint/tensorflow,snnn/tensorflow,jhaux/tensorflow,mengxn/tensorflow,codrut3/tensorflow,bowang/tensorflow,ZhangXinNan/tensorflow,tomasreimers/tensorflow-emscripten,tornadozou/tensorflow,tomasreimers/tensorflow-emscripten,dancingdan/tensorflow,unsiloai/syntaxnet-ops-hack,krikru/tensorflow-opencl,markslwong/tensorflow,xodus7/tensorflow,anand-c-goog/tensorflow,gibiansky/tensorflow,xzturn/tensorflow,tiagofrepereira2012/tensorflow,eaplatanios/tensorflow,scenarios/tensorflow,pcm17/tensorflow,jart/tensorflow,gunan/tensorflow,jbedorf/tensorflow,memo/tensorflow,bowang/tensorflow,a-doumoulakis/tensorflow,dongjoon-hyun/tensorflow,rabipanda/tensorflow,mortada/tensorflow,manipopopo/tensorflow,zycdragonball/tensorflow,brchiu/tensorflow,cancan101/tensorflow,guschmue/tensorflow,brchiu/tensorflow,tensorflow/tensorflow-pywrap_tf_optimizer,alheinecke/tensorflow-xsmm,whn09/tensorflow,rabipanda/tensorflow,anand-c-goog/tensorflow,yongtang/tensorflow,nikste/tensorflow,laosiaudi/tensorflow,benoitsteiner/tensorflow,manipopopo/tensorflow,kamcpp/tensorflow,apark263/tensorflow,lukeiwanski/tensorflow-opencl,tomasreimers/tensorflow-emscripten,ychfan/tensorflow,SnakeJenny/TensorFlow,manjunaths/tensorflow,asadziach/tensorflow,horance-liu/tensorflow,AndreasMadsen/tensorflow,guschmue/tensorflow,AnishShah/tensorflow,tensorflow/tensorflow,taknevski/tensorflow-xsmm,allenlavoie/tensorflow,asimshankar/tensorflow,chemelnucfin/tensorflow,theflofly/tensorflow,tensorflow/tensorflow-pywrap_tf_optimizer,tensorflow/tensorflow-experimental_link_static_libraries_once,kobejean/tensorflow,jhseu/tensorflow,thesuperzapper/tensorflow,kevin-coder/tensorflow-fork,alivecor/tensorflow,with-git/tensorflow,snnn/tensorflow,zasdfgbnm/tensorflow,handroissuazo/tensorflow,asadziach/tensorflow,whn09/tensorflow,anand-c-goog/tensorflow,asimshankar/tensorflow,jhseu/tensorflow,ychfan/tensorflow,dongjoon-hyun/tensorflow,ageron/tensorflow,guschmue/tensorflow,ychfan/tensorflow,hfp/tensorflow-xsmm,rdipietro/tensorflow,lukeiwanski/tensorflow,nolanliou/tensorflow,chenjun0210/tensorflow,suiyuan2009/tensorflow,tornadozou/tensorflow,Intel-tensorflow/tensorflow,mdrumond/tensorflow,kobejean/tensorflow,laosiaudi/tensorflow,guschmue/tensorflow,andrewcmyers/tensorflow,tntnatbry/tensorflow,horance-liu/tensorflow,renyi533/tensorflow,Kongsea/tensorflow,elingg/tensorflow,petewarden/tensorflow,yongtang/tensorflow,alivecor/tensorflow,scenarios/tensorflow,rabipanda/tensorflow,tensorflow/tensorflow-experimental_link_static_libraries_once,yongtang/tensorflow,cancan101/tensorflow,tensorflow/tensorflow-pywrap_saved_model,odejesush/tensorflow,ravindrapanda/tensorflow,SnakeJenny/TensorFlow,theflofly/tensorflow,alisidd/tensorflow,jhseu/tensorflow,whn09/tensorflow,Intel-tensorflow/tensorflow,ville-k/tensorflow,tensorflow/tensorflow-pywrap_saved_model,nolanliou/tensorflow,av8ramit/tensorflow,abhitopia/tensorflow,jwlawson/tensorflow,MoamerEncsConcordiaCa/tensorflow,aldian/tensorflow,DavidNorman/tensorflow,horance-liu/tensorflow,sandeepdsouza93/TensorFlow-15712,alistairlow/tensorflow,pavelchristof/gomoku-ai,markslwong/tensorflow,frreiss/tensorflow-fred,Intel-tensorflow/tensorflow,gojira/tensorflow,meteorcloudy/tensorflow,asimshankar/tensorflow,manazhao/tf_recsys,LUTAN/tensorflow,ghchinoy/tensorflow,alisidd/tensorflow,JVillella/tensorflow,lukeiwanski/tensorflow-opencl,yaroslavvb/tensorflow,seanli9jan/tensorflow,sandeepgupta2k4/tensorflow,aselle/tensorflow,pavelchristof/gomoku-ai,jwlawson/tensorflow,girving/tensorflow,anilmuthineni/tensorflow,petewarden/tensorflow,paolodedios/tensorflow,xodus7/tensorflow,jwlawson/tensorflow,gnieboer/tensorflow,kchodorow/tensorflow,dancingdan/tensorflow,johndpope/tensorflow,caisq/tensorflow,guschmue/tensorflow,with-git/tensorflow,laszlocsomor/tensorflow,alisidd/tensorflow,mortada/tensorflow,hsaputra/tensorflow,nolanliou/tensorflow,ppries/tensorflow,llhe/tensorflow,cxxgtxy/tensorflow,hehongliang/tensorflow,nburn42/tensorflow,lukeiwanski/tensorflow,JingJunYin/tensorflow,ran5515/DeepDecision,tongwang01/tensorflow,yaroslavvb/tensorflow,yufengg/tensorflow,DCSaunders/tensorflow,nikste/tensorflow,pierreg/tensorflow,AnishShah/tensorflow,seaotterman/tensorflow,laosiaudi/tensorflow,nightjean/Deep-Learning,alivecor/tensorflow,ville-k/tensorflow,benoitsteiner/tensorflow-opencl,SnakeJenny/TensorFlow,ppwwyyxx/tensorflow,jeffzheng1/tensorflow,LUTAN/tensorflow,apark263/tensorflow,kchodorow/tensorflow,MoamerEncsConcordiaCa/tensorflow,paolodedios/tensorflow,Bulochkin/tensorflow_pack,kamcpp/tensorflow,tntnatbry/tensorflow,tillahoffmann/tensorflow,tensorflow/tensorflow,seanli9jan/tensorflow,girving/tensorflow,Kongsea/tensorflow,dendisuhubdy/tensorflow,alshedivat/tensorflow,karllessard/tensorflow,taknevski/tensorflow-xsmm,nburn42/tensorflow,thjashin/tensorflow,rabipanda/tensorflow,sjperkins/tensorflow,benoitsteiner/tensorflow,asimshankar/tensorflow,laosiaudi/tensorflow,alheinecke/tensorflow-xsmm,lakshayg/tensorflow,scenarios/tensorflow,ibmsoe/tensorflow,a-doumoulakis/tensorflow,yongtang/tensorflow,sandeepdsouza93/TensorFlow-15712,codrut3/tensorflow,code-sauce/tensorflow,Intel-tensorflow/tensorflow,frreiss/tensorflow-fred,petewarden/tensorflow,Bismarrck/tensorflow,elingg/tensorflow,cancan101/tensorflow,JingJunYin/tensorflow,lukeiwanski/tensorflow-opencl,taknevski/tensorflow-xsmm,eaplatanios/tensorflow,nightjean/Deep-Learning,code-sauce/tensorflow,aldian/tensorflow,gautam1858/tensorflow,davidzchen/tensorflow,JingJunYin/tensorflow,Xeralux/tensorflow,ravindrapanda/tensorflow,caisq/tensorflow,code-sauce/tensorflow,paolodedios/tensorflow,jalexvig/tensorflow,jalexvig/tensorflow,MycChiu/tensorflow,thjashin/tensorflow,thjashin/tensorflow,pcm17/tensorflow,laszlocsomor/tensorflow,ppries/tensorflow,memo/tensorflow,seanli9jan/tensorflow,tiagofrepereira2012/tensorflow,laszlocsomor/tensorflow,Mistobaan/tensorflow,abhitopia/tensorflow,gautam1858/tensorflow,unsiloai/syntaxnet-ops-hack,lakshayg/tensorflow,tensorflow/tensorflow-pywrap_tf_optimizer,nburn42/tensorflow,tillahoffmann/tensorflow,jbedorf/tensorflow,Moriadry/tensorflow,tntnatbry/tensorflow,ville-k/tensorflow,taknevski/tensorflow-xsmm,wangyum/tensorflow,mdrumond/tensorflow,manipopopo/tensorflow,andrewcmyers/tensorflow,chenjun0210/tensorflow,raymondxyang/tensorflow,Mistobaan/tensorflow,drpngx/tensorflow,tensorflow/tensorflow,manjunaths/tensorflow,mengxn/tensorflow,thjashin/tensorflow,raymondxyang/tensorflow,gunan/tensorflow,sandeepgupta2k4/tensorflow,theflofly/tensorflow,ppwwyyxx/tensorflow,tongwang01/tensorflow,kobejean/tensorflow,cancan101/tensorflow,eadgarchen/tensorflow,Bulochkin/tensorflow_pack,AndreasMadsen/tensorflow,zasdfgbnm/tensorflow,pavelchristof/gomoku-ai,alistairlow/tensorflow,taknevski/tensorflow-xsmm,dendisuhubdy/tensorflow,rabipanda/tensorflow,asimshankar/tensorflow,asimshankar/tensorflow,aam-at/tensorflow,LUTAN/tensorflow,pcm17/tensorflow,alsrgv/tensorflow,dongjoon-hyun/tensorflow,ravindrapanda/tensorflow,meteorcloudy/tensorflow,benoitsteiner/tensorflow,handroissuazo/tensorflow,dendisuhubdy/tensorflow,hfp/tensorflow-xsmm,alistairlow/tensorflow,xodus7/tensorflow,jeffzheng1/tensorflow,sandeepdsouza93/TensorFlow-15712,hehongliang/tensorflow,RapidApplicationDevelopment/tensorflow,memo/tensorflow,chris-chris/tensorflow,xodus7/tensorflow,chris-chris/tensorflow,Intel-Corporation/tensorflow,pavelchristof/gomoku-ai,brchiu/tensorflow,sandeepdsouza93/TensorFlow-15712,AnishShah/tensorflow,jhseu/tensorflow,nanditav/15712-TensorFlow,mdrumond/tensorflow,HKUST-SING/tensorflow,Intel-tensorflow/tensorflow,aselle/tensorflow,frreiss/tensorflow-fred,DCSaunders/tensorflow,yaroslavvb/tensorflow,freedomtan/tensorflow,MoamerEncsConcordiaCa/tensorflow,yufengg/tensorflow,zycdragonball/tensorflow,annarev/tensorflow,allenlavoie/tensorflow,adamtiger/tensorflow,caisq/tensorflow,yanchen036/tensorflow,MycChiu/tensorflow,sandeepdsouza93/TensorFlow-15712,yongtang/tensorflow,laosiaudi/tensorflow,MycChiu/tensorflow,jeffzheng1/tensorflow,adamtiger/tensorflow,annarev/tensorflow,kchodorow/tensorflow,AndreasMadsen/tensorflow,johndpope/tensorflow,eadgarchen/tensorflow,drpngx/tensorflow,brchiu/tensorflow,abhitopia/tensorflow,ghchinoy/tensorflow,av8ramit/tensorflow,tillahoffmann/tensorflow,eerwitt/tensorflow,tomasreimers/tensorflow-emscripten,suiyuan2009/tensorflow,ran5515/DeepDecision,cancan101/tensorflow,MoamerEncsConcordiaCa/tensorflow,frreiss/tensorflow-fred,ppwwyyxx/tensorflow,mengxn/tensorflow,yongtang/tensorflow,gnieboer/tensorflow,sarvex/tensorflow,zasdfgbnm/tensorflow,dyoung418/tensorflow,cancan101/tensorflow,Moriadry/tensorflow,rabipanda/tensorflow,markslwong/tensorflow,yaroslavvb/tensorflow,karllessard/tensorflow,whn09/tensorflow,xodus7/tensorflow,yanchen036/tensorflow,apark263/tensorflow,tensorflow/tensorflow-pywrap_saved_model,scenarios/tensorflow,Moriadry/tensorflow,benoitsteiner/tensorflow,sjperkins/tensorflow,apark263/tensorflow,codrut3/tensorflow,ppwwyyxx/tensorflow,lukeiwanski/tensorflow,a-doumoulakis/tensorflow,gibiansky/tensorflow,jwlawson/tensorflow,asadziach/tensorflow,Kongsea/tensorflow,tntnatbry/tensorflow,johndpope/tensorflow,alshedivat/tensorflow,ghchinoy/tensorflow,jhaux/tensorflow,ZhangXinNan/tensorflow,theflofly/tensorflow,eadgarchen/tensorflow,freedomtan/tensorflow,RapidApplicationDevelopment/tensorflow,manjunaths/tensorflow,odejesush/tensorflow,ghchinoy/tensorflow,kamcpp/tensorflow,pcm17/tensorflow,ageron/tensorflow,gnieboer/tensorflow,yufengg/tensorflow,Intel-Corporation/tensorflow,alheinecke/tensorflow-xsmm,kamcpp/tensorflow,markslwong/tensorflow,benoitsteiner/tensorflow-xsmm,SnakeJenny/TensorFlow,kamcpp/tensorflow,hsaputra/tensorflow,codrut3/tensorflow,alsrgv/tensorflow,paolodedios/tensorflow,chris-chris/tensorflow,DavidNorman/tensorflow,with-git/tensorflow,davidzchen/tensorflow,brchiu/tensorflow,meteorcloudy/tensorflow,manjunaths/tensorflow,renyi533/tensorflow,freedomtan/tensorflow,nburn42/tensorflow,arborh/tensorflow,mengxn/tensorflow,freedomtan/tensorflow,DCSaunders/tensorflow,tntnatbry/tensorflow,av8ramit/tensorflow,dyoung418/tensorflow,jostep/tensorflow,kchodorow/tensorflow,sandeepgupta2k4/tensorflow,asimshankar/tensorflow,jendap/tensorflow,drpngx/tensorflow,ZhangXinNan/tensorflow,calebfoss/tensorflow,DavidNorman/tensorflow,AndreasMadsen/tensorflow,anilmuthineni/tensorflow,ville-k/tensorflow,nanditav/15712-TensorFlow,aam-at/tensorflow,kobejean/tensorflow,Moriadry/tensorflow,pierreg/tensorflow,johndpope/tensorflow,hsaputra/tensorflow,gautam1858/tensorflow,caisq/tensorflow,eadgarchen/tensorflow,jeffzheng1/tensorflow,maciekcc/tensorflow,odejesush/tensorflow,llhe/tensorflow,cxxgtxy/tensorflow,tillahoffmann/tensorflow,manipopopo/tensorflow,jbedorf/tensorflow,haeusser/tensorflow,alsrgv/tensorflow,ppwwyyxx/tensorflow,lukeiwanski/tensorflow-opencl,Xeralux/tensorflow,laosiaudi/tensorflow,lakshayg/tensorflow,memo/tensorflow,brchiu/tensorflow,xodus7/tensorflow,andrewcmyers/tensorflow,benoitsteiner/tensorflow,eadgarchen/tensorflow,benoitsteiner/tensorflow-xsmm,jhaux/tensorflow,benoitsteiner/tensorflow-opencl,renyi533/tensorflow,aselle/tensorflow,MoamerEncsConcordiaCa/tensorflow,codrut3/tensorflow,adit-chandra/tensorflow,tongwang01/tensorflow,asadziach/tensorflow,Mazecreator/tensorflow,gautam1858/tensorflow,jostep/tensorflow,ran5515/DeepDecision,cxxgtxy/tensorflow,Intel-Corporation/tensorflow,jostep/tensorflow,gojira/tensorflow,dongjoon-hyun/tensorflow,dendisuhubdy/tensorflow,nanditav/15712-TensorFlow,Xeralux/tensorflow,jart/tensorflow,paolodedios/tensorflow,sandeepgupta2k4/tensorflow,lukeiwanski/tensorflow-opencl,admcrae/tensorflow,sarvex/tensorflow,elingg/tensorflow,gunan/tensorflow,allenlavoie/tensorflow,davidzchen/tensorflow,ArtsiomCh/tensorflow,eerwitt/tensorflow,tillahoffmann/tensorflow,girving/tensorflow,Intel-Corporation/tensorflow,codrut3/tensorflow,mixturemodel-flow/tensorflow,manazhao/tf_recsys,dendisuhubdy/tensorflow,MoamerEncsConcordiaCa/tensorflow,lukeiwanski/tensorflow,maciekcc/tensorflow,LUTAN/tensorflow,JingJunYin/tensorflow,xzturn/tensorflow,LUTAN/tensorflow,rdipietro/tensorflow,tensorflow/tensorflow,frreiss/tensorflow-fred,tiagofrepereira2012/tensorflow,AndreasMadsen/tensorflow,zycdragonball/tensorflow,brchiu/tensorflow,jhaux/tensorflow,dongjoon-hyun/tensorflow,yaroslavvb/tensorflow,thjashin/tensorflow,nikste/tensorflow,drpngx/tensorflow,arborh/tensorflow,allenlavoie/tensorflow,mortada/tensorflow,taknevski/tensorflow-xsmm,snnn/tensorflow,maciekcc/tensorflow,RapidApplicationDevelopment/tensorflow,tornadozou/tensorflow,gnieboer/tensorflow,alsrgv/tensorflow,memo/tensorflow,memo/tensorflow,girving/tensorflow,ychfan/tensorflow,eaplatanios/tensorflow,markslwong/tensorflow,xzturn/tensorflow,benoitsteiner/tensorflow-xsmm,snnn/tensorflow,alsrgv/tensorflow,renyi533/tensorflow,Xeralux/tensorflow,taknevski/tensorflow-xsmm,annarev/tensorflow,strint/tensorflow,dongjoon-hyun/tensorflow,meteorcloudy/tensorflow,rabipanda/tensorflow,anilmuthineni/tensorflow,anand-c-goog/tensorflow,jbedorf/tensorflow,freedomtan/tensorflow,alshedivat/tensorflow,thesuperzapper/tensorflow,alheinecke/tensorflow-xsmm,jhseu/tensorflow,Mazecreator/tensorflow,tensorflow/tensorflow-experimental_link_static_libraries_once,drpngx/tensorflow,lukeiwanski/tensorflow-opencl,brchiu/tensorflow,av8ramit/tensorflow,aselle/tensorflow,tornadozou/tensorflow,ppries/tensorflow,cxxgtxy/tensorflow,calebfoss/tensorflow,jbedorf/tensorflow,petewarden/tensorflow,alsrgv/tensorflow,alsrgv/tensorflow,gunan/tensorflow,yanchen036/tensorflow,Bismarrck/tensorflow,Intel-Corporation/tensorflow,alheinecke/tensorflow-xsmm,scenarios/tensorflow,rdipietro/tensorflow,benoitsteiner/tensorflow-opencl,mavenlin/tensorflow,johndpope/tensorflow,jalexvig/tensorflow,MycChiu/tensorflow,Mistobaan/tensorflow,RapidApplicationDevelopment/tensorflow,yaroslavvb/tensorflow,eaplatanios/tensorflow,jhseu/tensorflow,pierreg/tensorflow,theflofly/tensorflow,asadziach/tensorflow,ppwwyyxx/tensorflow,hsaputra/tensorflow,ran5515/DeepDecision,mortada/tensorflow,pierreg/tensorflow,allenlavoie/tensorflow,manipopopo/tensorflow,jwlawson/tensorflow,zycdragonball/tensorflow,tomasreimers/tensorflow-emscripten,ravindrapanda/tensorflow,eaplatanios/tensorflow,Kongsea/tensorflow,MycChiu/tensorflow,alsrgv/tensorflow,mixturemodel-flow/tensorflow,benoitsteiner/tensorflow-opencl,kevin-coder/tensorflow-fork,wangyum/tensorflow,jhaux/tensorflow,meteorcloudy/tensorflow,odejesush/tensorflow,gojira/tensorflow,SnakeJenny/TensorFlow,asimshankar/tensorflow,with-git/tensorflow,frreiss/tensorflow-fred,strint/tensorflow,laosiaudi/tensorflow,lakshayg/tensorflow,nikste/tensorflow,jhaux/tensorflow,handroissuazo/tensorflow,admcrae/tensorflow,wangyum/tensorflow,codrut3/tensorflow,zasdfgbnm/tensorflow,dancingdan/tensorflow,horance-liu/tensorflow,gojira/tensorflow,nikste/tensorflow,vrv/tensorflow,ghchinoy/tensorflow,code-sauce/tensorflow,freedomtan/tensorflow,annarev/tensorflow,elingg/tensorflow,krikru/tensorflow-opencl,Mazecreator/tensorflow,nburn42/tensorflow,martinwicke/tensorflow,petewarden/tensorflow,whn09/tensorflow,Xeralux/tensorflow,ppries/tensorflow,tongwang01/tensorflow,caisq/tensorflow,llhe/tensorflow,jostep/tensorflow,calebfoss/tensorflow,ZhangXinNan/tensorflow,benoitsteiner/tensorflow-xsmm,strint/tensorflow,martinwicke/tensorflow,pierreg/tensorflow,johndpope/tensorflow,tensorflow/tensorflow,av8ramit/tensorflow,hfp/tensorflow-xsmm,tensorflow/tensorflow,chemelnucfin/tensorflow,ageron/tensorflow,gunan/tensorflow,jart/tensorflow,RapidApplicationDevelopment/tensorflow,code-sauce/tensorflow,elingg/tensorflow,tiagofrepereira2012/tensorflow,nightjean/Deep-Learning,pcm17/tensorflow,Carmezim/tensorflow,kobejean/tensorflow,eerwitt/tensorflow,Xeralux/tensorflow,lukeiwanski/tensorflow,llhe/tensorflow,jendap/tensorflow,seaotterman/tensorflow,allenlavoie/tensorflow,dendisuhubdy/tensorflow,memo/tensorflow,Bulochkin/tensorflow_pack,pcm17/tensorflow,suiyuan2009/tensorflow,kamcpp/tensorflow,ppwwyyxx/tensorflow,AnishShah/tensorflow,annarev/tensorflow,manipopopo/tensorflow,gunan/tensorflow,alivecor/tensorflow,AnishShah/tensorflow,llhe/tensorflow,tensorflow/tensorflow-pywrap_tf_optimizer,bowang/tensorflow,codrut3/tensorflow,eerwitt/tensorflow,davidzchen/tensorflow,benoitsteiner/tensorflow-opencl,alistairlow/tensorflow,HKUST-SING/tensorflow,jhaux/tensorflow,ravindrapanda/tensorflow,kchodorow/tensorflow,xzturn/tensorflow,a-doumoulakis/tensorflow,laszlocsomor/tensorflow,ibmsoe/tensorflow,johndpope/tensorflow,vrv/tensorflow,mixturemodel-flow/tensorflow,asadziach/tensorflow,jalexvig/tensorflow,gojira/tensorflow,calebfoss/tensorflow,dyoung418/tensorflow,alheinecke/tensorflow-xsmm,yongtang/tensorflow,jart/tensorflow,asimshankar/tensorflow,jwlawson/tensorflow,odejesush/tensorflow,ZhangXinNan/tensorflow,mixturemodel-flow/tensorflow,anilmuthineni/tensorflow,eaplatanios/tensorflow,kobejean/tensorflow,gibiansky/tensorflow,seaotterman/tensorflow,frreiss/tensorflow-fred,ishay2b/tensorflow,renyi533/tensorflow,tornadozou/tensorflow,Kongsea/tensorflow,sandeepdsouza93/TensorFlow-15712,thesuperzapper/tensorflow,jendap/tensorflow,pcm17/tensorflow,tensorflow/tensorflow-experimental_link_static_libraries_once,jart/tensorflow,gautam1858/tensorflow,kevin-coder/tensorflow-fork,tntnatbry/tensorflow,adit-chandra/tensorflow,ArtsiomCh/tensorflow,krikru/tensorflow-opencl,elingg/tensorflow,gibiansky/tensorflow,ychfan/tensorflow,martinwicke/tensorflow,chris-chris/tensorflow,mortada/tensorflow,AnishShah/tensorflow,jhseu/tensorflow,ageron/tensorflow,xzturn/tensorflow,jeffzheng1/tensorflow,jwlawson/tensorflow,theflofly/tensorflow,tensorflow/tensorflow-pywrap_tf_optimizer,Bulochkin/tensorflow_pack,dendisuhubdy/tensorflow,guschmue/tensorflow,jalexvig/tensorflow,laszlocsomor/tensorflow,manjunaths/tensorflow,ArtsiomCh/tensorflow,ville-k/tensorflow,Intel-tensorflow/tensorflow,ville-k/tensorflow,tillahoffmann/tensorflow,tensorflow/tensorflow-pywrap_saved_model,av8ramit/tensorflow,ageron/tensorflow,mdrumond/tensorflow,raymondxyang/tensorflow,handroissuazo/tensorflow,mixturemodel-flow/tensorflow,ishay2b/tensorflow,sjperkins/tensorflow,pcm17/tensorflow,renyi533/tensorflow,renyi533/tensorflow,chemelnucfin/tensorflow,ageron/tensorflow,tomasreimers/tensorflow-emscripten,paolodedios/tensorflow,pavelchristof/gomoku-ai,nikste/tensorflow,annarev/tensorflow,thesuperzapper/tensorflow,frreiss/tensorflow-fred,strint/tensorflow,chemelnucfin/tensorflow,gautam1858/tensorflow,gunan/tensorflow,JingJunYin/tensorflow,abhitopia/tensorflow,suiyuan2009/tensorflow,hfp/tensorflow-xsmm,lukeiwanski/tensorflow-opencl,alistairlow/tensorflow,gautam1858/tensorflow,cxxgtxy/tensorflow,DCSaunders/tensorflow,hsaputra/tensorflow,tensorflow/tensorflow-pywrap_saved_model,freedomtan/tensorflow,Mazecreator/tensorflow,chemelnucfin/tensorflow,martinwicke/tensorflow,ravindrapanda/tensorflow,nanditav/15712-TensorFlow,meteorcloudy/tensorflow,ville-k/tensorflow,ravindrapanda/tensorflow,calebfoss/tensorflow,freedomtan/tensorflow,davidzchen/tensorflow,paolodedios/tensorflow,jart/tensorflow,scenarios/tensorflow,jeffzheng1/tensorflow,Mazecreator/tensorflow,Mistobaan/tensorflow,abhitopia/tensorflow,petewarden/tensorflow,asadziach/tensorflow,alisidd/tensorflow,guschmue/tensorflow,anand-c-goog/tensorflow,arborh/tensorflow,MoamerEncsConcordiaCa/tensorflow,aselle/tensorflow,lakshayg/tensorflow,seanli9jan/tensorflow,Intel-tensorflow/tensorflow,a-doumoulakis/tensorflow,jostep/tensorflow,thjashin/tensorflow,tiagofrepereira2012/tensorflow,horance-liu/tensorflow,davidzchen/tensorflow,lukeiwanski/tensorflow-opencl,dendisuhubdy/tensorflow,sandeepgupta2k4/tensorflow,guschmue/tensorflow,anilmuthineni/tensorflow,jbedorf/tensorflow,DavidNorman/tensorflow,manipopopo/tensorflow,with-git/tensorflow,snnn/tensorflow,abhitopia/tensorflow,rabipanda/tensorflow,nanditav/15712-TensorFlow,xodus7/tensorflow,paolodedios/tensorflow,alshedivat/tensorflow,a-doumoulakis/tensorflow,tntnatbry/tensorflow,strint/tensorflow,zasdfgbnm/tensorflow,jbedorf/tensorflow,suiyuan2009/tensorflow,Carmezim/tensorflow,unsiloai/syntaxnet-ops-hack,eadgarchen/tensorflow,markslwong/tensorflow,dancingdan/tensorflow,dongjoon-hyun/tensorflow,gautam1858/tensorflow,jbedorf/tensorflow,tongwang01/tensorflow,Bulochkin/tensorflow_pack,eadgarchen/tensorflow,aselle/tensorflow,aam-at/tensorflow,Mistobaan/tensorflow,bowang/tensorflow,scenarios/tensorflow,alisidd/tensorflow,DavidNorman/tensorflow,markslwong/tensorflow,odejesush/tensorflow,seanli9jan/tensorflow,mixturemodel-flow/tensorflow,ZhangXinNan/tensorflow,seaotterman/tensorflow,nburn42/tensorflow,renyi533/tensorflow,Carmezim/tensorflow,bowang/tensorflow,adamtiger/tensorflow,theflofly/tensorflow,arborh/tensorflow,nburn42/tensorflow,gibiansky/tensorflow,odejesush/tensorflow,abhitopia/tensorflow,karllessard/tensorflow,calebfoss/tensorflow,dongjoon-hyun/tensorflow,sandeepdsouza93/TensorFlow-15712,sarvex/tensorflow,andrewcmyers/tensorflow,Moriadry/tensorflow,JingJunYin/tensorflow,elingg/tensorflow,raymondxyang/tensorflow,sandeepdsouza93/TensorFlow-15712,kamcpp/tensorflow,alivecor/tensorflow,asadziach/tensorflow,SnakeJenny/TensorFlow,yongtang/tensorflow,annarev/tensorflow,laosiaudi/tensorflow,admcrae/tensorflow,av8ramit/tensorflow,DCSaunders/tensorflow,seaotterman/tensorflow,gunan/tensorflow,seaotterman/tensorflow,yongtang/tensorflow,DavidNorman/tensorflow,xzturn/tensorflow,JVillella/tensorflow,thesuperzapper/tensorflow,ZhangXinNan/tensorflow,gautam1858/tensorflow,llhe/tensorflow,mengxn/tensorflow,kchodorow/tensorflow,caisq/tensorflow,zasdfgbnm/tensorflow,dancingdan/tensorflow,arborh/tensorflow,freedomtan/tensorflow,yongtang/tensorflow,girving/tensorflow,gojira/tensorflow,xzturn/tensorflow,ghchinoy/tensorflow,taknevski/tensorflow-xsmm,chris-chris/tensorflow,adit-chandra/tensorflow,whn09/tensorflow,seanli9jan/tensorflow,davidzchen/tensorflow,JVillella/tensorflow,jbedorf/tensorflow,ppwwyyxx/tensorflow,gunan/tensorflow,haeusser/tensorflow,JVillella/tensorflow,wangyum/tensorflow,xodus7/tensorflow,anand-c-goog/tensorflow,tntnatbry/tensorflow,thesuperzapper/tensorflow,aam-at/tensorflow,adamtiger/tensorflow,Mazecreator/tensorflow,karllessard/tensorflow,jendap/tensorflow,ran5515/DeepDecision,freedomtan/tensorflow,abhitopia/tensorflow,gunan/tensorflow,jbedorf/tensorflow,manazhao/tf_recsys,tillahoffmann/tensorflow,laosiaudi/tensorflow,alshedivat/tensorflow,whn09/tensorflow,ghchinoy/tensorflow,girving/tensorflow,chenjun0210/tensorflow,apark263/tensorflow,eadgarchen/tensorflow,sjperkins/tensorflow,mortada/tensorflow,tensorflow/tensorflow-experimental_link_static_libraries_once,tiagofrepereira2012/tensorflow,LUTAN/tensorflow,asimshankar/tensorflow,nikste/tensorflow,thesuperzapper/tensorflow,DCSaunders/tensorflow,ppries/tensorflow,allenlavoie/tensorflow,tornadozou/tensorflow,ageron/tensorflow,allenlavoie/tensorflow,girving/tensorflow,pierreg/tensorflow,ibmsoe/tensorflow,suiyuan2009/tensorflow,chemelnucfin/tensorflow,snnn/tensorflow,tensorflow/tensorflow-experimental_link_static_libraries_once,manazhao/tf_recsys,kobejean/tensorflow,andrewcmyers/tensorflow,kobejean/tensorflow,cxxgtxy/tensorflow,hehongliang/tensorflow,vrv/tensorflow,alheinecke/tensorflow-xsmm,Mistobaan/tensorflow,maciekcc/tensorflow,maciekcc/tensorflow,aldian/tensorflow,AnishShah/tensorflow,calebfoss/tensorflow,wangyum/tensorflow,seaotterman/tensorflow,lakshayg/tensorflow,calebfoss/tensorflow,nanditav/15712-TensorFlow,davidzchen/tensorflow,cxxgtxy/tensorflow,tillahoffmann/tensorflow,dancingdan/tensorflow,davidzchen/tensorflow,jbedorf/tensorflow,codrut3/tensorflow,DavidNorman/tensorflow,sjperkins/tensorflow,aam-at/tensorflow,horance-liu/tensorflow,Moriadry/tensorflow,DavidNorman/tensorflow,LUTAN/tensorflow,ageron/tensorflow,xzturn/tensorflow,rdipietro/tensorflow,annarev/tensorflow,DavidNorman/tensorflow,chenjun0210/tensorflow,jalexvig/tensorflow,ran5515/DeepDecision,adit-chandra/tensorflow,gnieboer/tensorflow,ArtsiomCh/tensorflow,snnn/tensorflow,with-git/tensorflow,AndreasMadsen/tensorflow,llhe/tensorflow,benoitsteiner/tensorflow,vrv/tensorflow,Intel-tensorflow/tensorflow,eerwitt/tensorflow,mixturemodel-flow/tensorflow,strint/tensorflow,ppries/tensorflow,tomasreimers/tensorflow-emscripten,Mistobaan/tensorflow,nburn42/tensorflow,apark263/tensorflow,eerwitt/tensorflow,kobejean/tensorflow,chemelnucfin/tensorflow,jostep/tensorflow,scenarios/tensorflow,xzturn/tensorflow,asimshankar/tensorflow,unsiloai/syntaxnet-ops-hack,jostep/tensorflow,rabipanda/tensorflow,ishay2b/tensorflow | shell | ## Code Before:
set -e
# Install dependencies from ubuntu deb repository.
apt-get update
apt-get install -y --no-install-recommends \
autoconf \
automake \
build-essential \
cmake \
curl \
ffmpeg \
git \
libcurl4-openssl-dev \
libtool \
openjdk-8-jdk \
openjdk-8-jre-headless \
pkg-config \
python-dev \
python-pip \
python-virtualenv \
python3-dev \
python3-pip \
rsync \
sudo \
swig \
unzip \
wget \
zip \
zlib1g-dev
apt-get clean
rm -rf /var/lib/apt/lists/*
## Instruction:
Fix certificate issues in bazel for ci build.
(cherry picked from commit 923c999d6e910d7566f10a852d2ff12c4a88fad8)
## Code After:
set -e
# Install dependencies from ubuntu deb repository.
apt-get update
apt-get install -y --no-install-recommends \
autoconf \
automake \
build-essential \
cmake \
curl \
ffmpeg \
git \
libcurl4-openssl-dev \
libtool \
openjdk-8-jdk \
openjdk-8-jre-headless \
pkg-config \
python-dev \
python-pip \
python-virtualenv \
python3-dev \
python3-pip \
rsync \
sudo \
swig \
unzip \
wget \
zip \
zlib1g-dev
# Install ca-certificates, and update the certificate store.
apt-get install ca-certificates-java
update-ca-certificates -f
apt-get clean
rm -rf /var/lib/apt/lists/*
|
set -e
# Install dependencies from ubuntu deb repository.
apt-get update
apt-get install -y --no-install-recommends \
autoconf \
automake \
build-essential \
cmake \
curl \
ffmpeg \
git \
libcurl4-openssl-dev \
libtool \
openjdk-8-jdk \
openjdk-8-jre-headless \
pkg-config \
python-dev \
python-pip \
python-virtualenv \
python3-dev \
python3-pip \
rsync \
sudo \
swig \
unzip \
wget \
zip \
zlib1g-dev
+
+ # Install ca-certificates, and update the certificate store.
+ apt-get install ca-certificates-java
+ update-ca-certificates -f
+
apt-get clean
rm -rf /var/lib/apt/lists/* | 5 | 0.151515 | 5 | 0 |
830a9b0111ba612109b204ce1b414b11b4946254 | src/main/java/com/github/pedrovgs/problem69/BitsToTransform.java | src/main/java/com/github/pedrovgs/problem69/BitsToTransform.java | /*
* Copyright (C) 2014 Pedro Vicente Gómez Sánchez.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package com.github.pedrovgs.problem69;
/**
* Given two integers passed as argument, can you write a function to determine the number of bits
* required to convert integer A to integer B.
*
* @author Pedro Vicente Gómez Sánchez.
*/
public class BitsToTransform {
}
| /*
* Copyright (C) 2014 Pedro Vicente Gómez Sánchez.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package com.github.pedrovgs.problem69;
/**
* Given two integers passed as argument, can you write a function to determine the number of bits
* required to convert integer A to integer B.
*
* @author Pedro Vicente Gómez Sánchez.
*/
public class BitsToTransform {
/**
* Iterative algorithm created to calculate a XOR using the input integers. Once you have the XOR
* you can iterate through the XOR to count the number of bits equals to 1. To iterate over the
* integer we are going to modify the value of the XOR integer using & operator with XOR & XOR
* -1.
*
* For example:
*
* numA = 001
* numB = 101
* XOR = 100
*
* Iteration 1 -> XOR = 100
* Iteration 2 -> XOR = 100 & 011 = 000 // Loop body not executed.
*/
public int calculate(int numA, int numB) {
int differentBits = 0;
for (int XOR = numA ^ numB; XOR != 0; XOR = XOR & (XOR - 1)) {
differentBits++;
}
return differentBits;
}
}
| Add solution to problem 69 | Add solution to problem 69
| Java | apache-2.0 | sridhar-newsdistill/Algorithms,mrgenco/Algorithms-1,Arkar-Aung/Algorithms,ajinkyakolhe112/Algorithms,chengjinqian/Algorithms,007slm/Algorithms,Ariloum/Algorithms,VeskoI/Algorithms,pedrovgs/Algorithms,JeffreyWei/Algorithms,inexistence/Algorithms,zmywly8866/Algorithms,ArlenLiu/Algorithms,zhdh2008/Algorithms,AppScientist/Algorithms,jibaro/Algorithms | java | ## Code Before:
/*
* Copyright (C) 2014 Pedro Vicente Gómez Sánchez.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package com.github.pedrovgs.problem69;
/**
* Given two integers passed as argument, can you write a function to determine the number of bits
* required to convert integer A to integer B.
*
* @author Pedro Vicente Gómez Sánchez.
*/
public class BitsToTransform {
}
## Instruction:
Add solution to problem 69
## Code After:
/*
* Copyright (C) 2014 Pedro Vicente Gómez Sánchez.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package com.github.pedrovgs.problem69;
/**
* Given two integers passed as argument, can you write a function to determine the number of bits
* required to convert integer A to integer B.
*
* @author Pedro Vicente Gómez Sánchez.
*/
public class BitsToTransform {
/**
* Iterative algorithm created to calculate a XOR using the input integers. Once you have the XOR
* you can iterate through the XOR to count the number of bits equals to 1. To iterate over the
* integer we are going to modify the value of the XOR integer using & operator with XOR & XOR
* -1.
*
* For example:
*
* numA = 001
* numB = 101
* XOR = 100
*
* Iteration 1 -> XOR = 100
* Iteration 2 -> XOR = 100 & 011 = 000 // Loop body not executed.
*/
public int calculate(int numA, int numB) {
int differentBits = 0;
for (int XOR = numA ^ numB; XOR != 0; XOR = XOR & (XOR - 1)) {
differentBits++;
}
return differentBits;
}
}
| /*
* Copyright (C) 2014 Pedro Vicente Gómez Sánchez.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package com.github.pedrovgs.problem69;
/**
* Given two integers passed as argument, can you write a function to determine the number of bits
* required to convert integer A to integer B.
*
* @author Pedro Vicente Gómez Sánchez.
*/
public class BitsToTransform {
+ /**
+ * Iterative algorithm created to calculate a XOR using the input integers. Once you have the XOR
+ * you can iterate through the XOR to count the number of bits equals to 1. To iterate over the
+ * integer we are going to modify the value of the XOR integer using & operator with XOR & XOR
+ * -1.
+ *
+ * For example:
+ *
+ * numA = 001
+ * numB = 101
+ * XOR = 100
+ *
+ * Iteration 1 -> XOR = 100
+ * Iteration 2 -> XOR = 100 & 011 = 000 // Loop body not executed.
+ */
+ public int calculate(int numA, int numB) {
+ int differentBits = 0;
+ for (int XOR = numA ^ numB; XOR != 0; XOR = XOR & (XOR - 1)) {
+ differentBits++;
+ }
+ return differentBits;
+ }
} | 22 | 0.846154 | 22 | 0 |
67eb9de5d4fef31746c1112b68a3ef5825e7beaf | app/services/document_parser.rb | app/services/document_parser.rb | require 'open3'
class DocumentParser
attr_accessor :title, :body
def initialize(filename)
@filename = filename
to_hash
end
def tika_path
Rails.root.join('vendor/java/tika-app-1.4.jar')
end
def output format = :text
format_argument = format == :html ? "-h" : "-t"
stdout, errors, _ = Open3.capture3 "java -jar #{tika_path} #{format_argument} #{@filename}"
stdout
end
def title
output.lines.detect { |l| l.present? }.strip
end
def body
Nokogiri::HTML(output :html).css('body').children[1..-1].to_html.strip
end
def to_hash
{ title: title, body: body }
end
class << self
def parse(filename)
new(filename).to_hash
end
end
end
| require 'open3'
class DocumentParser
def initialize(filename)
@filename = filename
to_hash
end
def tika_path
Rails.root.join('vendor/java/tika-app-1.4.jar')
end
def output format = :text
format_argument = format == :html ? "-h" : "-t"
stdout, errors, _ = Open3.capture3 "java -jar #{tika_path} #{format_argument} #{@filename}"
stdout
end
def title
output.lines.detect { |l| l.present? }.strip
end
def body
Nokogiri::HTML(output :html).css('body').children[1..-1].to_html.strip
end
def to_hash
{ title: title, body: body }
end
class << self
def parse(filename)
new(filename).to_hash
end
end
end
| Remove attr_accessor as it's not needed | Remove attr_accessor as it's not needed
| Ruby | mit | johan--/tahi,johan--/tahi,johan--/tahi,johan--/tahi | ruby | ## Code Before:
require 'open3'
class DocumentParser
attr_accessor :title, :body
def initialize(filename)
@filename = filename
to_hash
end
def tika_path
Rails.root.join('vendor/java/tika-app-1.4.jar')
end
def output format = :text
format_argument = format == :html ? "-h" : "-t"
stdout, errors, _ = Open3.capture3 "java -jar #{tika_path} #{format_argument} #{@filename}"
stdout
end
def title
output.lines.detect { |l| l.present? }.strip
end
def body
Nokogiri::HTML(output :html).css('body').children[1..-1].to_html.strip
end
def to_hash
{ title: title, body: body }
end
class << self
def parse(filename)
new(filename).to_hash
end
end
end
## Instruction:
Remove attr_accessor as it's not needed
## Code After:
require 'open3'
class DocumentParser
def initialize(filename)
@filename = filename
to_hash
end
def tika_path
Rails.root.join('vendor/java/tika-app-1.4.jar')
end
def output format = :text
format_argument = format == :html ? "-h" : "-t"
stdout, errors, _ = Open3.capture3 "java -jar #{tika_path} #{format_argument} #{@filename}"
stdout
end
def title
output.lines.detect { |l| l.present? }.strip
end
def body
Nokogiri::HTML(output :html).css('body').children[1..-1].to_html.strip
end
def to_hash
{ title: title, body: body }
end
class << self
def parse(filename)
new(filename).to_hash
end
end
end
| require 'open3'
class DocumentParser
- attr_accessor :title, :body
-
def initialize(filename)
@filename = filename
to_hash
end
def tika_path
Rails.root.join('vendor/java/tika-app-1.4.jar')
end
def output format = :text
format_argument = format == :html ? "-h" : "-t"
stdout, errors, _ = Open3.capture3 "java -jar #{tika_path} #{format_argument} #{@filename}"
stdout
end
def title
output.lines.detect { |l| l.present? }.strip
end
def body
Nokogiri::HTML(output :html).css('body').children[1..-1].to_html.strip
end
def to_hash
{ title: title, body: body }
end
class << self
def parse(filename)
new(filename).to_hash
end
end
end | 2 | 0.052632 | 0 | 2 |
7624c9323f8b01329408e2ca72a6e6a2a1478033 | spec/spec_helper.rb | spec/spec_helper.rb | def add_to_load_path(path, prepend=false)
path = File.expand_path("../#{path}", __FILE__)
if prepend
$LOAD_PATH.unshift(path) unless $LOAD_PATH.include?(path)
else
$LOAD_PATH << path unless $LOAD_PATH.include?(path)
end
end
add_to_load_path("../lib", :prepend)
add_to_load_path("../../rspec-core/lib")
add_to_load_path("../../rspec-mocks/lib")
require 'rspec/core'
require 'rspec/mocks'
Dir['./spec/support/**/*'].each do |f|
require f
end
def with_ruby(version)
yield if RUBY_PLATFORM =~ Regexp.compile("^#{version}")
end
module RSpec
module Ruby
class << self
def version
RUBY_VERSION
end
end
end
end
module RSpec
module Matchers
def fail
raise_error(RSpec::Expectations::ExpectationNotMetError)
end
def fail_with(message)
raise_error(RSpec::Expectations::ExpectationNotMetError, message)
end
end
end
RSpec::configure do |config|
config.mock_with(:rspec)
config.include RSpec::Mocks::Methods
config.color_enabled = true
end
| def add_to_load_path(path, prepend=false)
path = File.expand_path("../#{path}", __FILE__)
if prepend
$LOAD_PATH.unshift(path) unless $LOAD_PATH.include?(path)
else
$LOAD_PATH << path unless $LOAD_PATH.include?(path)
end
end
add_to_load_path("../lib", :prepend)
add_to_load_path("../../rspec-core/lib")
add_to_load_path("../../rspec-mocks/lib")
require 'rspec/core'
require 'rspec/mocks'
Dir['./spec/support/**/*'].each do |f|
require f
end
def with_ruby(version)
yield if RUBY_PLATFORM =~ Regexp.compile("^#{version}")
end
module RSpec
module Ruby
class << self
def version
RUBY_VERSION
end
end
end
end
module RSpec
module Matchers
def fail
raise_error(RSpec::Expectations::ExpectationNotMetError)
end
def fail_with(message)
raise_error(RSpec::Expectations::ExpectationNotMetError, message)
end
end
end
RSpec::configure do |config|
config.mock_with(:rspec)
config.include RSpec::Mocks::Methods
config.color_enabled = true
config.filter_run :focused => true
config.run_all_when_everything_filtered = true
end
| Add filtering for self-spec suite. | Add filtering for self-spec suite.
| Ruby | mit | zsyed91/rspec-expectations,jdax/rspec-expectations,emilyforst/rspec-expectations,h4ck3rm1k3/rspec-expectations,edwardpark/rspec-expectations,jamelablack/rspec-expectations,AEgan/rspec-expectations,danielfone/rspec-expectations,chrisarcand/rspec-expectations,rspec/rspec-expectations,alexaltair/rspec-expectations,jdax/rspec-expectations,bjpcjp/rspec-expectations,maclover7/rspec-expectations,danielfone/rspec-expectations,emilyforst/rspec-expectations,travis-repos/rspec-expectations,sferik/rspec-expectations,pcreux/rspec-expectations,jamelablack/rspec-expectations,alexaltair/rspec-expectations,bjpcjp/rspec-expectations,h4ck3rm1k3/rspec-expectations,unmanbearpig/rspec-expectations,zsyed91/rspec-expectations,rspec/rspec-expectations,yaoshipu/rspec-expectations,dg-ratiodata/rspec-expectations,maclover7/rspec-expectations,jasonkarns/rspec-expectations,chrisarcand/rspec-expectations,dg-ratiodata/rspec-expectations,sferik/rspec-expectations,unmanbearpig/rspec-expectations,edwardpark/rspec-expectations,yaoshipu/rspec-expectations,AEgan/rspec-expectations,jasonkarns/rspec-expectations | ruby | ## Code Before:
def add_to_load_path(path, prepend=false)
path = File.expand_path("../#{path}", __FILE__)
if prepend
$LOAD_PATH.unshift(path) unless $LOAD_PATH.include?(path)
else
$LOAD_PATH << path unless $LOAD_PATH.include?(path)
end
end
add_to_load_path("../lib", :prepend)
add_to_load_path("../../rspec-core/lib")
add_to_load_path("../../rspec-mocks/lib")
require 'rspec/core'
require 'rspec/mocks'
Dir['./spec/support/**/*'].each do |f|
require f
end
def with_ruby(version)
yield if RUBY_PLATFORM =~ Regexp.compile("^#{version}")
end
module RSpec
module Ruby
class << self
def version
RUBY_VERSION
end
end
end
end
module RSpec
module Matchers
def fail
raise_error(RSpec::Expectations::ExpectationNotMetError)
end
def fail_with(message)
raise_error(RSpec::Expectations::ExpectationNotMetError, message)
end
end
end
RSpec::configure do |config|
config.mock_with(:rspec)
config.include RSpec::Mocks::Methods
config.color_enabled = true
end
## Instruction:
Add filtering for self-spec suite.
## Code After:
def add_to_load_path(path, prepend=false)
path = File.expand_path("../#{path}", __FILE__)
if prepend
$LOAD_PATH.unshift(path) unless $LOAD_PATH.include?(path)
else
$LOAD_PATH << path unless $LOAD_PATH.include?(path)
end
end
add_to_load_path("../lib", :prepend)
add_to_load_path("../../rspec-core/lib")
add_to_load_path("../../rspec-mocks/lib")
require 'rspec/core'
require 'rspec/mocks'
Dir['./spec/support/**/*'].each do |f|
require f
end
def with_ruby(version)
yield if RUBY_PLATFORM =~ Regexp.compile("^#{version}")
end
module RSpec
module Ruby
class << self
def version
RUBY_VERSION
end
end
end
end
module RSpec
module Matchers
def fail
raise_error(RSpec::Expectations::ExpectationNotMetError)
end
def fail_with(message)
raise_error(RSpec::Expectations::ExpectationNotMetError, message)
end
end
end
RSpec::configure do |config|
config.mock_with(:rspec)
config.include RSpec::Mocks::Methods
config.color_enabled = true
config.filter_run :focused => true
config.run_all_when_everything_filtered = true
end
| def add_to_load_path(path, prepend=false)
path = File.expand_path("../#{path}", __FILE__)
if prepend
$LOAD_PATH.unshift(path) unless $LOAD_PATH.include?(path)
else
$LOAD_PATH << path unless $LOAD_PATH.include?(path)
end
end
add_to_load_path("../lib", :prepend)
add_to_load_path("../../rspec-core/lib")
add_to_load_path("../../rspec-mocks/lib")
require 'rspec/core'
require 'rspec/mocks'
Dir['./spec/support/**/*'].each do |f|
require f
end
def with_ruby(version)
yield if RUBY_PLATFORM =~ Regexp.compile("^#{version}")
end
module RSpec
module Ruby
class << self
def version
RUBY_VERSION
end
end
end
end
module RSpec
module Matchers
def fail
raise_error(RSpec::Expectations::ExpectationNotMetError)
end
def fail_with(message)
raise_error(RSpec::Expectations::ExpectationNotMetError, message)
end
end
end
RSpec::configure do |config|
config.mock_with(:rspec)
config.include RSpec::Mocks::Methods
config.color_enabled = true
+ config.filter_run :focused => true
+ config.run_all_when_everything_filtered = true
end | 2 | 0.039216 | 2 | 0 |
16e1cbc2457fc62cc432909b611eb734f48beeab | integration/helpers/images.go | integration/helpers/images.go | /*
Copyright IBM Corp All Rights Reserved.
SPDX-License-Identifier: Apache-2.0
*/
package helpers
import (
"encoding/base32"
"fmt"
"strings"
docker "github.com/fsouza/go-dockerclient"
"github.com/hyperledger/fabric/common/util"
. "github.com/onsi/ginkgo"
. "github.com/onsi/gomega"
)
func AssertImagesExist(imageNames ...string) {
dockerClient, err := docker.NewClientFromEnv()
Expect(err).NotTo(HaveOccurred())
for _, imageName := range imageNames {
images, err := dockerClient.ListImages(docker.ListImagesOptions{
Filter: imageName,
})
ExpectWithOffset(1, err).NotTo(HaveOccurred())
if len(images) != 1 {
Fail(fmt.Sprintf("missing required image: %s", imageName), 1)
}
}
}
// UniqueName generates base-32 enocded UUIDs for container names.
func UniqueName() string {
name := base32.StdEncoding.WithPadding(base32.NoPadding).EncodeToString(util.GenerateBytesUUID())
return strings.ToLower(name)
}
| /*
Copyright IBM Corp All Rights Reserved.
SPDX-License-Identifier: Apache-2.0
*/
package helpers
import (
"encoding/base32"
"fmt"
"strings"
docker "github.com/fsouza/go-dockerclient"
"github.com/hyperledger/fabric/common/util"
. "github.com/onsi/ginkgo"
. "github.com/onsi/gomega"
)
func AssertImagesExist(imageNames ...string) {
dockerClient, err := docker.NewClientFromEnv()
Expect(err).NotTo(HaveOccurred())
for _, imageName := range imageNames {
images, err := dockerClient.ListImages(docker.ListImagesOptions{
Filters: map[string][]string{"reference": {imageName}},
})
ExpectWithOffset(1, err).NotTo(HaveOccurred())
if len(images) != 1 {
Fail(fmt.Sprintf("missing required image: %s", imageName), 1)
}
}
}
// UniqueName generates base-32 enocded UUIDs for container names.
func UniqueName() string {
name := base32.StdEncoding.WithPadding(base32.NoPadding).EncodeToString(util.GenerateBytesUUID())
return strings.ToLower(name)
}
| Update image filter used by integration tests | Update image filter used by integration tests
Signed-off-by: Matthew Sykes <89c6e4c9c1046731a9c37abefc223a83de36a5d9@us.ibm.com>
| Go | apache-2.0 | stemlending/fabric,manish-sethi/fabric-sidedb,manish-sethi/fabric,manish-sethi/fabric,manish-sethi/fabric-sidedb,stemlending/fabric,manish-sethi/fabric-sidedb,jimthematrix/fabric,manish-sethi/fabric-sidedb,manish-sethi/fabric-sidedb,jimthematrix/fabric,hyperledger/fabric,hyperledger/fabric,manish-sethi/fabric-sidedb,manish-sethi/fabric-sidedb,stemlending/fabric,stemlending/fabric | go | ## Code Before:
/*
Copyright IBM Corp All Rights Reserved.
SPDX-License-Identifier: Apache-2.0
*/
package helpers
import (
"encoding/base32"
"fmt"
"strings"
docker "github.com/fsouza/go-dockerclient"
"github.com/hyperledger/fabric/common/util"
. "github.com/onsi/ginkgo"
. "github.com/onsi/gomega"
)
func AssertImagesExist(imageNames ...string) {
dockerClient, err := docker.NewClientFromEnv()
Expect(err).NotTo(HaveOccurred())
for _, imageName := range imageNames {
images, err := dockerClient.ListImages(docker.ListImagesOptions{
Filter: imageName,
})
ExpectWithOffset(1, err).NotTo(HaveOccurred())
if len(images) != 1 {
Fail(fmt.Sprintf("missing required image: %s", imageName), 1)
}
}
}
// UniqueName generates base-32 enocded UUIDs for container names.
func UniqueName() string {
name := base32.StdEncoding.WithPadding(base32.NoPadding).EncodeToString(util.GenerateBytesUUID())
return strings.ToLower(name)
}
## Instruction:
Update image filter used by integration tests
Signed-off-by: Matthew Sykes <89c6e4c9c1046731a9c37abefc223a83de36a5d9@us.ibm.com>
## Code After:
/*
Copyright IBM Corp All Rights Reserved.
SPDX-License-Identifier: Apache-2.0
*/
package helpers
import (
"encoding/base32"
"fmt"
"strings"
docker "github.com/fsouza/go-dockerclient"
"github.com/hyperledger/fabric/common/util"
. "github.com/onsi/ginkgo"
. "github.com/onsi/gomega"
)
func AssertImagesExist(imageNames ...string) {
dockerClient, err := docker.NewClientFromEnv()
Expect(err).NotTo(HaveOccurred())
for _, imageName := range imageNames {
images, err := dockerClient.ListImages(docker.ListImagesOptions{
Filters: map[string][]string{"reference": {imageName}},
})
ExpectWithOffset(1, err).NotTo(HaveOccurred())
if len(images) != 1 {
Fail(fmt.Sprintf("missing required image: %s", imageName), 1)
}
}
}
// UniqueName generates base-32 enocded UUIDs for container names.
func UniqueName() string {
name := base32.StdEncoding.WithPadding(base32.NoPadding).EncodeToString(util.GenerateBytesUUID())
return strings.ToLower(name)
}
| /*
Copyright IBM Corp All Rights Reserved.
SPDX-License-Identifier: Apache-2.0
*/
package helpers
import (
"encoding/base32"
"fmt"
"strings"
docker "github.com/fsouza/go-dockerclient"
"github.com/hyperledger/fabric/common/util"
. "github.com/onsi/ginkgo"
. "github.com/onsi/gomega"
)
func AssertImagesExist(imageNames ...string) {
dockerClient, err := docker.NewClientFromEnv()
Expect(err).NotTo(HaveOccurred())
for _, imageName := range imageNames {
images, err := dockerClient.ListImages(docker.ListImagesOptions{
- Filter: imageName,
+ Filters: map[string][]string{"reference": {imageName}},
})
ExpectWithOffset(1, err).NotTo(HaveOccurred())
if len(images) != 1 {
Fail(fmt.Sprintf("missing required image: %s", imageName), 1)
}
}
}
// UniqueName generates base-32 enocded UUIDs for container names.
func UniqueName() string {
name := base32.StdEncoding.WithPadding(base32.NoPadding).EncodeToString(util.GenerateBytesUUID())
return strings.ToLower(name)
} | 2 | 0.05 | 1 | 1 |
c40f3179ce8da686c2bb89120c907daeb7f53b1b | lib/brewery.rb | lib/brewery.rb | require "brewery/engine"
require "brewery/auth_core"
require "jquery-rails"
require "sass-rails"
require "coffee-rails"
require "bootstrap-sass"
require "authlogic"
require "cancan"
require "will_paginate"
require "crummy"
require "email_validator"
require "simple_form"
module Brewery
end
| require "brewery/engine"
require "brewery/auth_core"
require "jquery-rails"
require "coffee-rails"
require "bootstrap-sass"
require "authlogic"
require "cancan"
require "will_paginate"
require "crummy"
require "email_validator"
require "simple_form"
module Brewery
end
| Remove sass-rails as a dependency | Remove sass-rails as a dependency
| Ruby | mit | brewery-project/brewery,brewery-project/brewery,brewery-project/brewery,brewery-project/brewery | ruby | ## Code Before:
require "brewery/engine"
require "brewery/auth_core"
require "jquery-rails"
require "sass-rails"
require "coffee-rails"
require "bootstrap-sass"
require "authlogic"
require "cancan"
require "will_paginate"
require "crummy"
require "email_validator"
require "simple_form"
module Brewery
end
## Instruction:
Remove sass-rails as a dependency
## Code After:
require "brewery/engine"
require "brewery/auth_core"
require "jquery-rails"
require "coffee-rails"
require "bootstrap-sass"
require "authlogic"
require "cancan"
require "will_paginate"
require "crummy"
require "email_validator"
require "simple_form"
module Brewery
end
| require "brewery/engine"
require "brewery/auth_core"
require "jquery-rails"
- require "sass-rails"
require "coffee-rails"
require "bootstrap-sass"
require "authlogic"
require "cancan"
require "will_paginate"
require "crummy"
require "email_validator"
require "simple_form"
module Brewery
end | 1 | 0.058824 | 0 | 1 |
d8a613ea19ecf87097dee3cbcb121f60dba0726c | .travis.yml | .travis.yml | cache: bundler
language: ruby
rvm:
- 2.1.10
- 2.2.7
- 2.3.4
- 2.4.1
- 2.5.8
- 2.6.6
- 2.7.1
env:
global:
- NOKOGIRI_USE_SYSTEM_LIBRARIES=true
matrix:
- AR=5.0.3
- AR=5.1.1
- AR=5.2.2
- AR=6.0.3.3
matrix:
exclude:
- rvm: 2.1.10
env: AR=5.0.3
- rvm: 2.1.10
env: AR=5.1.1
- rvm: 2.1.10
env: AR=6.0.3.3
- rvm: 2.2.7
env: AR=6.0.3.3
- rvm: 2.3.4
env: AR=6.0.3.3
- rvm: 2.4.1
env: AR=6.0.3.3
- rvm: 2.5.8
env: AR=6.0.3.3
- rvm: 2.6.6
env: AR=6.0.3.3
before_script:
- mysql -e 'create database seeder_test;'
| cache: bundler
language: ruby
services:
- mysql
rvm:
- 2.1.10
- 2.2.7
- 2.3.4
- 2.4.1
- 2.5.8
- 2.6.6
- 2.7.1
env:
global:
- NOKOGIRI_USE_SYSTEM_LIBRARIES=true
matrix:
- AR=5.0.3
- AR=5.1.1
- AR=5.2.2
- AR=6.0.3.3
matrix:
exclude:
- rvm: 2.1.10
env: AR=5.0.3
- rvm: 2.1.10
env: AR=5.1.1
- rvm: 2.1.10
env: AR=6.0.3.3
- rvm: 2.2.7
env: AR=6.0.3.3
- rvm: 2.3.4
env: AR=6.0.3.3
- rvm: 2.4.1
env: AR=6.0.3.3
- rvm: 2.5.8
env: AR=6.0.3.3
- rvm: 2.6.6
env: AR=6.0.3.3
before_script:
- mysql -e 'create database seeder_test;'
| Add mysql to CI environment | Add mysql to CI environment
| YAML | mit | wegowise/seeder | yaml | ## Code Before:
cache: bundler
language: ruby
rvm:
- 2.1.10
- 2.2.7
- 2.3.4
- 2.4.1
- 2.5.8
- 2.6.6
- 2.7.1
env:
global:
- NOKOGIRI_USE_SYSTEM_LIBRARIES=true
matrix:
- AR=5.0.3
- AR=5.1.1
- AR=5.2.2
- AR=6.0.3.3
matrix:
exclude:
- rvm: 2.1.10
env: AR=5.0.3
- rvm: 2.1.10
env: AR=5.1.1
- rvm: 2.1.10
env: AR=6.0.3.3
- rvm: 2.2.7
env: AR=6.0.3.3
- rvm: 2.3.4
env: AR=6.0.3.3
- rvm: 2.4.1
env: AR=6.0.3.3
- rvm: 2.5.8
env: AR=6.0.3.3
- rvm: 2.6.6
env: AR=6.0.3.3
before_script:
- mysql -e 'create database seeder_test;'
## Instruction:
Add mysql to CI environment
## Code After:
cache: bundler
language: ruby
services:
- mysql
rvm:
- 2.1.10
- 2.2.7
- 2.3.4
- 2.4.1
- 2.5.8
- 2.6.6
- 2.7.1
env:
global:
- NOKOGIRI_USE_SYSTEM_LIBRARIES=true
matrix:
- AR=5.0.3
- AR=5.1.1
- AR=5.2.2
- AR=6.0.3.3
matrix:
exclude:
- rvm: 2.1.10
env: AR=5.0.3
- rvm: 2.1.10
env: AR=5.1.1
- rvm: 2.1.10
env: AR=6.0.3.3
- rvm: 2.2.7
env: AR=6.0.3.3
- rvm: 2.3.4
env: AR=6.0.3.3
- rvm: 2.4.1
env: AR=6.0.3.3
- rvm: 2.5.8
env: AR=6.0.3.3
- rvm: 2.6.6
env: AR=6.0.3.3
before_script:
- mysql -e 'create database seeder_test;'
| cache: bundler
language: ruby
+ services:
+ - mysql
rvm:
- 2.1.10
- 2.2.7
- 2.3.4
- 2.4.1
- 2.5.8
- 2.6.6
- 2.7.1
env:
global:
- NOKOGIRI_USE_SYSTEM_LIBRARIES=true
matrix:
- AR=5.0.3
- AR=5.1.1
- AR=5.2.2
- AR=6.0.3.3
matrix:
exclude:
- rvm: 2.1.10
env: AR=5.0.3
- rvm: 2.1.10
env: AR=5.1.1
- rvm: 2.1.10
env: AR=6.0.3.3
- rvm: 2.2.7
env: AR=6.0.3.3
- rvm: 2.3.4
env: AR=6.0.3.3
- rvm: 2.4.1
env: AR=6.0.3.3
- rvm: 2.5.8
env: AR=6.0.3.3
- rvm: 2.6.6
env: AR=6.0.3.3
before_script:
- mysql -e 'create database seeder_test;' | 2 | 0.046512 | 2 | 0 |
8076767deed5d79fb33f7e671c4901da673cd989 | lib/peoplesoft_models/base.rb | lib/peoplesoft_models/base.rb | class PeoplesoftModels::Base < ActiveRecord::Base
self.abstract_class = true
end
| module PeoplesoftModels
class Base < ActiveRecord::Base
self.abstract_class = true
end
end
| Make PeoplesoftModels::Base nesting consistent with other classes | Make PeoplesoftModels::Base nesting consistent with other classes
| Ruby | mit | cdinger/activerecord-peoplesoft_models,cdinger/activerecord-peoplesoft_models | ruby | ## Code Before:
class PeoplesoftModels::Base < ActiveRecord::Base
self.abstract_class = true
end
## Instruction:
Make PeoplesoftModels::Base nesting consistent with other classes
## Code After:
module PeoplesoftModels
class Base < ActiveRecord::Base
self.abstract_class = true
end
end
| + module PeoplesoftModels
- class PeoplesoftModels::Base < ActiveRecord::Base
? ------------------
+ class Base < ActiveRecord::Base
? ++
- self.abstract_class = true
+ self.abstract_class = true
? ++
+ end
end | 6 | 2 | 4 | 2 |
5533fec7239e3443983b12af819c3218fd376dd8 | README.md | README.md |
This workspace contains two projects that perform Rc4 encryption and decryption. One project is written in Objective-C, the other, Swift. Which language is fastest?
The code is written in a traditional way. I intentionally left the new functional hotness out. I didn't want it to distract from the goal. This code is also not as optimised as it could be. And there are no compiler 'ooh fast' switch optimisations.
##To Run The Tests
Open the workspace. Run the unit tests for each project and compare the performance results.
|
This workspace contains two projects that perform Rc4 encryption and decryption. One project is written in Objective-C, the other, Swift. Which language is fastest?
http://codetipping.weebly.com
##To Run The Tests
Open the workspace. Run the unit tests for each project and compare the performance results.
| Add link to reference url. | Add link to reference url. | Markdown | mit | RussVanBert/Rc4 | markdown | ## Code Before:
This workspace contains two projects that perform Rc4 encryption and decryption. One project is written in Objective-C, the other, Swift. Which language is fastest?
The code is written in a traditional way. I intentionally left the new functional hotness out. I didn't want it to distract from the goal. This code is also not as optimised as it could be. And there are no compiler 'ooh fast' switch optimisations.
##To Run The Tests
Open the workspace. Run the unit tests for each project and compare the performance results.
## Instruction:
Add link to reference url.
## Code After:
This workspace contains two projects that perform Rc4 encryption and decryption. One project is written in Objective-C, the other, Swift. Which language is fastest?
http://codetipping.weebly.com
##To Run The Tests
Open the workspace. Run the unit tests for each project and compare the performance results.
|
This workspace contains two projects that perform Rc4 encryption and decryption. One project is written in Objective-C, the other, Swift. Which language is fastest?
- The code is written in a traditional way. I intentionally left the new functional hotness out. I didn't want it to distract from the goal. This code is also not as optimised as it could be. And there are no compiler 'ooh fast' switch optimisations.
+ http://codetipping.weebly.com
##To Run The Tests
Open the workspace. Run the unit tests for each project and compare the performance results. | 2 | 0.25 | 1 | 1 |
aaf9227c127e9cb42514bc00f3416f18b8d14eb3 | app/views/passwords/index.html.erb | app/views/passwords/index.html.erb | <% content_for :heading do -%><%= t('myplaceonline.passwords.title') %><% end -%>
<h1><%= t('myplaceonline.passwords.title') %></h1>
<%= flashes! %>
<%
if @passwords.length == 0
%>
<p><%= t('myplaceonline.passwords.none') %></p>
<%
else
%>
<ul data-role="listview" data-inset="true" data-filter="true">
<%
@passwords.each do |password|
%>
<li>
<%= link_to password.name, password %>
</li>
<%
end
%>
</ul>
<%
end
%>
<% content_for :footer do -%>
<div data-role="footer" data-position="fixed">
<div data-role="navbar">
<ul>
<li><%= link_to t('myplaceonline.passwords.title_add'), new_password_path, "data-icon" => "plus" %></li>
<li><a href="#" data-icon="bars">Import</a></li>
</ul>
</div>
</div>
<% end -%> | <% content_for :heading do -%><%= t('myplaceonline.passwords.title') %><% end -%>
<h1><%= t('myplaceonline.passwords.title') %></h1>
<%= flashes! %>
<%
if @passwords.length == 0
%>
<p><%= t('myplaceonline.passwords.none') %></p>
<%
else
%>
<div class="searchable_container">
<ul class="searchable" data-role="listview" data-inset="true" data-filter="true">
<%
@passwords.each do |password|
%>
<li>
<%= link_to password.name, password %>
</li>
<%
end
%>
</ul>
</div>
<%
end
%>
<script type="text/javascript">
onPageLoad(function() {
var input = $(".searchable_container input");
input.keyup(function(e) {
if(e.which == 13) {
var searchList = $(".searchable li:not(.ui-screen-hidden)");
if (searchList.size() > 0) {
e.preventDefault();
navigate(searchList.filter(":first").children("a").attr("href"));
}
}
return true;
});
input.focus();
});
</script>
<% content_for :footer do -%>
<div data-role="footer" data-position="fixed">
<div data-role="navbar">
<ul>
<li><%= link_to t('myplaceonline.passwords.title_add'), new_password_path, "data-icon" => "plus" %></li>
<li><a href="#" data-icon="bars">Import</a></li>
</ul>
</div>
</div>
<% end -%> | Add ENTER support to passwords list | Add ENTER support to passwords list
| HTML+ERB | agpl-3.0 | myplaceonline/myplaceonline_rails,myplaceonline/myplaceonline_rails,myplaceonline/myplaceonline_rails | html+erb | ## Code Before:
<% content_for :heading do -%><%= t('myplaceonline.passwords.title') %><% end -%>
<h1><%= t('myplaceonline.passwords.title') %></h1>
<%= flashes! %>
<%
if @passwords.length == 0
%>
<p><%= t('myplaceonline.passwords.none') %></p>
<%
else
%>
<ul data-role="listview" data-inset="true" data-filter="true">
<%
@passwords.each do |password|
%>
<li>
<%= link_to password.name, password %>
</li>
<%
end
%>
</ul>
<%
end
%>
<% content_for :footer do -%>
<div data-role="footer" data-position="fixed">
<div data-role="navbar">
<ul>
<li><%= link_to t('myplaceonline.passwords.title_add'), new_password_path, "data-icon" => "plus" %></li>
<li><a href="#" data-icon="bars">Import</a></li>
</ul>
</div>
</div>
<% end -%>
## Instruction:
Add ENTER support to passwords list
## Code After:
<% content_for :heading do -%><%= t('myplaceonline.passwords.title') %><% end -%>
<h1><%= t('myplaceonline.passwords.title') %></h1>
<%= flashes! %>
<%
if @passwords.length == 0
%>
<p><%= t('myplaceonline.passwords.none') %></p>
<%
else
%>
<div class="searchable_container">
<ul class="searchable" data-role="listview" data-inset="true" data-filter="true">
<%
@passwords.each do |password|
%>
<li>
<%= link_to password.name, password %>
</li>
<%
end
%>
</ul>
</div>
<%
end
%>
<script type="text/javascript">
onPageLoad(function() {
var input = $(".searchable_container input");
input.keyup(function(e) {
if(e.which == 13) {
var searchList = $(".searchable li:not(.ui-screen-hidden)");
if (searchList.size() > 0) {
e.preventDefault();
navigate(searchList.filter(":first").children("a").attr("href"));
}
}
return true;
});
input.focus();
});
</script>
<% content_for :footer do -%>
<div data-role="footer" data-position="fixed">
<div data-role="navbar">
<ul>
<li><%= link_to t('myplaceonline.passwords.title_add'), new_password_path, "data-icon" => "plus" %></li>
<li><a href="#" data-icon="bars">Import</a></li>
</ul>
</div>
</div>
<% end -%> | <% content_for :heading do -%><%= t('myplaceonline.passwords.title') %><% end -%>
<h1><%= t('myplaceonline.passwords.title') %></h1>
<%= flashes! %>
<%
if @passwords.length == 0
%>
<p><%= t('myplaceonline.passwords.none') %></p>
<%
else
%>
+ <div class="searchable_container">
- <ul data-role="listview" data-inset="true" data-filter="true">
+ <ul class="searchable" data-role="listview" data-inset="true" data-filter="true">
? ++ +++++++++++++++++++
<%
- @passwords.each do |password|
+ @passwords.each do |password|
? ++
%>
- <li>
+ <li>
? ++
- <%= link_to password.name, password %>
+ <%= link_to password.name, password %>
? ++
- </li>
+ </li>
? ++
<%
- end
+ end
? ++
%>
- </ul>
+ </ul>
? ++
+ </div>
<%
end
%>
+
+ <script type="text/javascript">
+ onPageLoad(function() {
+
+ var input = $(".searchable_container input");
+
+ input.keyup(function(e) {
+ if(e.which == 13) {
+ var searchList = $(".searchable li:not(.ui-screen-hidden)");
+ if (searchList.size() > 0) {
+ e.preventDefault();
+ navigate(searchList.filter(":first").children("a").attr("href"));
+ }
+ }
+ return true;
+ });
+
+ input.focus();
+
+ });
+ </script>
<% content_for :footer do -%>
<div data-role="footer" data-position="fixed">
<div data-role="navbar">
<ul>
<li><%= link_to t('myplaceonline.passwords.title_add'), new_password_path, "data-icon" => "plus" %></li>
<li><a href="#" data-icon="bars">Import</a></li>
</ul>
</div>
</div>
<% end -%> | 37 | 1.027778 | 30 | 7 |
f68d5e9fd8dfa5bcd2925429c53b2048380a27d7 | .travis.yml | .travis.yml | language: sh
before_install:
- git clone git://github.com/bmizerany/roundup
- cd roundup
- ./configure
- make
- export PATH="$PATH:$PWD"
- cd -
script: make
| language: sh
os:
- linux
- osx
before_install:
- git clone git://github.com/bmizerany/roundup
- cd roundup
- ./configure
- make
- export PATH="$PATH:$PWD"
- cd -
script: make
| Add OSX to CI config | Add OSX to CI config
| YAML | mit | aaronjameslang/git-hook-commit-message-cbeams | yaml | ## Code Before:
language: sh
before_install:
- git clone git://github.com/bmizerany/roundup
- cd roundup
- ./configure
- make
- export PATH="$PATH:$PWD"
- cd -
script: make
## Instruction:
Add OSX to CI config
## Code After:
language: sh
os:
- linux
- osx
before_install:
- git clone git://github.com/bmizerany/roundup
- cd roundup
- ./configure
- make
- export PATH="$PATH:$PWD"
- cd -
script: make
| language: sh
+
+ os:
+ - linux
+ - osx
+
before_install:
- git clone git://github.com/bmizerany/roundup
- cd roundup
- ./configure
- make
- export PATH="$PATH:$PWD"
- cd -
script: make | 5 | 0.555556 | 5 | 0 |
081174d0e35a91d19da0d397440d49bfc80e31b8 | spec/factories/page_factory.rb | spec/factories/page_factory.rb | FactoryGirl.define do
factory :page do
title 'New Page'
slug 'page'
breadcrumb 'New Page'
status_id '1'
factory :page_with_layout do
layout
end
# :parent_id => nil
# factory :admin_user do
# spree_roles { [Spree::Role.find_by(name: 'admin') || create(:role, name: 'admin')] }
# end
end
end | FactoryGirl.define do
factory :page do
title 'New Page'
slug 'page'
breadcrumb 'New Page'
status_id '1'
factory :page_with_layout do
layout
end
end
end | Clean up copy/paste remnants in page factory | Clean up copy/paste remnants in page factory
| Ruby | mit | blj/radiant,ahjohannessen/radiant,radiant/radiant,ahjohannessen/radiant,ahjohannessen/radiant,blj/radiant,blj/radiant,radiant/radiant,radiant/radiant,LytayTOUCH/radiant,LytayTOUCH/radiant,blj/radiant,LytayTOUCH/radiant,radiant/radiant,LytayTOUCH/radiant,ahjohannessen/radiant | ruby | ## Code Before:
FactoryGirl.define do
factory :page do
title 'New Page'
slug 'page'
breadcrumb 'New Page'
status_id '1'
factory :page_with_layout do
layout
end
# :parent_id => nil
# factory :admin_user do
# spree_roles { [Spree::Role.find_by(name: 'admin') || create(:role, name: 'admin')] }
# end
end
end
## Instruction:
Clean up copy/paste remnants in page factory
## Code After:
FactoryGirl.define do
factory :page do
title 'New Page'
slug 'page'
breadcrumb 'New Page'
status_id '1'
factory :page_with_layout do
layout
end
end
end | FactoryGirl.define do
factory :page do
title 'New Page'
slug 'page'
breadcrumb 'New Page'
status_id '1'
factory :page_with_layout do
layout
end
-
- # :parent_id => nil
-
- # factory :admin_user do
- # spree_roles { [Spree::Role.find_by(name: 'admin') || create(:role, name: 'admin')] }
- # end
end
end | 6 | 0.333333 | 0 | 6 |
43ecb332c85edc585dae35d1c574c065da89ca43 | gulp/plugins/gulp-sanitize-translations.js | gulp/plugins/gulp-sanitize-translations.js | var through = require( 'through2' );
var gutil = require( 'gulp-util' );
var PluginError = gutil.PluginError;
var sanitize = require( 'sanitize-html' );
const PLUGIN_NAME = 'gulp-sanitize-translations.js';
module.exports = function( options )
{
// Creating a stream through which each file will pass
var stream = through.obj( function( file, enc, callback )
{
if ( file.isBuffer() ) {
var content = file.contents.toString();
var parsed = JSON.parse( content );
var lang = Object.keys( parsed )[0];
for ( var i in parsed[ lang ] ) {
parsed[ lang ][ i ] = sanitize( parsed[ lang ][ i ] );
}
file.contents = new Buffer( JSON.stringify( parsed ), 'utf-8' );
}
// Streams not supported.
else if ( file.isStream() ) {
this.emit( 'error', new gutil.PluginError( PLUGIN_NAME, 'Streaming not supported.' ) );
return callback();
}
// Anything else just falls through.
this.push( file );
return callback();
} );
// returning the file stream
return stream;
};
| var through = require( 'through2' );
var gutil = require( 'gulp-util' );
var PluginError = gutil.PluginError;
var sanitize = require( 'sanitize-html' );
const PLUGIN_NAME = 'gulp-sanitize-translations.js';
module.exports = function( options )
{
// Creating a stream through which each file will pass
var stream = through.obj( function( file, enc, callback )
{
if ( file.isBuffer() ) {
var content = file.contents.toString();
var parsed = JSON.parse( content );
var lang = Object.keys( parsed )[0];
for ( var i in parsed[ lang ] ) {
if ( Array.isArray( parsed[ lang ][ i ] ) ) {
for ( var n in parsed[ lang ][ i ] ) {
parsed[ lang ][ i ][ n ] = sanitize( parsed[ lang ][ i ][ n ] );
}
}
else {
parsed[ lang ][ i ] = sanitize( parsed[ lang ][ i ] );
}
}
file.contents = new Buffer( JSON.stringify( parsed ), 'utf-8' );
}
// Streams not supported.
else if ( file.isStream() ) {
this.emit( 'error', new gutil.PluginError( PLUGIN_NAME, 'Streaming not supported.' ) );
return callback();
}
// Anything else just falls through.
this.push( file );
return callback();
} );
// returning the file stream
return stream;
};
| Fix translation sanitizing to work for plurals. | Fix translation sanitizing to work for plurals.
| JavaScript | mit | gamejolt/frontend-lib,gamejolt/frontend-lib,gamejolt/frontend-lib,gamejolt/frontend-lib | javascript | ## Code Before:
var through = require( 'through2' );
var gutil = require( 'gulp-util' );
var PluginError = gutil.PluginError;
var sanitize = require( 'sanitize-html' );
const PLUGIN_NAME = 'gulp-sanitize-translations.js';
module.exports = function( options )
{
// Creating a stream through which each file will pass
var stream = through.obj( function( file, enc, callback )
{
if ( file.isBuffer() ) {
var content = file.contents.toString();
var parsed = JSON.parse( content );
var lang = Object.keys( parsed )[0];
for ( var i in parsed[ lang ] ) {
parsed[ lang ][ i ] = sanitize( parsed[ lang ][ i ] );
}
file.contents = new Buffer( JSON.stringify( parsed ), 'utf-8' );
}
// Streams not supported.
else if ( file.isStream() ) {
this.emit( 'error', new gutil.PluginError( PLUGIN_NAME, 'Streaming not supported.' ) );
return callback();
}
// Anything else just falls through.
this.push( file );
return callback();
} );
// returning the file stream
return stream;
};
## Instruction:
Fix translation sanitizing to work for plurals.
## Code After:
var through = require( 'through2' );
var gutil = require( 'gulp-util' );
var PluginError = gutil.PluginError;
var sanitize = require( 'sanitize-html' );
const PLUGIN_NAME = 'gulp-sanitize-translations.js';
module.exports = function( options )
{
// Creating a stream through which each file will pass
var stream = through.obj( function( file, enc, callback )
{
if ( file.isBuffer() ) {
var content = file.contents.toString();
var parsed = JSON.parse( content );
var lang = Object.keys( parsed )[0];
for ( var i in parsed[ lang ] ) {
if ( Array.isArray( parsed[ lang ][ i ] ) ) {
for ( var n in parsed[ lang ][ i ] ) {
parsed[ lang ][ i ][ n ] = sanitize( parsed[ lang ][ i ][ n ] );
}
}
else {
parsed[ lang ][ i ] = sanitize( parsed[ lang ][ i ] );
}
}
file.contents = new Buffer( JSON.stringify( parsed ), 'utf-8' );
}
// Streams not supported.
else if ( file.isStream() ) {
this.emit( 'error', new gutil.PluginError( PLUGIN_NAME, 'Streaming not supported.' ) );
return callback();
}
// Anything else just falls through.
this.push( file );
return callback();
} );
// returning the file stream
return stream;
};
| var through = require( 'through2' );
var gutil = require( 'gulp-util' );
var PluginError = gutil.PluginError;
var sanitize = require( 'sanitize-html' );
const PLUGIN_NAME = 'gulp-sanitize-translations.js';
module.exports = function( options )
{
// Creating a stream through which each file will pass
var stream = through.obj( function( file, enc, callback )
{
if ( file.isBuffer() ) {
var content = file.contents.toString();
var parsed = JSON.parse( content );
var lang = Object.keys( parsed )[0];
for ( var i in parsed[ lang ] ) {
+
+ if ( Array.isArray( parsed[ lang ][ i ] ) ) {
+ for ( var n in parsed[ lang ][ i ] ) {
+ parsed[ lang ][ i ][ n ] = sanitize( parsed[ lang ][ i ][ n ] );
+ }
+ }
+ else {
- parsed[ lang ][ i ] = sanitize( parsed[ lang ][ i ] );
+ parsed[ lang ][ i ] = sanitize( parsed[ lang ][ i ] );
? +
+ }
}
file.contents = new Buffer( JSON.stringify( parsed ), 'utf-8' );
}
// Streams not supported.
else if ( file.isStream() ) {
this.emit( 'error', new gutil.PluginError( PLUGIN_NAME, 'Streaming not supported.' ) );
return callback();
}
// Anything else just falls through.
this.push( file );
return callback();
} );
// returning the file stream
return stream;
}; | 10 | 0.263158 | 9 | 1 |
a6a1b0e34c4eb10a44b78c88877166ef03edf239 | lib/script.coffee | lib/script.coffee | ScriptView = require './script-view'
configUri = "atom://script"
grammarMap =
CoffeeScript:
interpreter: "coffee"
makeargs: (code) -> ['-e', code]
Python:
interpreter: "python"
makeargs: (code) -> ['-c', code]
module.exports =
activate: ->
atom.project.registerOpener (uri) =>
interpreter = grammarMap[@lang]["interpreter"]
makeargs = grammarMap[@lang]["makeargs"]
@scriptView = new ScriptView(interpreter, makeargs) if uri is configUri
atom.workspaceView.command "script:run-selection", =>
editor = atom.workspace.getActiveEditor()
code = editor.getSelectedText()
if not code? or not code
code = editor.getText()
if not editor?
console.log("Editor unavailable")
return
grammar = editor.getGrammar()
@lang = grammar.name
if grammar.name == "Null Grammar"
console.log("Need to select a language in the lower left or")
console.log("save the file with an appropriate extension.")
return
if ! grammar.name in grammarMap
console.log("Interpreter not configured for " + @lang)
console.log("Send a pull request to add support!")
return
atom.workspaceView.open(configUri, split: 'right')
@scriptView.runit(code)
| ScriptView = require './script-view'
configUri = "atom://script"
grammarMap =
CoffeeScript:
interpreter: "coffee"
makeargs: (code) -> ['-e', code]
Python:
interpreter: "python"
makeargs: (code) -> ['-c', code]
module.exports =
activate: ->
atom.project.registerOpener (uri) =>
interpreter = grammarMap[@lang]["interpreter"]
makeargs = grammarMap[@lang]["makeargs"]
@scriptView = new ScriptView(interpreter, makeargs) if uri is configUri
atom.workspaceView.command "script:run-selection", =>
editor = atom.workspace.getActiveEditor()
if not editor?
console.log("Editor unavailable")
return
code = editor.getSelectedText()
if not code? or not code
code = editor.getText()
grammar = editor.getGrammar()
@lang = grammar.name
if grammar.name == "Null Grammar"
console.log("Need to select a language in the lower left or")
console.log("save the file with an appropriate extension.")
return
if ! grammar.name in grammarMap
console.log("Interpreter not configured for " + @lang)
console.log("Send a pull request to add support!")
return
atom.workspaceView.open(configUri, split: 'right')
@scriptView.runit(code)
| Make sure editor is set before trying to get code. | Make sure editor is set before trying to get code.
| CoffeeScript | mit | jchannon/atom-script,rodionovd/atom-script,rodionovd/atom-script,anfedorov/atom-script,jchannon/atom-script,efatsi/atom-script,rodionovd/atom-script,jchannon/atom-script,jchannon/atom-script,efatsi/atom-script,anfedorov/atom-script,Calyhre/atom-script,anfedorov/atom-script,Calyhre/atom-script,rgbkrk/atom-script,rodionovd/atom-script,anfedorov/atom-script,Calyhre/atom-script,efatsi/atom-script,Calyhre/atom-script,anfedorov/atom-script,Calyhre/atom-script,rodionovd/atom-script,efatsi/atom-script,Calyhre/atom-script,rodionovd/atom-script,jchannon/atom-script,rodionovd/atom-script,efatsi/atom-script,rodionovd/atom-script,jchannon/atom-script,jchannon/atom-script,Calyhre/atom-script,jchannon/atom-script,rodionovd/atom-script,idleberg/atom-script,anfedorov/atom-script,anfedorov/atom-script,rodionovd/atom-script,anfedorov/atom-script,chenruixuan/atom-script,anfedorov/atom-script,fscherwi/atom-script,TomosBlack/atom-script,rodionovd/atom-script,rodionovd/atom-script,MichaelSp/atom-script,rodionovd/atom-script,Calyhre/atom-script,efatsi/atom-script,efatsi/atom-script,anfedorov/atom-script,Calyhre/atom-script,efatsi/atom-script,efatsi/atom-script,efatsi/atom-script,anfedorov/atom-script,jchannon/atom-script,anfedorov/atom-script,efatsi/atom-script,efatsi/atom-script,anfedorov/atom-script,Calyhre/atom-script,efatsi/atom-script,Calyhre/atom-script,Calyhre/atom-script,jchannon/atom-script,jchannon/atom-script | coffeescript | ## Code Before:
ScriptView = require './script-view'
configUri = "atom://script"
grammarMap =
CoffeeScript:
interpreter: "coffee"
makeargs: (code) -> ['-e', code]
Python:
interpreter: "python"
makeargs: (code) -> ['-c', code]
module.exports =
activate: ->
atom.project.registerOpener (uri) =>
interpreter = grammarMap[@lang]["interpreter"]
makeargs = grammarMap[@lang]["makeargs"]
@scriptView = new ScriptView(interpreter, makeargs) if uri is configUri
atom.workspaceView.command "script:run-selection", =>
editor = atom.workspace.getActiveEditor()
code = editor.getSelectedText()
if not code? or not code
code = editor.getText()
if not editor?
console.log("Editor unavailable")
return
grammar = editor.getGrammar()
@lang = grammar.name
if grammar.name == "Null Grammar"
console.log("Need to select a language in the lower left or")
console.log("save the file with an appropriate extension.")
return
if ! grammar.name in grammarMap
console.log("Interpreter not configured for " + @lang)
console.log("Send a pull request to add support!")
return
atom.workspaceView.open(configUri, split: 'right')
@scriptView.runit(code)
## Instruction:
Make sure editor is set before trying to get code.
## Code After:
ScriptView = require './script-view'
configUri = "atom://script"
grammarMap =
CoffeeScript:
interpreter: "coffee"
makeargs: (code) -> ['-e', code]
Python:
interpreter: "python"
makeargs: (code) -> ['-c', code]
module.exports =
activate: ->
atom.project.registerOpener (uri) =>
interpreter = grammarMap[@lang]["interpreter"]
makeargs = grammarMap[@lang]["makeargs"]
@scriptView = new ScriptView(interpreter, makeargs) if uri is configUri
atom.workspaceView.command "script:run-selection", =>
editor = atom.workspace.getActiveEditor()
if not editor?
console.log("Editor unavailable")
return
code = editor.getSelectedText()
if not code? or not code
code = editor.getText()
grammar = editor.getGrammar()
@lang = grammar.name
if grammar.name == "Null Grammar"
console.log("Need to select a language in the lower left or")
console.log("save the file with an appropriate extension.")
return
if ! grammar.name in grammarMap
console.log("Interpreter not configured for " + @lang)
console.log("Send a pull request to add support!")
return
atom.workspaceView.open(configUri, split: 'right')
@scriptView.runit(code)
| ScriptView = require './script-view'
configUri = "atom://script"
grammarMap =
CoffeeScript:
interpreter: "coffee"
makeargs: (code) -> ['-e', code]
Python:
interpreter: "python"
makeargs: (code) -> ['-c', code]
module.exports =
activate: ->
atom.project.registerOpener (uri) =>
interpreter = grammarMap[@lang]["interpreter"]
makeargs = grammarMap[@lang]["makeargs"]
@scriptView = new ScriptView(interpreter, makeargs) if uri is configUri
atom.workspaceView.command "script:run-selection", =>
editor = atom.workspace.getActiveEditor()
+
+ if not editor?
+ console.log("Editor unavailable")
+ return
+
code = editor.getSelectedText()
if not code? or not code
code = editor.getText()
-
- if not editor?
- console.log("Editor unavailable")
- return
grammar = editor.getGrammar()
@lang = grammar.name
if grammar.name == "Null Grammar"
console.log("Need to select a language in the lower left or")
console.log("save the file with an appropriate extension.")
return
if ! grammar.name in grammarMap
console.log("Interpreter not configured for " + @lang)
console.log("Send a pull request to add support!")
return
atom.workspaceView.open(configUri, split: 'right')
@scriptView.runit(code) | 9 | 0.183673 | 5 | 4 |
a2d3dea21c3aa90125cfe00ca70c0bab25de2147 | tasks/rules/styles.js | tasks/rules/styles.js | const MiniCSSExtractPlugin = require('mini-css-extract-plugin');
const styleFunctions = require('../../src/styles/functions/index.js');
module.exports = (env) => {
const styleLoaders = {
production: MiniCSSExtractPlugin.loader,
development: 'style-loader'
};
return {
test: /\.css$/,
exclude: /node_modules/,
use: [
{
loader: styleLoaders[env] // Creates style nodes from JS strings
},
{
loader: 'css-loader' // Translates CSS into CommonJS
},
{
loader: 'postcss-loader', // More CSS Plugins
options: {
postcssOptions: {
plugins: [
require('postcss-import'),
require('postcss-at-rules-variables')({ atRules: ['for', 'if', 'else', 'each', 'mixin', 'media'] }),
require('postcss-simple-vars'),
require('postcss-replace')({ pattern: /##/g, data: { replaceAll: '$' } }),
require('postcss-mixins'),
require('postcss-functions')({ functions: styleFunctions }),
require('postcss-each'),
require('postcss-calc'),
require('postcss-fontpath'),
require('postcss-nested'),
require('autoprefixer'),
require('postcss-discard-comments')
]
}
}
}
]
};
};
| const MiniCSSExtractPlugin = require('mini-css-extract-plugin');
const styleFunctions = require('../../src/styles/functions/index.js');
module.exports = (env) => {
const styleLoaders = {
production: MiniCSSExtractPlugin.loader,
development: 'style-loader'
};
return {
test: /\.css$/,
exclude: /node_modules/,
use: [
{
loader: styleLoaders[env] // Creates style nodes from JS strings
},
{
loader: 'css-loader' // Translates CSS into CommonJS
},
{
loader: 'postcss-loader', // More CSS Plugins
options: {
postcssOptions: {
plugins: [
require('postcss-import'),
require('postcss-at-rules-variables')({ atRules: ['each', 'mixin', 'media'] }),
require('postcss-simple-vars'),
require('postcss-replace')({ pattern: /##/g, data: { replaceAll: '$' } }),
require('postcss-mixins'),
require('postcss-functions')({ functions: styleFunctions }),
require('postcss-each'),
require('postcss-calc'),
require('postcss-fontpath'),
require('postcss-nested'),
require('autoprefixer'),
require('postcss-discard-comments')
]
}
}
}
]
};
};
| Remove conditional in at rules | Remove conditional in at rules
| JavaScript | mit | CKGrafico/Frontend-Boilerplates,CKGrafico/Frontend-Boilerplates,CKGrafico/Gulp-Boilerplates,CKGrafico/Gulp-Boilerplates | javascript | ## Code Before:
const MiniCSSExtractPlugin = require('mini-css-extract-plugin');
const styleFunctions = require('../../src/styles/functions/index.js');
module.exports = (env) => {
const styleLoaders = {
production: MiniCSSExtractPlugin.loader,
development: 'style-loader'
};
return {
test: /\.css$/,
exclude: /node_modules/,
use: [
{
loader: styleLoaders[env] // Creates style nodes from JS strings
},
{
loader: 'css-loader' // Translates CSS into CommonJS
},
{
loader: 'postcss-loader', // More CSS Plugins
options: {
postcssOptions: {
plugins: [
require('postcss-import'),
require('postcss-at-rules-variables')({ atRules: ['for', 'if', 'else', 'each', 'mixin', 'media'] }),
require('postcss-simple-vars'),
require('postcss-replace')({ pattern: /##/g, data: { replaceAll: '$' } }),
require('postcss-mixins'),
require('postcss-functions')({ functions: styleFunctions }),
require('postcss-each'),
require('postcss-calc'),
require('postcss-fontpath'),
require('postcss-nested'),
require('autoprefixer'),
require('postcss-discard-comments')
]
}
}
}
]
};
};
## Instruction:
Remove conditional in at rules
## Code After:
const MiniCSSExtractPlugin = require('mini-css-extract-plugin');
const styleFunctions = require('../../src/styles/functions/index.js');
module.exports = (env) => {
const styleLoaders = {
production: MiniCSSExtractPlugin.loader,
development: 'style-loader'
};
return {
test: /\.css$/,
exclude: /node_modules/,
use: [
{
loader: styleLoaders[env] // Creates style nodes from JS strings
},
{
loader: 'css-loader' // Translates CSS into CommonJS
},
{
loader: 'postcss-loader', // More CSS Plugins
options: {
postcssOptions: {
plugins: [
require('postcss-import'),
require('postcss-at-rules-variables')({ atRules: ['each', 'mixin', 'media'] }),
require('postcss-simple-vars'),
require('postcss-replace')({ pattern: /##/g, data: { replaceAll: '$' } }),
require('postcss-mixins'),
require('postcss-functions')({ functions: styleFunctions }),
require('postcss-each'),
require('postcss-calc'),
require('postcss-fontpath'),
require('postcss-nested'),
require('autoprefixer'),
require('postcss-discard-comments')
]
}
}
}
]
};
};
| const MiniCSSExtractPlugin = require('mini-css-extract-plugin');
const styleFunctions = require('../../src/styles/functions/index.js');
module.exports = (env) => {
const styleLoaders = {
production: MiniCSSExtractPlugin.loader,
development: 'style-loader'
};
return {
test: /\.css$/,
exclude: /node_modules/,
use: [
{
loader: styleLoaders[env] // Creates style nodes from JS strings
},
{
loader: 'css-loader' // Translates CSS into CommonJS
},
{
loader: 'postcss-loader', // More CSS Plugins
options: {
postcssOptions: {
plugins: [
require('postcss-import'),
- require('postcss-at-rules-variables')({ atRules: ['for', 'if', 'else', 'each', 'mixin', 'media'] }),
? ---------------------
+ require('postcss-at-rules-variables')({ atRules: ['each', 'mixin', 'media'] }),
require('postcss-simple-vars'),
require('postcss-replace')({ pattern: /##/g, data: { replaceAll: '$' } }),
require('postcss-mixins'),
require('postcss-functions')({ functions: styleFunctions }),
require('postcss-each'),
require('postcss-calc'),
require('postcss-fontpath'),
require('postcss-nested'),
require('autoprefixer'),
require('postcss-discard-comments')
]
}
}
}
]
};
}; | 2 | 0.045455 | 1 | 1 |
4588c34efbbef1db59271c2ddb65ec45dbcd739a | website/source/docs/command-line/fix.html.md | website/source/docs/command-line/fix.html.md | ---
description: |
The `packer fix` Packer command takes a template and finds backwards
incompatible parts of it and brings it up to date so it can be used with the
latest version of Packer. After you update to a new Packer release, you should
run the fix command to make sure your templates work with the new release.
layout: docs
page_title: 'Fix - Command-Line'
...
# Command-Line: Fix
The `packer fix` Packer command takes a template and finds backwards
incompatible parts of it and brings it up to date so it can be used with the
latest version of Packer. After you update to a new Packer release, you should
run the fix command to make sure your templates work with the new release.
The fix command will output the changed template to standard out, so you should
redirect standard using standard OS-specific techniques if you want to save it
to a file. For example, on Linux systems, you may want to do this:
\$ packer fix old.json > new.json
If fixing fails for any reason, the fix command will exit with a non-zero exit
status. Error messages appear on standard error, so if you're redirecting
output, you'll still see error messages.
-> **Even when Packer fix doesn't do anything** to the template, the template
will be outputted to standard out. Things such as configuration key ordering and
indentation may be changed. The output format however, is pretty-printed for
human readability.
The full list of fixes that the fix command performs is visible in the help
output, which can be seen via `packer fix -h`.
| ---
description: |
The `packer fix` Packer command takes a template and finds backwards
incompatible parts of it and brings it up to date so it can be used with the
latest version of Packer. After you update to a new Packer release, you should
run the fix command to make sure your templates work with the new release.
layout: docs
page_title: 'Fix - Command-Line'
...
# Command-Line: Fix
The `packer fix` Packer command takes a template and finds backwards
incompatible parts of it and brings it up to date so it can be used with the
latest version of Packer. After you update to a new Packer release, you should
run the fix command to make sure your templates work with the new release.
The fix command will output the changed template to standard out, so you should
redirect standard using standard OS-specific techniques if you want to save it
to a file. For example, on Linux systems, you may want to do this:
$ packer fix old.json > new.json
If fixing fails for any reason, the fix command will exit with a non-zero exit
status. Error messages appear on standard error, so if you're redirecting
output, you'll still see error messages.
-> **Even when Packer fix doesn't do anything** to the template, the template
will be outputted to standard out. Things such as configuration key ordering and
indentation may be changed. The output format however, is pretty-printed for
human readability.
The full list of fixes that the fix command performs is visible in the help
output, which can be seen via `packer fix -h`.
| Fix formatting for fix docs | Fix formatting for fix docs
| Markdown | mpl-2.0 | timsutton/packer,grubernaut/packer,mitchellh/packer,KohlsTechnology/packer,grubernaut/packer,threatstream/packer,tas50/packer,paulmey/packer,mayflower/packer,markpeek/packer,mansunkuo/packer,sean-/packer,paulmey/packer,Akasurde/packer,timsutton/packer,samdunne/packer,mohae/packer,arizvisa/packer,jimmythedog/packer,sneal/packer,taliesins/packer,KohlsTechnology/packer,ChrisLundquist/packer,rnaveiras/packer,markpeek/packer,taliesins/packer,mansunkuo/packer,ChrisLundquist/packer,routelastresort/packer,vtolstov/packer,pvandervelde/packer,life360/packer,pvandervelde/packer,monkeylittleinc/packer,mayflower/packer,bryson/packer,paulmey/packer,tb3088/packer,dave2/packer,Akasurde/packer,mayflower/packer,mayflower/packer,mkuzmin/packer,life360/packer,sean-/packer,sean-/packer,rickard-von-essen/packer,monkeylittleinc/packer,taliesins/packer,dave2/packer,p1-nico/packer,monkeylittleinc/packer,sean-/packer,rickard-von-essen/packer,KohlsTechnology/packer,yoctocloud/packer,sean-/packer,sneal/packer,quantang/packer,quantang/packer,tb3088/packer,pecigonzalo/packer,mitchellh/packer,marc-ta/packer,tsmolka/packer,youhong316/packer,atlassian/packer,routelastresort/packer,mkuzmin/packer,ricardclau/packer,mkuzmin/packer,tsmolka/packer,ricardclau/packer,mefellows/packer,pecigonzalo/packer,sean-/packer,zhuzhih2017/packer,taliesins/packer,sneal/packer,tsmolka/packer,pvandervelde/packer,bryson/packer,pieter-lazzaro/packer,markpeek/packer,grubernaut/packer,c22/packer,mkuzmin/packer,ricardclau/packer,timsutton/packer,taliesins/packer,jimmythedog/packer,tb3088/packer,mafrosis/packer,hashicorp/packer,youhong316/packer,zhuzhih2017/packer,legal90/packer,c22/packer,atlassian/packer,smaato/packer,youhong316/packer,mefellows/packer,tas50/packer,boumenot/packer,timsutton/packer,danschaffer/packer,pvanbuijtene/packer,grubernaut/packer,esemplare/packer,monkeylittleinc/packer,danschaffer/packer,taliesins/packer,jen20/packer,yoctocloud/packer,rnaveiras/packer,tb3088/packer,rickard-von-essen/packer,boumenot/packer,mafrosis/packer,arizvisa/packer,routelastresort/packer,samdunne/packer,markpeek/packer,mefellows/packer,yoctocloud/packer,boumenot/packer,vtolstov/packer,sneal/packer,mitchellh/packer,Akasurde/packer,jen20/packer,samdunne/packer,bryson/packer,ChrisLundquist/packer,samdunne/packer,markpeek/packer,grubernaut/packer,routelastresort/packer,marc-ta/packer,mohae/packer,atlassian/packer,arizvisa/packer,mansunkuo/packer,threatstream/packer,ChrisLundquist/packer,pieter-lazzaro/packer,zhuzhih2017/packer,rnaveiras/packer,esemplare/packer,p1-nico/packer,tb3088/packer,boumenot/packer,pieter-lazzaro/packer,smaato/packer,markpeek/packer,monkeylittleinc/packer,rnaveiras/packer,mansunkuo/packer,esemplare/packer,ricardclau/packer,mwhooker/packer,life360/packer,legal90/packer,arizvisa/packer,atlassian/packer,mkuzmin/packer,quantang/packer,boumenot/packer,atlassian/packer,quantang/packer,tsmolka/packer,danschaffer/packer,mefellows/packer,rickard-von-essen/packer,mwhooker/packer,KohlsTechnology/packer,ricardclau/packer,mansunkuo/packer,dave2/packer,mitchellh/packer,KohlsTechnology/packer,c22/packer,routelastresort/packer,ChrisLundquist/packer,jen20/packer,zhuzhih2017/packer,smaato/packer,dayglojesus/packer,smaato/packer,p1-nico/packer,pvanbuijtene/packer,arizvisa/packer,jen20/packer,bryson/packer,pvandervelde/packer,p1-nico/packer,jimmythedog/packer,legal90/packer,mafrosis/packer,marc-ta/packer,c22/packer,grange74/packer,dave2/packer,mohae/packer,Akasurde/packer,pvandervelde/packer,timsutton/packer,tb3088/packer,yoctocloud/packer,esemplare/packer,mitchellh/packer,life360/packer,grange74/packer,pieter-lazzaro/packer,tsmolka/packer,c22/packer,mohae/packer,rnaveiras/packer,jimmythedog/packer,bryson/packer,mohae/packer,mwhooker/packer,zhuzhih2017/packer,dave2/packer,c22/packer,vtolstov/packer,threatstream/packer,mafrosis/packer,mafrosis/packer,legal90/packer,sneal/packer,pvanbuijtene/packer,pieter-lazzaro/packer,p1-nico/packer,dayglojesus/packer,pecigonzalo/packer,tas50/packer,marc-ta/packer,youhong316/packer,grange74/packer,ChrisLundquist/packer,esemplare/packer,vtolstov/packer,grange74/packer,mayflower/packer,pvandervelde/packer,yoctocloud/packer,jimmythedog/packer,mkuzmin/packer,pieter-lazzaro/packer,timsutton/packer,tas50/packer,dave2/packer,life360/packer,threatstream/packer,smaato/packer,dayglojesus/packer,mwhooker/packer,smaato/packer,legal90/packer,arizvisa/packer,jen20/packer,atlassian/packer,threatstream/packer,Akasurde/packer,mwhooker/packer,danschaffer/packer,dayglojesus/packer,rickard-von-essen/packer,pecigonzalo/packer,zhuzhih2017/packer,paulmey/packer,ricardclau/packer,samdunne/packer,paulmey/packer,youhong316/packer,hashicorp/packer,pvanbuijtene/packer,tas50/packer,p1-nico/packer,vtolstov/packer,hashicorp/packer,grange74/packer,bryson/packer,danschaffer/packer,hashicorp/packer,dayglojesus/packer,KohlsTechnology/packer,quantang/packer,grubernaut/packer,youhong316/packer,paulmey/packer,pvanbuijtene/packer,tsmolka/packer,jen20/packer,rnaveiras/packer,tas50/packer | markdown | ## Code Before:
---
description: |
The `packer fix` Packer command takes a template and finds backwards
incompatible parts of it and brings it up to date so it can be used with the
latest version of Packer. After you update to a new Packer release, you should
run the fix command to make sure your templates work with the new release.
layout: docs
page_title: 'Fix - Command-Line'
...
# Command-Line: Fix
The `packer fix` Packer command takes a template and finds backwards
incompatible parts of it and brings it up to date so it can be used with the
latest version of Packer. After you update to a new Packer release, you should
run the fix command to make sure your templates work with the new release.
The fix command will output the changed template to standard out, so you should
redirect standard using standard OS-specific techniques if you want to save it
to a file. For example, on Linux systems, you may want to do this:
\$ packer fix old.json > new.json
If fixing fails for any reason, the fix command will exit with a non-zero exit
status. Error messages appear on standard error, so if you're redirecting
output, you'll still see error messages.
-> **Even when Packer fix doesn't do anything** to the template, the template
will be outputted to standard out. Things such as configuration key ordering and
indentation may be changed. The output format however, is pretty-printed for
human readability.
The full list of fixes that the fix command performs is visible in the help
output, which can be seen via `packer fix -h`.
## Instruction:
Fix formatting for fix docs
## Code After:
---
description: |
The `packer fix` Packer command takes a template and finds backwards
incompatible parts of it and brings it up to date so it can be used with the
latest version of Packer. After you update to a new Packer release, you should
run the fix command to make sure your templates work with the new release.
layout: docs
page_title: 'Fix - Command-Line'
...
# Command-Line: Fix
The `packer fix` Packer command takes a template and finds backwards
incompatible parts of it and brings it up to date so it can be used with the
latest version of Packer. After you update to a new Packer release, you should
run the fix command to make sure your templates work with the new release.
The fix command will output the changed template to standard out, so you should
redirect standard using standard OS-specific techniques if you want to save it
to a file. For example, on Linux systems, you may want to do this:
$ packer fix old.json > new.json
If fixing fails for any reason, the fix command will exit with a non-zero exit
status. Error messages appear on standard error, so if you're redirecting
output, you'll still see error messages.
-> **Even when Packer fix doesn't do anything** to the template, the template
will be outputted to standard out. Things such as configuration key ordering and
indentation may be changed. The output format however, is pretty-printed for
human readability.
The full list of fixes that the fix command performs is visible in the help
output, which can be seen via `packer fix -h`.
| ---
description: |
The `packer fix` Packer command takes a template and finds backwards
incompatible parts of it and brings it up to date so it can be used with the
latest version of Packer. After you update to a new Packer release, you should
run the fix command to make sure your templates work with the new release.
layout: docs
page_title: 'Fix - Command-Line'
...
# Command-Line: Fix
The `packer fix` Packer command takes a template and finds backwards
incompatible parts of it and brings it up to date so it can be used with the
latest version of Packer. After you update to a new Packer release, you should
run the fix command to make sure your templates work with the new release.
The fix command will output the changed template to standard out, so you should
redirect standard using standard OS-specific techniques if you want to save it
to a file. For example, on Linux systems, you may want to do this:
- \$ packer fix old.json > new.json
? ^ ^^^^
+ $ packer fix old.json > new.json
? ^^^^ ^
If fixing fails for any reason, the fix command will exit with a non-zero exit
status. Error messages appear on standard error, so if you're redirecting
output, you'll still see error messages.
- -> **Even when Packer fix doesn't do anything** to the template, the template
? ^^^^
+ -> **Even when Packer fix doesn't do anything** to the template, the template
? ^
will be outputted to standard out. Things such as configuration key ordering and
indentation may be changed. The output format however, is pretty-printed for
human readability.
The full list of fixes that the fix command performs is visible in the help
output, which can be seen via `packer fix -h`. | 4 | 0.117647 | 2 | 2 |
8c7ee77516f4f07906be059edf20fc62d8716465 | src/server/utils/areObjectValuesDefined.js | src/server/utils/areObjectValuesDefined.js | export default function areObjectValuesDefined(obj) {
return Object.keys(obj)
.map((key) => obj[key])
.reduce((prev, value) => {
return prev && value !== undefined;
}, true);
}
| /**
* Checks if all values in a object are defined
* @param {Object} obj The object
* @return {boolean}
*/
export default function areObjectValuesDefined(obj) {
return Object.keys(obj)
.map((key) => obj[key])
.every(isDefined);
}
function isDefined(value) {
return value !== undefined;
}
| Use every() instead of reduce() | Use every() instead of reduce()
| JavaScript | mit | danistefanovic/hooka | javascript | ## Code Before:
export default function areObjectValuesDefined(obj) {
return Object.keys(obj)
.map((key) => obj[key])
.reduce((prev, value) => {
return prev && value !== undefined;
}, true);
}
## Instruction:
Use every() instead of reduce()
## Code After:
/**
* Checks if all values in a object are defined
* @param {Object} obj The object
* @return {boolean}
*/
export default function areObjectValuesDefined(obj) {
return Object.keys(obj)
.map((key) => obj[key])
.every(isDefined);
}
function isDefined(value) {
return value !== undefined;
}
| + /**
+ * Checks if all values in a object are defined
+ * @param {Object} obj The object
+ * @return {boolean}
+ */
export default function areObjectValuesDefined(obj) {
return Object.keys(obj)
.map((key) => obj[key])
+ .every(isDefined);
- .reduce((prev, value) => {
- return prev && value !== undefined;
- }, true);
}
+
+ function isDefined(value) {
+ return value !== undefined;
+ } | 13 | 1.857143 | 10 | 3 |
f940b3c33783993f7f6364d029fcac37ac0d6ec4 | README.md | README.md |
This is a repository of example [DebOps](http://debops.org/) projects - mostly
Ansible inventories and some additional Bash scripts, when user needs to
perform actions outside Ansible.
Further documentation will be added at a later time.
|
This is a repository of example [DebOps](http://debops.org/) projects - mostly
Ansible inventories and some additional Bash scripts, when user needs to
perform actions outside Ansible.
Further documentation will be added at a later time.
### Basic examples
* `vagrant-multi-machine`: Using Debops with a multi-machine
Vagrantfile
* `webhost-gitusers-dokuwiki`: Setting up a Webserver with nginx,
docuwiki and gituser.
### Advances examples
* `devlab`: manage a "development lab" on an Ubuntu/Xubuntu host with
KVM virtual machines guests, which also are LXC hosts for
containers.
### Other examples
* `vagrant-ansible-single-machine`: Using Ansible (but not debops)
with a single Vagrant box.
| Add short description about contained examples. | Add short description about contained examples. | Markdown | mit | bborysenko/examples,hvisage/examples,hvisage/examples,ypid/examples,OneManOpsTeam/examples,bborysenko/examples,debops/examples,ypid/examples,OneManOpsTeam/examples,debops/examples | markdown | ## Code Before:
This is a repository of example [DebOps](http://debops.org/) projects - mostly
Ansible inventories and some additional Bash scripts, when user needs to
perform actions outside Ansible.
Further documentation will be added at a later time.
## Instruction:
Add short description about contained examples.
## Code After:
This is a repository of example [DebOps](http://debops.org/) projects - mostly
Ansible inventories and some additional Bash scripts, when user needs to
perform actions outside Ansible.
Further documentation will be added at a later time.
### Basic examples
* `vagrant-multi-machine`: Using Debops with a multi-machine
Vagrantfile
* `webhost-gitusers-dokuwiki`: Setting up a Webserver with nginx,
docuwiki and gituser.
### Advances examples
* `devlab`: manage a "development lab" on an Ubuntu/Xubuntu host with
KVM virtual machines guests, which also are LXC hosts for
containers.
### Other examples
* `vagrant-ansible-single-machine`: Using Ansible (but not debops)
with a single Vagrant box.
|
This is a repository of example [DebOps](http://debops.org/) projects - mostly
Ansible inventories and some additional Bash scripts, when user needs to
perform actions outside Ansible.
Further documentation will be added at a later time.
+ ### Basic examples
+
+ * `vagrant-multi-machine`: Using Debops with a multi-machine
+ Vagrantfile
+ * `webhost-gitusers-dokuwiki`: Setting up a Webserver with nginx,
+ docuwiki and gituser.
+
+ ### Advances examples
+
+ * `devlab`: manage a "development lab" on an Ubuntu/Xubuntu host with
+ KVM virtual machines guests, which also are LXC hosts for
+ containers.
+
+ ### Other examples
+
+ * `vagrant-ansible-single-machine`: Using Ansible (but not debops)
+ with a single Vagrant box.
+ | 18 | 2.571429 | 18 | 0 |
035583945d5fc560238e3a8fa6132318bbf54f32 | install.sh | install.sh |
readonly SRC="https://github.com/rinatz/dotfiles/archive/master.tar.gz"
readonly DEST="${HOME}/.dotfiles"
function main() {
mkdir -p "${DEST}"
curl -fsSL "${SRC}" | tar zxv -C "${DEST}" --strip-components 1
local dotfiles
dotfiles=$(find "${DEST}" -name ".*" | grep -v windows)
for dotfile in ${dotfiles[@]}; do
[[ "${dotfile}" == "${DEST}" ]] && continue
ln -svf "${dotfile}" "${HOME}"
done
}
main "$@"
|
readonly SRC="https://github.com/rinatz/dotfiles/archive/master.tar.gz"
readonly DEST="/tmp/dotfiles"
function main() {
mkdir -p "${DEST}"
curl -fsSL "${SRC}" | tar zxv -C "${DEST}" --strip-components 1
local dotfiles
dotfiles=$(find "${DEST}" -name ".*")
for dotfile in ${dotfiles[@]}; do
[[ "${dotfile}" == "${DEST}" ]] && continue
[[ "${dotfile}" == "${DEST}/.git" ]] && continue
if [[ ! $(uname) =~ ^MINGW.*$ ]]; then
[[ "${dotfile}" =~ ^.*\/windows\/.* ]] && continue
fi
\cp -rv "${dotfile}" "${HOME}"
done
rm -rf "${DEST}"
}
main "$@"
| Copy files instead of symlinking | Copy files instead of symlinking
| Shell | unlicense | rinatz/dotfiles | shell | ## Code Before:
readonly SRC="https://github.com/rinatz/dotfiles/archive/master.tar.gz"
readonly DEST="${HOME}/.dotfiles"
function main() {
mkdir -p "${DEST}"
curl -fsSL "${SRC}" | tar zxv -C "${DEST}" --strip-components 1
local dotfiles
dotfiles=$(find "${DEST}" -name ".*" | grep -v windows)
for dotfile in ${dotfiles[@]}; do
[[ "${dotfile}" == "${DEST}" ]] && continue
ln -svf "${dotfile}" "${HOME}"
done
}
main "$@"
## Instruction:
Copy files instead of symlinking
## Code After:
readonly SRC="https://github.com/rinatz/dotfiles/archive/master.tar.gz"
readonly DEST="/tmp/dotfiles"
function main() {
mkdir -p "${DEST}"
curl -fsSL "${SRC}" | tar zxv -C "${DEST}" --strip-components 1
local dotfiles
dotfiles=$(find "${DEST}" -name ".*")
for dotfile in ${dotfiles[@]}; do
[[ "${dotfile}" == "${DEST}" ]] && continue
[[ "${dotfile}" == "${DEST}/.git" ]] && continue
if [[ ! $(uname) =~ ^MINGW.*$ ]]; then
[[ "${dotfile}" =~ ^.*\/windows\/.* ]] && continue
fi
\cp -rv "${dotfile}" "${HOME}"
done
rm -rf "${DEST}"
}
main "$@"
|
readonly SRC="https://github.com/rinatz/dotfiles/archive/master.tar.gz"
- readonly DEST="${HOME}/.dotfiles"
? ------- ^
+ readonly DEST="/tmp/dotfiles"
? ^^^^
function main() {
mkdir -p "${DEST}"
curl -fsSL "${SRC}" | tar zxv -C "${DEST}" --strip-components 1
local dotfiles
- dotfiles=$(find "${DEST}" -name ".*" | grep -v windows)
? ------------------
+ dotfiles=$(find "${DEST}" -name ".*")
for dotfile in ${dotfiles[@]}; do
[[ "${dotfile}" == "${DEST}" ]] && continue
+ [[ "${dotfile}" == "${DEST}/.git" ]] && continue
+ if [[ ! $(uname) =~ ^MINGW.*$ ]]; then
+ [[ "${dotfile}" =~ ^.*\/windows\/.* ]] && continue
+ fi
+
- ln -svf "${dotfile}" "${HOME}"
? ^^ ^ -
+ \cp -rv "${dotfile}" "${HOME}"
? ^^^ ^
done
+
+ rm -rf "${DEST}"
}
main "$@" | 13 | 0.684211 | 10 | 3 |
d0754e21827c804222ffc3963498deb120d3800e | lib/beaglebone/GPIO.rb | lib/beaglebone/GPIO.rb |
module Beaglebone
module GPIO
def self.list
`ls /sys/class/gpio/`
end
end
end
|
module Beaglebone
module GPIO
def self.list
`ls /sys/class/gpio/`.split('\n')
end
end
end
| Convert list string to array | Convert list string to array
| Ruby | mit | DivXZero/beaglebone-rails,DivXZero/beaglebone-rails | ruby | ## Code Before:
module Beaglebone
module GPIO
def self.list
`ls /sys/class/gpio/`
end
end
end
## Instruction:
Convert list string to array
## Code After:
module Beaglebone
module GPIO
def self.list
`ls /sys/class/gpio/`.split('\n')
end
end
end
|
module Beaglebone
module GPIO
def self.list
- `ls /sys/class/gpio/`
+ `ls /sys/class/gpio/`.split('\n')
? ++++++++++++
end
end
end | 2 | 0.25 | 1 | 1 |
faa372f7779a9b8deadfdd50cc6281145ce9ce0e | README.md | README.md |
A Symfony2 distrib improved with FOSUserBundle, DoctrineFixtures and DoctrineExtensions. It also includes Yui-compressor. Everything is preconfigured.
You just have to follow these short steps to start your project:
* copy the app/config/parameter.yml.dist and fill with your database information
* install the vendor with composer
* reconfigure the secured area (default is /admin)
* create your own bundle
* create the User entity in your bundle and change default entry in config.yml
I hope this will help kickstart your projects. |
A Symfony2 distrib improved with FOSUserBundle, DoctrineFixtures and DoctrineExtensions. It also includes Yui-compressor. Everything is preconfigured as possible.
You just have to follow these short steps to start your project:
* copy the app/config/parameter.yml.dist and fill with your database information
* install the vendor with composer
* reconfigure the secured area (default is /admin)
* create your own bundle
* create the User entity in your bundle : [Read the FOSUserBundle Documentation](https://github.com/FriendsOfSymfony/FOSUserBundle/blob/master/Resources/doc/index.md)
* change default entry in app/config/config.yml
* if you use Assetic add your bundle in its configuration in app/config/config.yml
I hope this will help kickstart your projects. | Add explanation in the readme file | Add explanation in the readme file
| Markdown | mit | emerick42/Symfony2-Improved-Distrib,emerick42/Symfony2-Improved-Distrib | markdown | ## Code Before:
A Symfony2 distrib improved with FOSUserBundle, DoctrineFixtures and DoctrineExtensions. It also includes Yui-compressor. Everything is preconfigured.
You just have to follow these short steps to start your project:
* copy the app/config/parameter.yml.dist and fill with your database information
* install the vendor with composer
* reconfigure the secured area (default is /admin)
* create your own bundle
* create the User entity in your bundle and change default entry in config.yml
I hope this will help kickstart your projects.
## Instruction:
Add explanation in the readme file
## Code After:
A Symfony2 distrib improved with FOSUserBundle, DoctrineFixtures and DoctrineExtensions. It also includes Yui-compressor. Everything is preconfigured as possible.
You just have to follow these short steps to start your project:
* copy the app/config/parameter.yml.dist and fill with your database information
* install the vendor with composer
* reconfigure the secured area (default is /admin)
* create your own bundle
* create the User entity in your bundle : [Read the FOSUserBundle Documentation](https://github.com/FriendsOfSymfony/FOSUserBundle/blob/master/Resources/doc/index.md)
* change default entry in app/config/config.yml
* if you use Assetic add your bundle in its configuration in app/config/config.yml
I hope this will help kickstart your projects. |
- A Symfony2 distrib improved with FOSUserBundle, DoctrineFixtures and DoctrineExtensions. It also includes Yui-compressor. Everything is preconfigured.
+ A Symfony2 distrib improved with FOSUserBundle, DoctrineFixtures and DoctrineExtensions. It also includes Yui-compressor. Everything is preconfigured as possible.
? ++++++++++++
You just have to follow these short steps to start your project:
* copy the app/config/parameter.yml.dist and fill with your database information
* install the vendor with composer
* reconfigure the secured area (default is /admin)
* create your own bundle
- * create the User entity in your bundle and change default entry in config.yml
+ * create the User entity in your bundle : [Read the FOSUserBundle Documentation](https://github.com/FriendsOfSymfony/FOSUserBundle/blob/master/Resources/doc/index.md)
+ * change default entry in app/config/config.yml
+ * if you use Assetic add your bundle in its configuration in app/config/config.yml
I hope this will help kickstart your projects. | 6 | 0.545455 | 4 | 2 |
d8b9ff420d60510c8d74de728923a60af78677b4 | node_modules_build/kwf-webpack/trl/trl-html-webpack-plugin.js | node_modules_build/kwf-webpack/trl/trl-html-webpack-plugin.js | 'use strict';
var HtmlWebpackPlugin = require('html-webpack-plugin');
var fs = require('fs');
class TrlHtmlWebpackPlugin extends HtmlWebpackPlugin
{
htmlWebpackPluginAssets(compilation, chunks) {
let language = this.options.language;
let ret = super.htmlWebpackPluginAssets(compilation, chunks);
chunks.forEach(chunk => {
chunk.files.forEach(file => {
if (fs.existsSync(compilation.options.output.path+'/' + language + '.' + file)) {
ret.js.unshift(ret.publicPath + language + '.' + file);
}
})
});
return ret;
}
}
module.exports = TrlHtmlWebpackPlugin;
| 'use strict';
var HtmlWebpackPlugin = require('html-webpack-plugin');
var fs = require('fs');
class TrlHtmlWebpackPlugin extends HtmlWebpackPlugin
{
htmlWebpackPluginAssets(compilation, chunks) {
let language = this.options.language;
let ret = super.htmlWebpackPluginAssets(compilation, chunks);
chunks.forEach(chunk => {
chunk.files.forEach(file => {
if (compilation.assets[language + '.' + file]) {
ret.js.unshift(ret.publicPath + language + '.' + file);
}
})
});
return ret;
}
}
module.exports = TrlHtmlWebpackPlugin;
| Fix trl: Don't access filesystem, file doesn't get written when using dev-server | Fix trl: Don't access filesystem, file doesn't get written when using dev-server
| JavaScript | bsd-2-clause | koala-framework/koala-framework,koala-framework/koala-framework | javascript | ## Code Before:
'use strict';
var HtmlWebpackPlugin = require('html-webpack-plugin');
var fs = require('fs');
class TrlHtmlWebpackPlugin extends HtmlWebpackPlugin
{
htmlWebpackPluginAssets(compilation, chunks) {
let language = this.options.language;
let ret = super.htmlWebpackPluginAssets(compilation, chunks);
chunks.forEach(chunk => {
chunk.files.forEach(file => {
if (fs.existsSync(compilation.options.output.path+'/' + language + '.' + file)) {
ret.js.unshift(ret.publicPath + language + '.' + file);
}
})
});
return ret;
}
}
module.exports = TrlHtmlWebpackPlugin;
## Instruction:
Fix trl: Don't access filesystem, file doesn't get written when using dev-server
## Code After:
'use strict';
var HtmlWebpackPlugin = require('html-webpack-plugin');
var fs = require('fs');
class TrlHtmlWebpackPlugin extends HtmlWebpackPlugin
{
htmlWebpackPluginAssets(compilation, chunks) {
let language = this.options.language;
let ret = super.htmlWebpackPluginAssets(compilation, chunks);
chunks.forEach(chunk => {
chunk.files.forEach(file => {
if (compilation.assets[language + '.' + file]) {
ret.js.unshift(ret.publicPath + language + '.' + file);
}
})
});
return ret;
}
}
module.exports = TrlHtmlWebpackPlugin;
| 'use strict';
var HtmlWebpackPlugin = require('html-webpack-plugin');
var fs = require('fs');
class TrlHtmlWebpackPlugin extends HtmlWebpackPlugin
{
htmlWebpackPluginAssets(compilation, chunks) {
let language = this.options.language;
let ret = super.htmlWebpackPluginAssets(compilation, chunks);
chunks.forEach(chunk => {
chunk.files.forEach(file => {
- if (fs.existsSync(compilation.options.output.path+'/' + language + '.' + file)) {
+ if (compilation.assets[language + '.' + file]) {
ret.js.unshift(ret.publicPath + language + '.' + file);
}
})
});
return ret;
}
}
module.exports = TrlHtmlWebpackPlugin; | 2 | 0.095238 | 1 | 1 |
fb7d488d7cdf7452306c144d4bd78c14cf97fd91 | transifex/arch/extract-and-upload-translations.sh | transifex/arch/extract-and-upload-translations.sh |
set -o nounset # Treat unset variables as an error
SCREEN_WIDTH=${COLUMNS:-$(tput cols 2>/dev/null || echo 80)}
ruller() {
printf '%*s\n' $SCREEN_WIDTH '' | tr ' ' -
}
if [ $# -ne 0 ]; then
ruller
echo "This script does not require any arguments to be passed"
exit 1
fi
ruller
/salt-source/doc/.scripts/setup-transifex-config
ruller
/salt-source/doc/.scripts/update-transifex-source-translations
ruller
echo DONE
|
set -o nounset # Treat unset variables as an error
SCREEN_WIDTH=${COLUMNS:-$(tput cols 2>/dev/null || echo 80)}
ruller() {
printf '%*s\n' $SCREEN_WIDTH '' | tr ' ' -
}
if [ $# -ne 0 ]; then
ruller
echo "This script does not require any arguments to be passed"
exit 1
fi
ruller
/salt-source/doc/.scripts/setup-transifex-config
exc=$?
if [ $exc -ne 0 ]; then
exit $exc
fi
ruller
/salt-source/doc/.scripts/update-transifex-source-translations
exc=$?
ruller
echo DONE
exit $exc
| Exit using the proper exit code. | Exit using the proper exit code.
| Shell | apache-2.0 | saltstack/docker-containers | shell | ## Code Before:
set -o nounset # Treat unset variables as an error
SCREEN_WIDTH=${COLUMNS:-$(tput cols 2>/dev/null || echo 80)}
ruller() {
printf '%*s\n' $SCREEN_WIDTH '' | tr ' ' -
}
if [ $# -ne 0 ]; then
ruller
echo "This script does not require any arguments to be passed"
exit 1
fi
ruller
/salt-source/doc/.scripts/setup-transifex-config
ruller
/salt-source/doc/.scripts/update-transifex-source-translations
ruller
echo DONE
## Instruction:
Exit using the proper exit code.
## Code After:
set -o nounset # Treat unset variables as an error
SCREEN_WIDTH=${COLUMNS:-$(tput cols 2>/dev/null || echo 80)}
ruller() {
printf '%*s\n' $SCREEN_WIDTH '' | tr ' ' -
}
if [ $# -ne 0 ]; then
ruller
echo "This script does not require any arguments to be passed"
exit 1
fi
ruller
/salt-source/doc/.scripts/setup-transifex-config
exc=$?
if [ $exc -ne 0 ]; then
exit $exc
fi
ruller
/salt-source/doc/.scripts/update-transifex-source-translations
exc=$?
ruller
echo DONE
exit $exc
|
set -o nounset # Treat unset variables as an error
SCREEN_WIDTH=${COLUMNS:-$(tput cols 2>/dev/null || echo 80)}
ruller() {
printf '%*s\n' $SCREEN_WIDTH '' | tr ' ' -
}
if [ $# -ne 0 ]; then
ruller
echo "This script does not require any arguments to be passed"
exit 1
fi
ruller
/salt-source/doc/.scripts/setup-transifex-config
+ exc=$?
+
+ if [ $exc -ne 0 ]; then
+ exit $exc
+ fi
ruller
/salt-source/doc/.scripts/update-transifex-source-translations
+ exc=$?
ruller
echo DONE
+ exit $exc | 7 | 0.291667 | 7 | 0 |
90b92bc4680523f37132a081d0ce2fb5d4e967b2 | _config.yml | _config.yml |
paginate: 10 # pagination based on number of posts
paginate_path: "page:num"
exclude: ["README.md"] # files to exclude
highlighter: rouge
markdown: kramdown
plugins:
- jekyll-paginate
- jekyll-seo-tag
google_analytics : UA-72064362-1
defaults:
-
scope:
path: "" # empty string for all files
values:
title: Bhushan Authankar
description: Bhushan Vinod Authankar
author:
name: Bhushan Authankar
email: bhushanauthankar@gmail.com
github: archbloom
twitter: archbloom
instagram: archbloom
pinterest: archbloom
linkedin: bhushan-authankar-05953abb
bio: It will change our Lives
email_md5:
rss_path: feed.xml
categories_path: categories.html
tags_path: tags.html
about_path: about.html
style_css_path: assets/css/style.css
BASE_PATH:
|
paginate: 10 # pagination based on number of posts
paginate_path: "page:num"
exclude: ["README.md"] # files to exclude
highlighter: rouge
markdown: kramdown
plugins:
- jekyll-paginate
- jekyll-seo-tag
- jekyll-sitemap
google_analytics : UA-72064362-1
permalink: /:title
defaults:
-
scope:
path: "" # empty string for all files
values:
title: Bhushan Authankar
description: Bhushan Vinod Authankar
author:
name: Bhushan Authankar
email: bhushanauthankar@gmail.com
github: archbloom
twitter: archbloom
instagram: archbloom
pinterest: archbloom
linkedin: bhushan-authankar-05953abb
bio: It will change our Lives
email_md5:
rss_path: feed.xml
categories_path: categories.html
tags_path: tags.html
about_path: about.html
style_css_path: assets/css/style.css
BASE_PATH:
| Add sitemap plugin to config | Add sitemap plugin to config
| YAML | mit | archbloom/archbloom.github.io,archbloom/archbloom.github.io,archbloom/archbloom.github.io | yaml | ## Code Before:
paginate: 10 # pagination based on number of posts
paginate_path: "page:num"
exclude: ["README.md"] # files to exclude
highlighter: rouge
markdown: kramdown
plugins:
- jekyll-paginate
- jekyll-seo-tag
google_analytics : UA-72064362-1
defaults:
-
scope:
path: "" # empty string for all files
values:
title: Bhushan Authankar
description: Bhushan Vinod Authankar
author:
name: Bhushan Authankar
email: bhushanauthankar@gmail.com
github: archbloom
twitter: archbloom
instagram: archbloom
pinterest: archbloom
linkedin: bhushan-authankar-05953abb
bio: It will change our Lives
email_md5:
rss_path: feed.xml
categories_path: categories.html
tags_path: tags.html
about_path: about.html
style_css_path: assets/css/style.css
BASE_PATH:
## Instruction:
Add sitemap plugin to config
## Code After:
paginate: 10 # pagination based on number of posts
paginate_path: "page:num"
exclude: ["README.md"] # files to exclude
highlighter: rouge
markdown: kramdown
plugins:
- jekyll-paginate
- jekyll-seo-tag
- jekyll-sitemap
google_analytics : UA-72064362-1
permalink: /:title
defaults:
-
scope:
path: "" # empty string for all files
values:
title: Bhushan Authankar
description: Bhushan Vinod Authankar
author:
name: Bhushan Authankar
email: bhushanauthankar@gmail.com
github: archbloom
twitter: archbloom
instagram: archbloom
pinterest: archbloom
linkedin: bhushan-authankar-05953abb
bio: It will change our Lives
email_md5:
rss_path: feed.xml
categories_path: categories.html
tags_path: tags.html
about_path: about.html
style_css_path: assets/css/style.css
BASE_PATH:
|
paginate: 10 # pagination based on number of posts
paginate_path: "page:num"
exclude: ["README.md"] # files to exclude
highlighter: rouge
markdown: kramdown
plugins:
- jekyll-paginate
- jekyll-seo-tag
+ - jekyll-sitemap
google_analytics : UA-72064362-1
+ permalink: /:title
defaults:
-
scope:
path: "" # empty string for all files
values:
title: Bhushan Authankar
description: Bhushan Vinod Authankar
author:
name: Bhushan Authankar
email: bhushanauthankar@gmail.com
github: archbloom
twitter: archbloom
instagram: archbloom
pinterest: archbloom
linkedin: bhushan-authankar-05953abb
bio: It will change our Lives
email_md5:
rss_path: feed.xml
categories_path: categories.html
tags_path: tags.html
about_path: about.html
style_css_path: assets/css/style.css
BASE_PATH: | 2 | 0.052632 | 2 | 0 |
b8a5ba58e62899008845ad58911ebf21ab330117 | requirements.txt | requirements.txt | django==1.8.3
psycopg2==2.6.1
pytz==2015.4
django-councilmatic==0.4.1
| django==1.8.3
psycopg2==2.6.1
pytz==2015.4
# django-councilmatic==0.4.1
-e git+https://github.com/datamade/django-councilmatic.git#egg=django-councilmatic
| Change to pull django councilmatic from github | Change to pull django councilmatic from github
| Text | mit | CivicTechTO/tor-councilmatic,patcon/sfo-councilmatic,datamade/chi-councilmatic,tor-councilmatic/tor-councilmatic,datamade/chi-councilmatic,tor-councilmatic/tor-councilmatic,patcon/sfo-councilmatic,datamade/chi-councilmatic,datamade/chi-councilmatic,CivicTechTO/tor-councilmatic,patcon/sfo-councilmatic,tor-councilmatic/tor-councilmatic,datamade/chi-councilmatic,CivicTechTO/tor-councilmatic | text | ## Code Before:
django==1.8.3
psycopg2==2.6.1
pytz==2015.4
django-councilmatic==0.4.1
## Instruction:
Change to pull django councilmatic from github
## Code After:
django==1.8.3
psycopg2==2.6.1
pytz==2015.4
# django-councilmatic==0.4.1
-e git+https://github.com/datamade/django-councilmatic.git#egg=django-councilmatic
| django==1.8.3
psycopg2==2.6.1
pytz==2015.4
+
- django-councilmatic==0.4.1
+ # django-councilmatic==0.4.1
? ++
+
+ -e git+https://github.com/datamade/django-councilmatic.git#egg=django-councilmatic | 5 | 1.25 | 4 | 1 |
5ab889f1a77ec0390ff2a20b80c215f9acc5d050 | CONTRIBUTING.md | CONTRIBUTING.md | How to contribute
=================
1. Fork the repo.
1. Clone the repo.
1. Run `npm install`.
1. Make your changes. Please add tests!
1. Make sure `npm test` doesn't have any errors.
1. Make sure `npm run hint` doesn't give any JSHint errors.
1. Submit your pull request!
Thanks so much.
| How to contribute
=================
1. Fork the repo.
1. Clone the repo.
1. Run `npm install`.
1. Make your changes. Please add tests! If adding a new language, define some tests in *test/definitions*.
1. Make sure `npm test` doesn't have any errors.
1. Make sure `npm run hint` doesn't give any JSHint errors.
1. Submit your pull request!
Thanks so much.
| Add note about contributing a new language | Add note about contributing a new language
| Markdown | unlicense | smnbbrv/HumanizeDuration.js,EvanHahn/HumanizeDuration.js,TioNoob/HumanizeDuration.js | markdown | ## Code Before:
How to contribute
=================
1. Fork the repo.
1. Clone the repo.
1. Run `npm install`.
1. Make your changes. Please add tests!
1. Make sure `npm test` doesn't have any errors.
1. Make sure `npm run hint` doesn't give any JSHint errors.
1. Submit your pull request!
Thanks so much.
## Instruction:
Add note about contributing a new language
## Code After:
How to contribute
=================
1. Fork the repo.
1. Clone the repo.
1. Run `npm install`.
1. Make your changes. Please add tests! If adding a new language, define some tests in *test/definitions*.
1. Make sure `npm test` doesn't have any errors.
1. Make sure `npm run hint` doesn't give any JSHint errors.
1. Submit your pull request!
Thanks so much.
| How to contribute
=================
1. Fork the repo.
1. Clone the repo.
1. Run `npm install`.
- 1. Make your changes. Please add tests!
+ 1. Make your changes. Please add tests! If adding a new language, define some tests in *test/definitions*.
1. Make sure `npm test` doesn't have any errors.
1. Make sure `npm run hint` doesn't give any JSHint errors.
1. Submit your pull request!
Thanks so much. | 2 | 0.166667 | 1 | 1 |
010794696bd14f807a25086ef516df26f0834412 | spec/data_set/sleep_spec.rb | spec/data_set/sleep_spec.rb | require 'spec_helper'
describe RubyJawbone::DataSet::Sleep do
#
end
| require 'spec_helper'
describe RubyJawbone::DataSet::Sleep do
let(:date) { Date.new(2014, 1, 21) }
let(:bed_time) { Time.local(2014, 1, 20, 22, 59) }
let(:asleep_time) { Time.local(2014, 1, 20, 23, 11) }
let(:awake_time) { Time.local(2014, 1, 21, 8, 4) }
let(:total_time_asleep) { 27478 }
let(:total_time_awake) { 5222 }
let(:total_time_in_light_sleep) { 17338 }
let(:total_time_in_deep_sleep) { 10140 }
let(:times_woken_up) { 3 }
let(:sleep) { RubyJawbone::DataSet::Sleep.new(date, bed_time, asleep_time, awake_time, total_time_asleep, total_time_awake, total_time_in_light_sleep, total_time_in_deep_sleep, times_woken_up) }
describe "#initialize" do
it "receives the sleep values and sets them into the correct properties" do
expect(sleep.date).to eq date
expect(sleep.bed_time).to eq bed_time
expect(sleep.asleep_time).to eq asleep_time
expect(sleep.awake_time).to eq awake_time
expect(sleep.total_time_asleep).to eq total_time_asleep
expect(sleep.total_time_awake).to eq total_time_awake
expect(sleep.total_time_in_light_sleep).to eq total_time_in_light_sleep
expect(sleep.total_time_in_deep_sleep).to eq total_time_in_deep_sleep
expect(sleep.times_woken_up).to eq times_woken_up
end
end
end
| Add spec to describe initializing a sleep data set object with sleep values. | Add spec to describe initializing a sleep data set object with sleep values.
| Ruby | mit | GlenCrawford/ruby_jawbone | ruby | ## Code Before:
require 'spec_helper'
describe RubyJawbone::DataSet::Sleep do
#
end
## Instruction:
Add spec to describe initializing a sleep data set object with sleep values.
## Code After:
require 'spec_helper'
describe RubyJawbone::DataSet::Sleep do
let(:date) { Date.new(2014, 1, 21) }
let(:bed_time) { Time.local(2014, 1, 20, 22, 59) }
let(:asleep_time) { Time.local(2014, 1, 20, 23, 11) }
let(:awake_time) { Time.local(2014, 1, 21, 8, 4) }
let(:total_time_asleep) { 27478 }
let(:total_time_awake) { 5222 }
let(:total_time_in_light_sleep) { 17338 }
let(:total_time_in_deep_sleep) { 10140 }
let(:times_woken_up) { 3 }
let(:sleep) { RubyJawbone::DataSet::Sleep.new(date, bed_time, asleep_time, awake_time, total_time_asleep, total_time_awake, total_time_in_light_sleep, total_time_in_deep_sleep, times_woken_up) }
describe "#initialize" do
it "receives the sleep values and sets them into the correct properties" do
expect(sleep.date).to eq date
expect(sleep.bed_time).to eq bed_time
expect(sleep.asleep_time).to eq asleep_time
expect(sleep.awake_time).to eq awake_time
expect(sleep.total_time_asleep).to eq total_time_asleep
expect(sleep.total_time_awake).to eq total_time_awake
expect(sleep.total_time_in_light_sleep).to eq total_time_in_light_sleep
expect(sleep.total_time_in_deep_sleep).to eq total_time_in_deep_sleep
expect(sleep.times_woken_up).to eq times_woken_up
end
end
end
| require 'spec_helper'
describe RubyJawbone::DataSet::Sleep do
- #
+ let(:date) { Date.new(2014, 1, 21) }
+ let(:bed_time) { Time.local(2014, 1, 20, 22, 59) }
+ let(:asleep_time) { Time.local(2014, 1, 20, 23, 11) }
+ let(:awake_time) { Time.local(2014, 1, 21, 8, 4) }
+ let(:total_time_asleep) { 27478 }
+ let(:total_time_awake) { 5222 }
+ let(:total_time_in_light_sleep) { 17338 }
+ let(:total_time_in_deep_sleep) { 10140 }
+ let(:times_woken_up) { 3 }
+
+ let(:sleep) { RubyJawbone::DataSet::Sleep.new(date, bed_time, asleep_time, awake_time, total_time_asleep, total_time_awake, total_time_in_light_sleep, total_time_in_deep_sleep, times_woken_up) }
+
+ describe "#initialize" do
+ it "receives the sleep values and sets them into the correct properties" do
+ expect(sleep.date).to eq date
+ expect(sleep.bed_time).to eq bed_time
+ expect(sleep.asleep_time).to eq asleep_time
+ expect(sleep.awake_time).to eq awake_time
+ expect(sleep.total_time_asleep).to eq total_time_asleep
+ expect(sleep.total_time_awake).to eq total_time_awake
+ expect(sleep.total_time_in_light_sleep).to eq total_time_in_light_sleep
+ expect(sleep.total_time_in_deep_sleep).to eq total_time_in_deep_sleep
+ expect(sleep.times_woken_up).to eq times_woken_up
+ end
+ end
end | 26 | 5.2 | 25 | 1 |
37aa5265bc137b48ca5f15de725357216c4ef7c2 | src/routes/invalidHeaders.js | src/routes/invalidHeaders.js | const _ = require('lodash');
const router = require('express').Router();
router.get('/cookie', (req, res) => {
const data = JSON.stringify(req.headers);
let response = '';
response += 'HTTP/1.1 200 OK\r\n';
response += 'Connection: close\r\n';
response += `Content-Length: ${data.length}\r\n`;
response += 'Content-Type: application/json\r\n';
response += `Cookie: ${_.repeat('a', req.query.length)}\r\n`;
response += 'Date: Wed, 30 Nov 2016 20:19:57 GMT\r\n\r\n';
response += data;
res.socket.end(response);
res.socket.destroy();
});
router.get('/content-length', (req, res) => {
let response = '';
response += 'HTTP/1.1 200 OK\r\n';
response += 'Connection: close\r\n';
response += 'Content-Length: 3\r\n';
response += 'Content-Type: text/plain\r\n';
response += 'Date: Wed, 30 Nov 2016 20:19:57 GMT\r\n\r\n';
response += 'abcdefghijklmnopqrstuvwxyz';
res.socket.end(response);
res.socket.destroy();
});
module.exports = router;
| const _ = require('lodash');
const router = require('express').Router();
router.get('/cookie', (req, res) => {
const data = JSON.stringify(req.headers);
let response = '';
response += 'HTTP/1.1 200 OK\r\n';
response += 'Connection: close\r\n';
response += `Content-Length: ${data.length}\r\n`;
response += 'Content-Type: application/json\r\n';
response += `Cookie: ${_.repeat('a', req.query.length)}\r\n`;
response += 'Date: Wed, 30 Nov 2016 20:19:57 GMT\r\n\r\n';
response += data;
res.socket.end(response);
res.socket.destroy();
});
router.get('/content-length', (req, res) => {
let response = '';
response += 'HTTP/1.1 200 OK\r\n';
response += 'Connection: close\r\n';
response += 'Content-Length: 3\r\n';
response += 'Access-Control-Allow-Origin: *\r\n';
response += 'Content-Type: text/plain\r\n';
response += 'Date: Wed, 30 Nov 2016 20:19:57 GMT\r\n\r\n';
response += 'abcdefghijklmnopqrstuvwxyz';
res.socket.end(response);
res.socket.destroy();
});
module.exports = router;
| Add cors header for browser testing | Add cors header for browser testing
| JavaScript | mit | apiaryio/node-hamms | javascript | ## Code Before:
const _ = require('lodash');
const router = require('express').Router();
router.get('/cookie', (req, res) => {
const data = JSON.stringify(req.headers);
let response = '';
response += 'HTTP/1.1 200 OK\r\n';
response += 'Connection: close\r\n';
response += `Content-Length: ${data.length}\r\n`;
response += 'Content-Type: application/json\r\n';
response += `Cookie: ${_.repeat('a', req.query.length)}\r\n`;
response += 'Date: Wed, 30 Nov 2016 20:19:57 GMT\r\n\r\n';
response += data;
res.socket.end(response);
res.socket.destroy();
});
router.get('/content-length', (req, res) => {
let response = '';
response += 'HTTP/1.1 200 OK\r\n';
response += 'Connection: close\r\n';
response += 'Content-Length: 3\r\n';
response += 'Content-Type: text/plain\r\n';
response += 'Date: Wed, 30 Nov 2016 20:19:57 GMT\r\n\r\n';
response += 'abcdefghijklmnopqrstuvwxyz';
res.socket.end(response);
res.socket.destroy();
});
module.exports = router;
## Instruction:
Add cors header for browser testing
## Code After:
const _ = require('lodash');
const router = require('express').Router();
router.get('/cookie', (req, res) => {
const data = JSON.stringify(req.headers);
let response = '';
response += 'HTTP/1.1 200 OK\r\n';
response += 'Connection: close\r\n';
response += `Content-Length: ${data.length}\r\n`;
response += 'Content-Type: application/json\r\n';
response += `Cookie: ${_.repeat('a', req.query.length)}\r\n`;
response += 'Date: Wed, 30 Nov 2016 20:19:57 GMT\r\n\r\n';
response += data;
res.socket.end(response);
res.socket.destroy();
});
router.get('/content-length', (req, res) => {
let response = '';
response += 'HTTP/1.1 200 OK\r\n';
response += 'Connection: close\r\n';
response += 'Content-Length: 3\r\n';
response += 'Access-Control-Allow-Origin: *\r\n';
response += 'Content-Type: text/plain\r\n';
response += 'Date: Wed, 30 Nov 2016 20:19:57 GMT\r\n\r\n';
response += 'abcdefghijklmnopqrstuvwxyz';
res.socket.end(response);
res.socket.destroy();
});
module.exports = router;
| const _ = require('lodash');
const router = require('express').Router();
router.get('/cookie', (req, res) => {
const data = JSON.stringify(req.headers);
let response = '';
response += 'HTTP/1.1 200 OK\r\n';
response += 'Connection: close\r\n';
response += `Content-Length: ${data.length}\r\n`;
response += 'Content-Type: application/json\r\n';
response += `Cookie: ${_.repeat('a', req.query.length)}\r\n`;
response += 'Date: Wed, 30 Nov 2016 20:19:57 GMT\r\n\r\n';
response += data;
res.socket.end(response);
res.socket.destroy();
});
router.get('/content-length', (req, res) => {
let response = '';
response += 'HTTP/1.1 200 OK\r\n';
response += 'Connection: close\r\n';
response += 'Content-Length: 3\r\n';
+ response += 'Access-Control-Allow-Origin: *\r\n';
response += 'Content-Type: text/plain\r\n';
response += 'Date: Wed, 30 Nov 2016 20:19:57 GMT\r\n\r\n';
response += 'abcdefghijklmnopqrstuvwxyz';
res.socket.end(response);
res.socket.destroy();
});
module.exports = router; | 1 | 0.030303 | 1 | 0 |
11ef542aaf042c15cc67dfc1627b1added75b16c | README.md | README.md | A .Net framework for calling stored procedures. This is an on-going project and will be strongly based upon the great work carried out by "bluemoonsailor" at "Mindless Passenger". See link: https://mindlesspassenger.wordpress.com/2011/02/02/code-first-and-stored-procedures/
| A .Net framework for calling stored procedures. This is an on-going project and will be strongly based upon the great work carried out by "bluemoonsailor" at "Mindless Passenger". See link: https://mindlesspassenger.wordpress.com/2011/02/02/code-first-and-stored-procedures/
## Project Brief
The aim of this project is to provide the following:
* (Must) Ability to support a POCO that represent a stored procedure
* (Must) Ability to support a PCOC that represents a row that is returned by a stored procedure
* (Must) Ability to support a PCOC that represents the parameters
* (Must) Ability to execute the stored procedure represented by the POCO against DBConnection using extensions
* (Must) Ability to execute the stored procedure represented by the POCO against SqlConnection using extensions
* (Must) Ability to execute the stored procedure represented by the POCO against DBContext using extensions
* (Must) Ability to handle output parameters
* (Must) Ability to handle all common parameter types
* (Must) Ability to handle all common return data types
* (Must) Contain a suite of unit tests that test all public accessors
* (Must) Contain a suite of integration tests that document usage of the assembly
* (Should) Ability to handle lesser used parameter types
* (Should) Ability to handle lesser used return data types
* (Should) Ability to handle multiple recordsets returned from a stored procedure
* (Should) warn calling code if parameter value data may be truncated due to smaller pameter type
| Update ReadMe.md to include project brief | Update ReadMe.md to include project brief | Markdown | mit | dibley1973/StoredProcedureFramework | markdown | ## Code Before:
A .Net framework for calling stored procedures. This is an on-going project and will be strongly based upon the great work carried out by "bluemoonsailor" at "Mindless Passenger". See link: https://mindlesspassenger.wordpress.com/2011/02/02/code-first-and-stored-procedures/
## Instruction:
Update ReadMe.md to include project brief
## Code After:
A .Net framework for calling stored procedures. This is an on-going project and will be strongly based upon the great work carried out by "bluemoonsailor" at "Mindless Passenger". See link: https://mindlesspassenger.wordpress.com/2011/02/02/code-first-and-stored-procedures/
## Project Brief
The aim of this project is to provide the following:
* (Must) Ability to support a POCO that represent a stored procedure
* (Must) Ability to support a PCOC that represents a row that is returned by a stored procedure
* (Must) Ability to support a PCOC that represents the parameters
* (Must) Ability to execute the stored procedure represented by the POCO against DBConnection using extensions
* (Must) Ability to execute the stored procedure represented by the POCO against SqlConnection using extensions
* (Must) Ability to execute the stored procedure represented by the POCO against DBContext using extensions
* (Must) Ability to handle output parameters
* (Must) Ability to handle all common parameter types
* (Must) Ability to handle all common return data types
* (Must) Contain a suite of unit tests that test all public accessors
* (Must) Contain a suite of integration tests that document usage of the assembly
* (Should) Ability to handle lesser used parameter types
* (Should) Ability to handle lesser used return data types
* (Should) Ability to handle multiple recordsets returned from a stored procedure
* (Should) warn calling code if parameter value data may be truncated due to smaller pameter type
| A .Net framework for calling stored procedures. This is an on-going project and will be strongly based upon the great work carried out by "bluemoonsailor" at "Mindless Passenger". See link: https://mindlesspassenger.wordpress.com/2011/02/02/code-first-and-stored-procedures/
+
+ ## Project Brief
+ The aim of this project is to provide the following:
+ * (Must) Ability to support a POCO that represent a stored procedure
+ * (Must) Ability to support a PCOC that represents a row that is returned by a stored procedure
+ * (Must) Ability to support a PCOC that represents the parameters
+ * (Must) Ability to execute the stored procedure represented by the POCO against DBConnection using extensions
+ * (Must) Ability to execute the stored procedure represented by the POCO against SqlConnection using extensions
+ * (Must) Ability to execute the stored procedure represented by the POCO against DBContext using extensions
+ * (Must) Ability to handle output parameters
+ * (Must) Ability to handle all common parameter types
+ * (Must) Ability to handle all common return data types
+ * (Must) Contain a suite of unit tests that test all public accessors
+ * (Must) Contain a suite of integration tests that document usage of the assembly
+ * (Should) Ability to handle lesser used parameter types
+ * (Should) Ability to handle lesser used return data types
+ * (Should) Ability to handle multiple recordsets returned from a stored procedure
+ * (Should) warn calling code if parameter value data may be truncated due to smaller pameter type | 18 | 18 | 18 | 0 |
44beaca6f2f681122e78437e454e52075d17da1b | db/sql/547__alter_table_dw_f_virta_otp_kvliikkuvuus_add_cimo_opintoaste.sql | db/sql/547__alter_table_dw_f_virta_otp_kvliikkuvuus_add_cimo_opintoaste.sql | IF EXISTS (
select * from INFORMATION_SCHEMA.COLUMNS
where TABLE_SCHEMA='dw' and TABLE_NAME='f_virta_otp_kvliikkuvuus'
and COLUMN_NAME='d_opiskeluoikeudentyyppi_id'
) BEGIN
ALTER TABLE dw.f_virta_otp_kvliikkuvuus DROP COLUMN d_opiskeluoikeudentyyppi_id
END
IF NOT EXISTS (
select * from INFORMATION_SCHEMA.COLUMNS
where TABLE_SCHEMA='dw' and TABLE_NAME='f_virta_otp_kvliikkuvuus'
and COLUMN_NAME='cimo_opintoaste'
) BEGIN
ALTER TABLE dw.f_virta_otp_kvliikkuvuus ADD cimo_opintoaste varchar(2) NULL
END
/* revert
ALTER TABLE dw.f_virta_otp_kvliikkuvuus ADD d_opiskeluoikeudentyyppi_id int NOT NULL
ALTER TABLE dw.f_virta_otp_kvliikkuvuus DROP COLUMN cimo_opintoaste
--*/
| IF EXISTS (
select * from INFORMATION_SCHEMA.COLUMNS
where TABLE_SCHEMA='dw' and TABLE_NAME='f_virta_otp_kvliikkuvuus'
and COLUMN_NAME='d_opiskeluoikeudentyyppi_id'
) BEGIN
ALTER TABLE dw.f_virta_otp_kvliikkuvuus DROP COLUMN d_opiskeluoikeudentyyppi_id
END
IF NOT EXISTS (
select * from INFORMATION_SCHEMA.COLUMNS
where TABLE_SCHEMA='dw' and TABLE_NAME='f_virta_otp_kvliikkuvuus'
and COLUMN_NAME='cimo_opintoaste'
) BEGIN
ALTER TABLE dw.f_virta_otp_kvliikkuvuus ADD cimo_opintoaste varchar(2) NULL
END
/* revert
-- should empty table before or something similar!
ALTER TABLE dw.f_virta_otp_kvliikkuvuus ADD d_opiskeluoikeudentyyppi_id int NOT NULL
ALTER TABLE dw.f_virta_otp_kvliikkuvuus DROP COLUMN cimo_opintoaste
--*/
| Add some note (just to get scripts executed again). | Add some note (just to get scripts executed again).
| SQL | mit | CSCfi/antero,CSCfi/antero,CSCfi/antero,CSCfi/antero,CSCfi/antero,CSCfi/antero | sql | ## Code Before:
IF EXISTS (
select * from INFORMATION_SCHEMA.COLUMNS
where TABLE_SCHEMA='dw' and TABLE_NAME='f_virta_otp_kvliikkuvuus'
and COLUMN_NAME='d_opiskeluoikeudentyyppi_id'
) BEGIN
ALTER TABLE dw.f_virta_otp_kvliikkuvuus DROP COLUMN d_opiskeluoikeudentyyppi_id
END
IF NOT EXISTS (
select * from INFORMATION_SCHEMA.COLUMNS
where TABLE_SCHEMA='dw' and TABLE_NAME='f_virta_otp_kvliikkuvuus'
and COLUMN_NAME='cimo_opintoaste'
) BEGIN
ALTER TABLE dw.f_virta_otp_kvliikkuvuus ADD cimo_opintoaste varchar(2) NULL
END
/* revert
ALTER TABLE dw.f_virta_otp_kvliikkuvuus ADD d_opiskeluoikeudentyyppi_id int NOT NULL
ALTER TABLE dw.f_virta_otp_kvliikkuvuus DROP COLUMN cimo_opintoaste
--*/
## Instruction:
Add some note (just to get scripts executed again).
## Code After:
IF EXISTS (
select * from INFORMATION_SCHEMA.COLUMNS
where TABLE_SCHEMA='dw' and TABLE_NAME='f_virta_otp_kvliikkuvuus'
and COLUMN_NAME='d_opiskeluoikeudentyyppi_id'
) BEGIN
ALTER TABLE dw.f_virta_otp_kvliikkuvuus DROP COLUMN d_opiskeluoikeudentyyppi_id
END
IF NOT EXISTS (
select * from INFORMATION_SCHEMA.COLUMNS
where TABLE_SCHEMA='dw' and TABLE_NAME='f_virta_otp_kvliikkuvuus'
and COLUMN_NAME='cimo_opintoaste'
) BEGIN
ALTER TABLE dw.f_virta_otp_kvliikkuvuus ADD cimo_opintoaste varchar(2) NULL
END
/* revert
-- should empty table before or something similar!
ALTER TABLE dw.f_virta_otp_kvliikkuvuus ADD d_opiskeluoikeudentyyppi_id int NOT NULL
ALTER TABLE dw.f_virta_otp_kvliikkuvuus DROP COLUMN cimo_opintoaste
--*/
| IF EXISTS (
select * from INFORMATION_SCHEMA.COLUMNS
where TABLE_SCHEMA='dw' and TABLE_NAME='f_virta_otp_kvliikkuvuus'
and COLUMN_NAME='d_opiskeluoikeudentyyppi_id'
) BEGIN
ALTER TABLE dw.f_virta_otp_kvliikkuvuus DROP COLUMN d_opiskeluoikeudentyyppi_id
END
IF NOT EXISTS (
select * from INFORMATION_SCHEMA.COLUMNS
where TABLE_SCHEMA='dw' and TABLE_NAME='f_virta_otp_kvliikkuvuus'
and COLUMN_NAME='cimo_opintoaste'
) BEGIN
ALTER TABLE dw.f_virta_otp_kvliikkuvuus ADD cimo_opintoaste varchar(2) NULL
END
/* revert
+ -- should empty table before or something similar!
ALTER TABLE dw.f_virta_otp_kvliikkuvuus ADD d_opiskeluoikeudentyyppi_id int NOT NULL
ALTER TABLE dw.f_virta_otp_kvliikkuvuus DROP COLUMN cimo_opintoaste
--*/ | 1 | 0.05 | 1 | 0 |
4809778bc9e985b54c23b5ce36a17a2567625a04 | README.rst | README.rst | Django-Analog
=============
Simple per-model log models for Django apps.
|Build Status| |Coverage Status| |Documentation Status|
Compatibility
-------------
* Django 1.8+
* Python 2.7 or Python 3.4+
Basic Usage
-----------
.. code:: python
from django.db import models
from analog import define_log_model
class MyModel(models.Model):
value = models.IntegerField(default=0)
MyModelLogEntry = define_log_model(MyModel)
m = MyModel.objects.create(value=42)
m.add_log_entry('Something occurred')
assert m.log_entries.last().message == 'Something occurred'
Development
-----------
::
pip install -e .
pip install -r requirements-dev.txt
Tests
~~~~~
::
py.test
Documentation
~~~~~~~~~~~~~
::
sphinx-build -b html docs docs/_build
.. |Build Status|
image:: https://travis-ci.org/andersinno/django-analog.svg?branch=master
:target: https://travis-ci.org/andersinno/django-analog
.. |Coverage Status|
image:: https://codecov.io/gh/andersinno/django-analog/branch/master/graph/badge.svg
:target: https://codecov.io/gh/andersinno/django-analog
.. |Documentation Status|
image:: https://readthedocs.org/projects/django-analog/badge/?version=latest
:target: http://django-analog.readthedocs.org/en/latest/?badge=latest
| Django-Analog
=============
Simple per-model log models for Django apps.
|Build Status| |Coverage Status| |Documentation Status|
Compatibility
-------------
* Django 1.8+
* Python 2.7 or Python 3.4+
Basic Usage
-----------
.. code:: python
from django.db import models
from analog import define_log_model
class MyModel(models.Model):
value = models.IntegerField(default=0)
MyModelLogEntry = define_log_model(MyModel)
m = MyModel.objects.create(value=42)
m.add_log_entry('Something occurred')
assert m.log_entries.last().message == 'Something occurred'
Development
-----------
::
pip install -e .
pip install -r requirements-dev.txt
Tests
~~~~~
::
py.test
Documentation
~~~~~~~~~~~~~
::
sphinx-build -b html docs docs/_build
.. |Build Status|
image:: https://github.com/andersinno/django-analog/workflows/Test/badge.svg
:target: https://github.com/andersinno/django-analog/actions
.. |Coverage Status|
image:: https://codecov.io/gh/andersinno/django-analog/branch/master/graph/badge.svg
:target: https://codecov.io/gh/andersinno/django-analog
.. |Documentation Status|
image:: https://readthedocs.org/projects/django-analog/badge/?version=latest
:target: http://django-analog.readthedocs.org/en/latest/?badge=latest
| Update CI badge to point to GitHub | Update CI badge to point to GitHub
| reStructuredText | mit | andersinno/django-analog | restructuredtext | ## Code Before:
Django-Analog
=============
Simple per-model log models for Django apps.
|Build Status| |Coverage Status| |Documentation Status|
Compatibility
-------------
* Django 1.8+
* Python 2.7 or Python 3.4+
Basic Usage
-----------
.. code:: python
from django.db import models
from analog import define_log_model
class MyModel(models.Model):
value = models.IntegerField(default=0)
MyModelLogEntry = define_log_model(MyModel)
m = MyModel.objects.create(value=42)
m.add_log_entry('Something occurred')
assert m.log_entries.last().message == 'Something occurred'
Development
-----------
::
pip install -e .
pip install -r requirements-dev.txt
Tests
~~~~~
::
py.test
Documentation
~~~~~~~~~~~~~
::
sphinx-build -b html docs docs/_build
.. |Build Status|
image:: https://travis-ci.org/andersinno/django-analog.svg?branch=master
:target: https://travis-ci.org/andersinno/django-analog
.. |Coverage Status|
image:: https://codecov.io/gh/andersinno/django-analog/branch/master/graph/badge.svg
:target: https://codecov.io/gh/andersinno/django-analog
.. |Documentation Status|
image:: https://readthedocs.org/projects/django-analog/badge/?version=latest
:target: http://django-analog.readthedocs.org/en/latest/?badge=latest
## Instruction:
Update CI badge to point to GitHub
## Code After:
Django-Analog
=============
Simple per-model log models for Django apps.
|Build Status| |Coverage Status| |Documentation Status|
Compatibility
-------------
* Django 1.8+
* Python 2.7 or Python 3.4+
Basic Usage
-----------
.. code:: python
from django.db import models
from analog import define_log_model
class MyModel(models.Model):
value = models.IntegerField(default=0)
MyModelLogEntry = define_log_model(MyModel)
m = MyModel.objects.create(value=42)
m.add_log_entry('Something occurred')
assert m.log_entries.last().message == 'Something occurred'
Development
-----------
::
pip install -e .
pip install -r requirements-dev.txt
Tests
~~~~~
::
py.test
Documentation
~~~~~~~~~~~~~
::
sphinx-build -b html docs docs/_build
.. |Build Status|
image:: https://github.com/andersinno/django-analog/workflows/Test/badge.svg
:target: https://github.com/andersinno/django-analog/actions
.. |Coverage Status|
image:: https://codecov.io/gh/andersinno/django-analog/branch/master/graph/badge.svg
:target: https://codecov.io/gh/andersinno/django-analog
.. |Documentation Status|
image:: https://readthedocs.org/projects/django-analog/badge/?version=latest
:target: http://django-analog.readthedocs.org/en/latest/?badge=latest
| Django-Analog
=============
Simple per-model log models for Django apps.
|Build Status| |Coverage Status| |Documentation Status|
Compatibility
-------------
* Django 1.8+
* Python 2.7 or Python 3.4+
Basic Usage
-----------
.. code:: python
from django.db import models
from analog import define_log_model
class MyModel(models.Model):
value = models.IntegerField(default=0)
MyModelLogEntry = define_log_model(MyModel)
m = MyModel.objects.create(value=42)
m.add_log_entry('Something occurred')
assert m.log_entries.last().message == 'Something occurred'
Development
-----------
::
pip install -e .
pip install -r requirements-dev.txt
Tests
~~~~~
::
py.test
Documentation
~~~~~~~~~~~~~
::
sphinx-build -b html docs docs/_build
.. |Build Status|
- image:: https://travis-ci.org/andersinno/django-analog.svg?branch=master
+ image:: https://github.com/andersinno/django-analog/workflows/Test/badge.svg
- :target: https://travis-ci.org/andersinno/django-analog
? ^^^^^^ -- ^^
+ :target: https://github.com/andersinno/django-analog/actions
? ++ ^^^^ ^ ++++++++
.. |Coverage Status|
image:: https://codecov.io/gh/andersinno/django-analog/branch/master/graph/badge.svg
:target: https://codecov.io/gh/andersinno/django-analog
.. |Documentation Status|
image:: https://readthedocs.org/projects/django-analog/badge/?version=latest
:target: http://django-analog.readthedocs.org/en/latest/?badge=latest | 4 | 0.065574 | 2 | 2 |
2f1b7da82b5f6c6057475b1efdbcb03cb6479f8b | server/requirements.txt | server/requirements.txt | gunicorn==19.3.0
honcho==0.6.6
python3-ldap==0.9.8.4
behave==1.2.5
wooper==0.4.2
git+git://github.com/superdesk/eve@75aa97d#egg=Eve==0.6.0-dev
-e git+git://github.com/superdesk/superdesk-core@7676f6a#egg=Superdesk-Core==0.0.1-dev
| gunicorn==19.3.0
honcho==0.6.6
python3-ldap==0.9.8.4
behave==1.2.5
wooper==0.4.2
pymongo==2.8
git+git://github.com/nicolaiarocci/eve@21ac82b#egg=Eve==0.6.0-dev
-e git+git://github.com/superdesk/superdesk-core@2f553d15#egg=Superdesk-Core==0.0.1-dev
| Prepare for Eve 0.6 release: | Prepare for Eve 0.6 release:
| Text | agpl-3.0 | fritzSF/superdesk,superdesk/superdesk-ntb,akintolga/superdesk,petrjasek/superdesk-ntb,sjunaid/superdesk,darconny/superdesk,sivakuna-aap/superdesk,plamut/superdesk,amagdas/superdesk,pavlovicnemanja/superdesk,ancafarcas/superdesk,fritzSF/superdesk,ioanpocol/superdesk,ioanpocol/superdesk,superdesk/superdesk-aap,pavlovicnemanja92/superdesk,thnkloud9/superdesk,amagdas/superdesk,verifiedpixel/superdesk,plamut/superdesk,ancafarcas/superdesk,pavlovicnemanja/superdesk,gbbr/superdesk,sjunaid/superdesk,marwoodandrew/superdesk,superdesk/superdesk-aap,liveblog/superdesk,mdhaman/superdesk-aap,superdesk/superdesk-aap,ioanpocol/superdesk-ntb,fritzSF/superdesk,superdesk/superdesk,fritzSF/superdesk,ioanpocol/superdesk-ntb,akintolga/superdesk,pavlovicnemanja92/superdesk,mdhaman/superdesk,petrjasek/superdesk,akintolga/superdesk,amagdas/superdesk,sjunaid/superdesk,sivakuna-aap/superdesk,superdesk/superdesk-ntb,petrjasek/superdesk,liveblog/superdesk,superdesk/superdesk,marwoodandrew/superdesk,superdesk/superdesk,gbbr/superdesk,superdesk/superdesk-ntb,hlmnrmr/superdesk,petrjasek/superdesk,Aca-jov/superdesk,thnkloud9/superdesk,plamut/superdesk,sivakuna-aap/superdesk,mugurrus/superdesk,petrjasek/superdesk-ntb,mugurrus/superdesk,mugurrus/superdesk,marwoodandrew/superdesk-aap,liveblog/superdesk,verifiedpixel/superdesk,mdhaman/superdesk-aap,marwoodandrew/superdesk,akintolga/superdesk-aap,hlmnrmr/superdesk,hlmnrmr/superdesk,amagdas/superdesk,mdhaman/superdesk,liveblog/superdesk,Aca-jov/superdesk,pavlovicnemanja92/superdesk,marwoodandrew/superdesk-aap,verifiedpixel/superdesk,akintolga/superdesk-aap,akintolga/superdesk,gbbr/superdesk,pavlovicnemanja/superdesk,verifiedpixel/superdesk,marwoodandrew/superdesk-aap,pavlovicnemanja92/superdesk,superdesk/superdesk-aap,sivakuna-aap/superdesk,marwoodandrew/superdesk-aap,akintolga/superdesk-aap,marwoodandrew/superdesk,superdesk/superdesk-ntb,liveblog/superdesk,fritzSF/superdesk,Aca-jov/superdesk,darconny/superdesk,petrjasek/superdesk,petrjasek/superdesk-ntb,akintolga/superdesk-aap,thnkloud9/superdesk,ioanpocol/superdesk,amagdas/superdesk,darconny/superdesk,plamut/superdesk,ioanpocol/superdesk-ntb,sivakuna-aap/superdesk,plamut/superdesk,akintolga/superdesk,marwoodandrew/superdesk,petrjasek/superdesk-ntb,pavlovicnemanja92/superdesk,verifiedpixel/superdesk,superdesk/superdesk,mdhaman/superdesk,pavlovicnemanja/superdesk,ancafarcas/superdesk,mdhaman/superdesk-aap,mdhaman/superdesk-aap | text | ## Code Before:
gunicorn==19.3.0
honcho==0.6.6
python3-ldap==0.9.8.4
behave==1.2.5
wooper==0.4.2
git+git://github.com/superdesk/eve@75aa97d#egg=Eve==0.6.0-dev
-e git+git://github.com/superdesk/superdesk-core@7676f6a#egg=Superdesk-Core==0.0.1-dev
## Instruction:
Prepare for Eve 0.6 release:
## Code After:
gunicorn==19.3.0
honcho==0.6.6
python3-ldap==0.9.8.4
behave==1.2.5
wooper==0.4.2
pymongo==2.8
git+git://github.com/nicolaiarocci/eve@21ac82b#egg=Eve==0.6.0-dev
-e git+git://github.com/superdesk/superdesk-core@2f553d15#egg=Superdesk-Core==0.0.1-dev
| gunicorn==19.3.0
honcho==0.6.6
python3-ldap==0.9.8.4
behave==1.2.5
wooper==0.4.2
+ pymongo==2.8
- git+git://github.com/superdesk/eve@75aa97d#egg=Eve==0.6.0-dev
+ git+git://github.com/nicolaiarocci/eve@21ac82b#egg=Eve==0.6.0-dev
- -e git+git://github.com/superdesk/superdesk-core@7676f6a#egg=Superdesk-Core==0.0.1-dev
? ^^^^ ^^
+ -e git+git://github.com/superdesk/superdesk-core@2f553d15#egg=Superdesk-Core==0.0.1-dev
? ^ ^^^^^^
| 5 | 0.625 | 3 | 2 |
3fe40e91f70e8256d7c86c46f866e82e3ccf26e2 | commandment/profiles/cert.py | commandment/profiles/cert.py | '''
Copyright (c) 2015 Jesse Peterson
Licensed under the MIT license. See the included LICENSE.txt file for details.
'''
from . import Payload
import plistlib # needed for Data() wrapper
class PEMCertificatePayload(Payload):
'''PEM-encoded certificate without private key. May contain root
certificates.
Payload type of "com.apple.security.pem". Further encodes cert_data as
plistlib.Data instance (Base64 data).'''
payload_type = 'com.apple.security.pem'
def __init__(self, identifier, cert_data, uuid=None, **kwargs):
Payload.__init__(self, self.payload_type, identifier, uuid, **kwargs)
self.payload['PayloadContent'] = plistlib.Data(cert_data)
class PKCS12CertificatePayload(Payload):
'''Password-protected identity certificate. Only one certificate may be
included.
Payload type of "com.apple.security.pkcs12". Include a PKCS#12 (.p12)
identity as cert_data. Further encodes cert_data as plistlib.Data instance
(Base64 data). Include a password argument for the PKCS#12 identity.'''
payload_type = 'com.apple.security.pkcs12'
def __init__(self, identifier, cert_data, password=None, uuid=None, **kwargs):
Payload.__init__(self, self.payload_type, identifier, uuid, **kwargs)
self.payload['PayloadContent'] = plistlib.Data(cert_data)
if password:
self.payload['Password'] = password
| '''
Copyright (c) 2015 Jesse Peterson
Licensed under the MIT license. See the included LICENSE.txt file for details.
'''
from . import Payload
import plistlib # needed for Data() wrapper
class PEMCertificatePayload(Payload):
'''PEM-encoded certificate without private key. May contain root
certificates.
Payload type of "com.apple.security.pem". Further encodes cert_data as
plistlib.Data instance (Base64 data).'''
payload_type = 'com.apple.security.pem'
def __init__(self, identifier, cert_data, uuid=None, **kwargs):
kwargs['PayloadContent'] = plistlib.Data(cert_data)
Payload.__init__(self, self.payload_type, identifier, uuid, **kwargs)
class PKCS12CertificatePayload(Payload):
'''Password-protected identity certificate. Only one certificate may be
included.
Payload type of "com.apple.security.pkcs12". Include a PKCS#12 (.p12)
identity as cert_data. Further encodes cert_data as plistlib.Data instance
(Base64 data). Include a password argument for the PKCS#12 identity.'''
payload_type = 'com.apple.security.pkcs12'
def __init__(self, identifier, cert_data, password=None, uuid=None, **kwargs):
kwargs['PayloadContent'] = plistlib.Data(cert_data)
if password:
kwargs['Password'] = password
Payload.__init__(self, self.payload_type, identifier, uuid, **kwargs)
| Change style of Payload suclasses to better encapsulate internal structure | Change style of Payload suclasses to better encapsulate internal structure
| Python | mit | mosen/commandment,jessepeterson/commandment,mosen/commandment,mosen/commandment,mosen/commandment,jessepeterson/commandment,mosen/commandment | python | ## Code Before:
'''
Copyright (c) 2015 Jesse Peterson
Licensed under the MIT license. See the included LICENSE.txt file for details.
'''
from . import Payload
import plistlib # needed for Data() wrapper
class PEMCertificatePayload(Payload):
'''PEM-encoded certificate without private key. May contain root
certificates.
Payload type of "com.apple.security.pem". Further encodes cert_data as
plistlib.Data instance (Base64 data).'''
payload_type = 'com.apple.security.pem'
def __init__(self, identifier, cert_data, uuid=None, **kwargs):
Payload.__init__(self, self.payload_type, identifier, uuid, **kwargs)
self.payload['PayloadContent'] = plistlib.Data(cert_data)
class PKCS12CertificatePayload(Payload):
'''Password-protected identity certificate. Only one certificate may be
included.
Payload type of "com.apple.security.pkcs12". Include a PKCS#12 (.p12)
identity as cert_data. Further encodes cert_data as plistlib.Data instance
(Base64 data). Include a password argument for the PKCS#12 identity.'''
payload_type = 'com.apple.security.pkcs12'
def __init__(self, identifier, cert_data, password=None, uuid=None, **kwargs):
Payload.__init__(self, self.payload_type, identifier, uuid, **kwargs)
self.payload['PayloadContent'] = plistlib.Data(cert_data)
if password:
self.payload['Password'] = password
## Instruction:
Change style of Payload suclasses to better encapsulate internal structure
## Code After:
'''
Copyright (c) 2015 Jesse Peterson
Licensed under the MIT license. See the included LICENSE.txt file for details.
'''
from . import Payload
import plistlib # needed for Data() wrapper
class PEMCertificatePayload(Payload):
'''PEM-encoded certificate without private key. May contain root
certificates.
Payload type of "com.apple.security.pem". Further encodes cert_data as
plistlib.Data instance (Base64 data).'''
payload_type = 'com.apple.security.pem'
def __init__(self, identifier, cert_data, uuid=None, **kwargs):
kwargs['PayloadContent'] = plistlib.Data(cert_data)
Payload.__init__(self, self.payload_type, identifier, uuid, **kwargs)
class PKCS12CertificatePayload(Payload):
'''Password-protected identity certificate. Only one certificate may be
included.
Payload type of "com.apple.security.pkcs12". Include a PKCS#12 (.p12)
identity as cert_data. Further encodes cert_data as plistlib.Data instance
(Base64 data). Include a password argument for the PKCS#12 identity.'''
payload_type = 'com.apple.security.pkcs12'
def __init__(self, identifier, cert_data, password=None, uuid=None, **kwargs):
kwargs['PayloadContent'] = plistlib.Data(cert_data)
if password:
kwargs['Password'] = password
Payload.__init__(self, self.payload_type, identifier, uuid, **kwargs)
| '''
Copyright (c) 2015 Jesse Peterson
Licensed under the MIT license. See the included LICENSE.txt file for details.
'''
from . import Payload
import plistlib # needed for Data() wrapper
class PEMCertificatePayload(Payload):
'''PEM-encoded certificate without private key. May contain root
certificates.
Payload type of "com.apple.security.pem". Further encodes cert_data as
plistlib.Data instance (Base64 data).'''
payload_type = 'com.apple.security.pem'
def __init__(self, identifier, cert_data, uuid=None, **kwargs):
+ kwargs['PayloadContent'] = plistlib.Data(cert_data)
Payload.__init__(self, self.payload_type, identifier, uuid, **kwargs)
-
- self.payload['PayloadContent'] = plistlib.Data(cert_data)
class PKCS12CertificatePayload(Payload):
'''Password-protected identity certificate. Only one certificate may be
included.
Payload type of "com.apple.security.pkcs12". Include a PKCS#12 (.p12)
identity as cert_data. Further encodes cert_data as plistlib.Data instance
(Base64 data). Include a password argument for the PKCS#12 identity.'''
payload_type = 'com.apple.security.pkcs12'
def __init__(self, identifier, cert_data, password=None, uuid=None, **kwargs):
+ kwargs['PayloadContent'] = plistlib.Data(cert_data)
+ if password:
+ kwargs['Password'] = password
Payload.__init__(self, self.payload_type, identifier, uuid, **kwargs)
-
- self.payload['PayloadContent'] = plistlib.Data(cert_data)
-
- if password:
- self.payload['Password'] = password | 11 | 0.282051 | 4 | 7 |
d36bb068fe3a122df9543cace2ead6b711d5f372 | test/suite/SafeCWD.hs | test/suite/SafeCWD.hs | module SafeCWD
( inDir
, removeDirectoryRecursiveSafe
) where
import Control.Concurrent.QSem
import Control.Exception
import Control.Monad
import System.Directory
import System.IO.Unsafe
sem :: QSem
sem = unsafePerformIO $ newQSem 1
inDir :: Bool -> FilePath -> IO a -> IO a
inDir startClean dir action = bracket before after (const action)
where
before = do
waitQSem sem
cwd <- getCurrentDirectory
when startClean $ removeDirectoryRecursiveSafe dir
createDirectoryIfMissing True dir
setCurrentDirectory dir
return cwd
after cwd = do
setCurrentDirectory cwd
signalQSem sem
removeDirectoryRecursiveSafe p =
doesDirectoryExist p >>= flip when (removeDirectoryRecursive p)
| module SafeCWD
( inDir
, removeDirectoryRecursiveSafe
) where
import Control.Concurrent.QSem
import Control.Exception
import Control.Monad
import System.Directory
import System.IO.Unsafe
sem :: QSem
sem = unsafePerformIO $ newQSem 1
{-# NOINLINE sem #-}
inDir :: Bool -> FilePath -> IO a -> IO a
inDir startClean dir action = bracket before after (const action)
where
before = do
waitQSem sem
cwd <- getCurrentDirectory
when startClean $ removeDirectoryRecursiveSafe dir
createDirectoryIfMissing True dir
setCurrentDirectory dir
return cwd
after cwd = do
setCurrentDirectory cwd
signalQSem sem
removeDirectoryRecursiveSafe p =
doesDirectoryExist p >>= flip when (removeDirectoryRecursive p)
| Add a NOINLINE around unsafePerformIO-guarded semaphore in testsuite. | Add a NOINLINE around unsafePerformIO-guarded semaphore in testsuite.
| Haskell | bsd-3-clause | 23Skidoo/snap,snapframework/snap,bitemyapp/snap,sopvop/snap | haskell | ## Code Before:
module SafeCWD
( inDir
, removeDirectoryRecursiveSafe
) where
import Control.Concurrent.QSem
import Control.Exception
import Control.Monad
import System.Directory
import System.IO.Unsafe
sem :: QSem
sem = unsafePerformIO $ newQSem 1
inDir :: Bool -> FilePath -> IO a -> IO a
inDir startClean dir action = bracket before after (const action)
where
before = do
waitQSem sem
cwd <- getCurrentDirectory
when startClean $ removeDirectoryRecursiveSafe dir
createDirectoryIfMissing True dir
setCurrentDirectory dir
return cwd
after cwd = do
setCurrentDirectory cwd
signalQSem sem
removeDirectoryRecursiveSafe p =
doesDirectoryExist p >>= flip when (removeDirectoryRecursive p)
## Instruction:
Add a NOINLINE around unsafePerformIO-guarded semaphore in testsuite.
## Code After:
module SafeCWD
( inDir
, removeDirectoryRecursiveSafe
) where
import Control.Concurrent.QSem
import Control.Exception
import Control.Monad
import System.Directory
import System.IO.Unsafe
sem :: QSem
sem = unsafePerformIO $ newQSem 1
{-# NOINLINE sem #-}
inDir :: Bool -> FilePath -> IO a -> IO a
inDir startClean dir action = bracket before after (const action)
where
before = do
waitQSem sem
cwd <- getCurrentDirectory
when startClean $ removeDirectoryRecursiveSafe dir
createDirectoryIfMissing True dir
setCurrentDirectory dir
return cwd
after cwd = do
setCurrentDirectory cwd
signalQSem sem
removeDirectoryRecursiveSafe p =
doesDirectoryExist p >>= flip when (removeDirectoryRecursive p)
| module SafeCWD
( inDir
, removeDirectoryRecursiveSafe
) where
import Control.Concurrent.QSem
import Control.Exception
import Control.Monad
import System.Directory
import System.IO.Unsafe
sem :: QSem
sem = unsafePerformIO $ newQSem 1
+ {-# NOINLINE sem #-}
+
inDir :: Bool -> FilePath -> IO a -> IO a
inDir startClean dir action = bracket before after (const action)
where
before = do
waitQSem sem
cwd <- getCurrentDirectory
when startClean $ removeDirectoryRecursiveSafe dir
createDirectoryIfMissing True dir
setCurrentDirectory dir
return cwd
after cwd = do
setCurrentDirectory cwd
signalQSem sem
+
removeDirectoryRecursiveSafe p =
doesDirectoryExist p >>= flip when (removeDirectoryRecursive p) | 3 | 0.1 | 3 | 0 |
651200c44df19a443a2637c89bd38a2dbf94e5ea | test/test_jekyll_mentions.rb | test/test_jekyll_mentions.rb | require 'helper'
class TestJekyllMentions < Minitest::Test
include MentionsTestHelpers
def setup
@site = fixture_site
@site.read
@mentions = Jekyll::Mentions.new(@site)
@mention = "test <a href='https://github.com/test' class='user-mention'>@test</a> test"
end
should "replace @mention with link" do
page = page_with_name(@site, "index.md")
@mentions.mentionify page
assert_equal @mention, page.content
end
should "replace page content on generate" do
@mentions.generate(@site)
assert_equal @mention, @site.pages.first.content
end
should "not mangle liquid templating" do
page = page_with_name(@site, "leave-liquid-alone.md")
@mentions.mentionify page
assert_equal "#{@mention}<a href=\"{{ test }}\">test</a>", page.content
end
should "not mangle markdown" do
page = page_with_name(@site, "mentioned-markdown.md")
@mentions.mentionify page
assert_equal "#{@mention}\n> test", page.content
end
should "not mangle non-mentioned content" do
page = page_with_name(@site, "non-mentioned.md")
@mentions.mentionify page
assert_equal "test test test\n> test", page.content
end
should "not touch non-HTML pages" do
assert_equal "test @test test", page_with_name(@site, "test.json").content
end
end
| require 'helper'
class TestJekyllMentions < Minitest::Test
include MentionsTestHelpers
def setup
@site = fixture_site
@site.read
@mentions = Jekyll::Mentions.new(@site)
@mention = "test <a href='https://github.com/test' class='user-mention'>@test</a> test"
end
should "replace @mention with link" do
page = page_with_name(@site, "index.md")
@mentions.mentionify page
assert_equal @mention, page.content
end
should "replace page content on generate" do
@mentions.generate(@site)
assert_equal @mention, @site.pages.first.content
end
should "not mangle liquid templating" do
page = page_with_name(@site, "leave-liquid-alone.md")
@mentions.mentionify page
assert_equal "#{@mention}<a href=\"{{ test }}\">test</a>", page.content
end
should "not mangle markdown" do
page = page_with_name(@site, "mentioned-markdown.md")
@mentions.mentionify page
assert_equal "#{@mention}\n> test", page.content
end
should "not mangle non-mentioned content" do
page = page_with_name(@site, "non-mentioned.md")
@mentions.mentionify page
assert_equal "test test test\n> test", page.content
end
should "not touch non-HTML pages" do
@mentions.generate(@site)
assert_equal "test @test test", page_with_name(@site, "test.json").content
end
end
| Make sure to generate the site in the test about the non-HTML files. | Make sure to generate the site in the test about the non-HTML files.
| Ruby | mit | jekyll/jekyll-mentions,jekyll/jekyll-mentions | ruby | ## Code Before:
require 'helper'
class TestJekyllMentions < Minitest::Test
include MentionsTestHelpers
def setup
@site = fixture_site
@site.read
@mentions = Jekyll::Mentions.new(@site)
@mention = "test <a href='https://github.com/test' class='user-mention'>@test</a> test"
end
should "replace @mention with link" do
page = page_with_name(@site, "index.md")
@mentions.mentionify page
assert_equal @mention, page.content
end
should "replace page content on generate" do
@mentions.generate(@site)
assert_equal @mention, @site.pages.first.content
end
should "not mangle liquid templating" do
page = page_with_name(@site, "leave-liquid-alone.md")
@mentions.mentionify page
assert_equal "#{@mention}<a href=\"{{ test }}\">test</a>", page.content
end
should "not mangle markdown" do
page = page_with_name(@site, "mentioned-markdown.md")
@mentions.mentionify page
assert_equal "#{@mention}\n> test", page.content
end
should "not mangle non-mentioned content" do
page = page_with_name(@site, "non-mentioned.md")
@mentions.mentionify page
assert_equal "test test test\n> test", page.content
end
should "not touch non-HTML pages" do
assert_equal "test @test test", page_with_name(@site, "test.json").content
end
end
## Instruction:
Make sure to generate the site in the test about the non-HTML files.
## Code After:
require 'helper'
class TestJekyllMentions < Minitest::Test
include MentionsTestHelpers
def setup
@site = fixture_site
@site.read
@mentions = Jekyll::Mentions.new(@site)
@mention = "test <a href='https://github.com/test' class='user-mention'>@test</a> test"
end
should "replace @mention with link" do
page = page_with_name(@site, "index.md")
@mentions.mentionify page
assert_equal @mention, page.content
end
should "replace page content on generate" do
@mentions.generate(@site)
assert_equal @mention, @site.pages.first.content
end
should "not mangle liquid templating" do
page = page_with_name(@site, "leave-liquid-alone.md")
@mentions.mentionify page
assert_equal "#{@mention}<a href=\"{{ test }}\">test</a>", page.content
end
should "not mangle markdown" do
page = page_with_name(@site, "mentioned-markdown.md")
@mentions.mentionify page
assert_equal "#{@mention}\n> test", page.content
end
should "not mangle non-mentioned content" do
page = page_with_name(@site, "non-mentioned.md")
@mentions.mentionify page
assert_equal "test test test\n> test", page.content
end
should "not touch non-HTML pages" do
@mentions.generate(@site)
assert_equal "test @test test", page_with_name(@site, "test.json").content
end
end
| require 'helper'
class TestJekyllMentions < Minitest::Test
include MentionsTestHelpers
def setup
@site = fixture_site
@site.read
@mentions = Jekyll::Mentions.new(@site)
@mention = "test <a href='https://github.com/test' class='user-mention'>@test</a> test"
end
should "replace @mention with link" do
page = page_with_name(@site, "index.md")
@mentions.mentionify page
assert_equal @mention, page.content
end
should "replace page content on generate" do
@mentions.generate(@site)
assert_equal @mention, @site.pages.first.content
end
should "not mangle liquid templating" do
page = page_with_name(@site, "leave-liquid-alone.md")
@mentions.mentionify page
assert_equal "#{@mention}<a href=\"{{ test }}\">test</a>", page.content
end
should "not mangle markdown" do
page = page_with_name(@site, "mentioned-markdown.md")
@mentions.mentionify page
assert_equal "#{@mention}\n> test", page.content
end
should "not mangle non-mentioned content" do
page = page_with_name(@site, "non-mentioned.md")
@mentions.mentionify page
assert_equal "test test test\n> test", page.content
end
should "not touch non-HTML pages" do
+ @mentions.generate(@site)
assert_equal "test @test test", page_with_name(@site, "test.json").content
end
end | 1 | 0.02 | 1 | 0 |
89c11946dc2ecb0736be805b06a559cf35a5c664 | impressionist.gemspec | impressionist.gemspec | require File.expand_path('../lib/impressionist/version', __FILE__)
Gem::Specification.new do |s|
s.add_dependency 'httpclient', '~> 2.2'
s.add_dependency 'nokogiri', '~> 1.5'
s.add_development_dependency 'rdoc', '>= 2.4.2'
s.authors = ["cowboycoded"]
s.description = "Log impressions from controller actions or from a model"
s.email = "john.mcaliley@gmail.com"
s.files = `git ls-files`.split("\n")
s.homepage = "https://github.com/charlotte-ruby/impressionist"
s.licenses = ["MIT"]
s.name = "impressionist"
s.require_paths = ["lib"]
s.required_rubygems_version = Gem::Requirement.new('>= 1.3.6') if s.respond_to? :required_rubygems_version=
s.summary = "Easy way to log impressions"
s.test_files = `git ls-files -- test_app/*`.split("\n")
s.version = Impressionist::VERSION
end
| require File.expand_path('../lib/impressionist/version', __FILE__)
Gem::Specification.new do |s|
s.add_dependency 'httpclient', '~> 2.2'
s.add_dependency 'nokogiri', '~> 1.5'
s.add_development_dependency 'rake', '>= 0.9'
s.add_development_dependency 'rdoc', '>= 2.4.2'
s.authors = ["cowboycoded"]
s.description = "Log impressions from controller actions or from a model"
s.email = "john.mcaliley@gmail.com"
s.files = `git ls-files`.split("\n")
s.homepage = "https://github.com/charlotte-ruby/impressionist"
s.licenses = ["MIT"]
s.name = "impressionist"
s.require_paths = ["lib"]
s.required_rubygems_version = Gem::Requirement.new('>= 1.3.6') if s.respond_to? :required_rubygems_version=
s.summary = "Easy way to log impressions"
s.test_files = `git ls-files -- test_app/*`.split("\n")
s.version = Impressionist::VERSION
end
| Add rake as a development dependency | Add rake as a development dependency
| Ruby | mit | guodongme/impressionist,firmanm/impressionist,samchen2009/impressionist,samchen2009/impressionist,charlotte-ruby/impressionist,guodongme/impressionist,charlotte-ruby/impressionist,guodongme/impressionist,samchen2009/impressionist,firmanm/impressionist,charlotte-ruby/impressionist,firmanm/impressionist | ruby | ## Code Before:
require File.expand_path('../lib/impressionist/version', __FILE__)
Gem::Specification.new do |s|
s.add_dependency 'httpclient', '~> 2.2'
s.add_dependency 'nokogiri', '~> 1.5'
s.add_development_dependency 'rdoc', '>= 2.4.2'
s.authors = ["cowboycoded"]
s.description = "Log impressions from controller actions or from a model"
s.email = "john.mcaliley@gmail.com"
s.files = `git ls-files`.split("\n")
s.homepage = "https://github.com/charlotte-ruby/impressionist"
s.licenses = ["MIT"]
s.name = "impressionist"
s.require_paths = ["lib"]
s.required_rubygems_version = Gem::Requirement.new('>= 1.3.6') if s.respond_to? :required_rubygems_version=
s.summary = "Easy way to log impressions"
s.test_files = `git ls-files -- test_app/*`.split("\n")
s.version = Impressionist::VERSION
end
## Instruction:
Add rake as a development dependency
## Code After:
require File.expand_path('../lib/impressionist/version', __FILE__)
Gem::Specification.new do |s|
s.add_dependency 'httpclient', '~> 2.2'
s.add_dependency 'nokogiri', '~> 1.5'
s.add_development_dependency 'rake', '>= 0.9'
s.add_development_dependency 'rdoc', '>= 2.4.2'
s.authors = ["cowboycoded"]
s.description = "Log impressions from controller actions or from a model"
s.email = "john.mcaliley@gmail.com"
s.files = `git ls-files`.split("\n")
s.homepage = "https://github.com/charlotte-ruby/impressionist"
s.licenses = ["MIT"]
s.name = "impressionist"
s.require_paths = ["lib"]
s.required_rubygems_version = Gem::Requirement.new('>= 1.3.6') if s.respond_to? :required_rubygems_version=
s.summary = "Easy way to log impressions"
s.test_files = `git ls-files -- test_app/*`.split("\n")
s.version = Impressionist::VERSION
end
| require File.expand_path('../lib/impressionist/version', __FILE__)
Gem::Specification.new do |s|
s.add_dependency 'httpclient', '~> 2.2'
s.add_dependency 'nokogiri', '~> 1.5'
+ s.add_development_dependency 'rake', '>= 0.9'
s.add_development_dependency 'rdoc', '>= 2.4.2'
s.authors = ["cowboycoded"]
s.description = "Log impressions from controller actions or from a model"
s.email = "john.mcaliley@gmail.com"
s.files = `git ls-files`.split("\n")
s.homepage = "https://github.com/charlotte-ruby/impressionist"
s.licenses = ["MIT"]
s.name = "impressionist"
s.require_paths = ["lib"]
s.required_rubygems_version = Gem::Requirement.new('>= 1.3.6') if s.respond_to? :required_rubygems_version=
s.summary = "Easy way to log impressions"
s.test_files = `git ls-files -- test_app/*`.split("\n")
s.version = Impressionist::VERSION
end | 1 | 0.052632 | 1 | 0 |
d3c435bb558d8bb3e3f44224fb8844a838acb608 | lib/hacky_hal/device_controllers/linux_computer.rb | lib/hacky_hal/device_controllers/linux_computer.rb | require_relative "generic_ssh"
module HackyHAL
module DeviceControllers
class LinuxComputer < GenericSsh
def mirror_screens(source_screen, dest_screen)
xrandr_command("--output #{dest_screen} --same-as #{source_screen}")
end
def set_screen_position(screen_1, screen_2, position)
xrandr_command("--output #{screen_1} --#{position}-of #{screen_2}")
end
def reset_display_settings(screen)
xrandr_command("--output #{screen} --auto")
end
private
def xrandr_command(options)
x_env_variables = [
"export CONSOLE=`sudo fgconsole`",
"export SESSION=`who -s | grep tty$CONSOLE | tr '()' ' '`",
"export DISPLAY=`echo $SESSION | awk '{print $5}'`",
"export XUSER=`echo $SESSION | awk '{print $1}'`",
"export XAUTHORITY=/home/$XUSER/.Xauthority"
].join("; ")
command = "#{x_env_variables}; xrandr -d $DISPLAY #{options}"
exec(command)
end
end
end
end
| require_relative "generic_ssh"
module HackyHAL
module DeviceControllers
class LinuxComputer < GenericSsh
def mirror_screens(source_screen, dest_screen)
xrandr_command("--output #{dest_screen} --same-as #{source_screen}")
end
def set_screen_position(screen_1, screen_2, position)
xrandr_command("--output #{screen_1} --#{position}-of #{screen_2}")
end
def reset_display_settings(screen)
xrandr_command("--output #{screen} --auto")
end
private
def xrandr_command(options)
x_env_variables = [
"CONSOLE=$(sudo fgconsole)",
"SESSION=$(who -s | grep tty$CONSOLE | tr '()' ' ')",
"XUSER=$(echo $SESSION | awk '{print $1}')",
"DISPLAY=$(echo $SESSION | awk '{print $5}')",
"XAUTHORITY=/home/$XUSER/.Xauthority",
"export DISPLAY",
"export XAUTHORITY"
].join("; ")
command = "#{x_env_variables}; xrandr -d $DISPLAY #{options}"
exec(command)
end
end
end
end
| Fix bug with LinuxComputer controller xrandr command builder | Fix bug with LinuxComputer controller xrandr command builder
| Ruby | mit | nickewing/hacky_hal | ruby | ## Code Before:
require_relative "generic_ssh"
module HackyHAL
module DeviceControllers
class LinuxComputer < GenericSsh
def mirror_screens(source_screen, dest_screen)
xrandr_command("--output #{dest_screen} --same-as #{source_screen}")
end
def set_screen_position(screen_1, screen_2, position)
xrandr_command("--output #{screen_1} --#{position}-of #{screen_2}")
end
def reset_display_settings(screen)
xrandr_command("--output #{screen} --auto")
end
private
def xrandr_command(options)
x_env_variables = [
"export CONSOLE=`sudo fgconsole`",
"export SESSION=`who -s | grep tty$CONSOLE | tr '()' ' '`",
"export DISPLAY=`echo $SESSION | awk '{print $5}'`",
"export XUSER=`echo $SESSION | awk '{print $1}'`",
"export XAUTHORITY=/home/$XUSER/.Xauthority"
].join("; ")
command = "#{x_env_variables}; xrandr -d $DISPLAY #{options}"
exec(command)
end
end
end
end
## Instruction:
Fix bug with LinuxComputer controller xrandr command builder
## Code After:
require_relative "generic_ssh"
module HackyHAL
module DeviceControllers
class LinuxComputer < GenericSsh
def mirror_screens(source_screen, dest_screen)
xrandr_command("--output #{dest_screen} --same-as #{source_screen}")
end
def set_screen_position(screen_1, screen_2, position)
xrandr_command("--output #{screen_1} --#{position}-of #{screen_2}")
end
def reset_display_settings(screen)
xrandr_command("--output #{screen} --auto")
end
private
def xrandr_command(options)
x_env_variables = [
"CONSOLE=$(sudo fgconsole)",
"SESSION=$(who -s | grep tty$CONSOLE | tr '()' ' ')",
"XUSER=$(echo $SESSION | awk '{print $1}')",
"DISPLAY=$(echo $SESSION | awk '{print $5}')",
"XAUTHORITY=/home/$XUSER/.Xauthority",
"export DISPLAY",
"export XAUTHORITY"
].join("; ")
command = "#{x_env_variables}; xrandr -d $DISPLAY #{options}"
exec(command)
end
end
end
end
| require_relative "generic_ssh"
module HackyHAL
module DeviceControllers
class LinuxComputer < GenericSsh
def mirror_screens(source_screen, dest_screen)
xrandr_command("--output #{dest_screen} --same-as #{source_screen}")
end
def set_screen_position(screen_1, screen_2, position)
xrandr_command("--output #{screen_1} --#{position}-of #{screen_2}")
end
def reset_display_settings(screen)
xrandr_command("--output #{screen} --auto")
end
private
def xrandr_command(options)
x_env_variables = [
- "export CONSOLE=`sudo fgconsole`",
? ------- ^ ^
+ "CONSOLE=$(sudo fgconsole)",
? ^^ ^
- "export SESSION=`who -s | grep tty$CONSOLE | tr '()' ' '`",
? ------- ^ ^
+ "SESSION=$(who -s | grep tty$CONSOLE | tr '()' ' ')",
? ^^ ^
+ "XUSER=$(echo $SESSION | awk '{print $1}')",
- "export DISPLAY=`echo $SESSION | awk '{print $5}'`",
? ------- ^ ^
+ "DISPLAY=$(echo $SESSION | awk '{print $5}')",
? ^^ ^
- "export XUSER=`echo $SESSION | awk '{print $1}'`",
- "export XAUTHORITY=/home/$XUSER/.Xauthority"
? -------
+ "XAUTHORITY=/home/$XUSER/.Xauthority",
? +
+ "export DISPLAY",
+ "export XAUTHORITY"
].join("; ")
command = "#{x_env_variables}; xrandr -d $DISPLAY #{options}"
exec(command)
end
end
end
end | 12 | 0.352941 | 7 | 5 |
33a2946e04dca399df1884f97b6d18616edeb0f1 | app/views/comments/index.haml | app/views/comments/index.haml | !!! XML
= xml_stylesheet 'main'
= xml_stylesheet 'coderay'
%feed{:xmlns => "http://www.w3.org/2005/Atom"}
%title
- if @post
== #{h @post.title} :
Comments : Nex3
%id= objects_url
- unless @comments.empty? && @post.nil?
%updated= (@comments[-1] || @post).created_at.xmlschema
%icon= image_path('favicon.png')
%link{:rel => 'self', :href => objects_path + '.atom'}/
- for comment in current_objects
%entry
%id= comment.uid
%title= h truncate(comment.content.gsub(/\s+/, ' '), 50)
%updated= comment.updated_at.xmlschema
%published= comment.created_at.xmlschema
%author
%name= h comment.user.name
%uri= h comment.user.href
%content{:type => 'xhtml'}
.model{:xmlns => "http://www.w3.org/1999/xhtml"}[comment]
= stylesheet_link_tag 'main'
= stylesheet_link_tag 'coderay'
.content= find_and_preserve(comment.render)
%link{:rel => 'alternate', :href => object_path(comment)}/
| !!! XML
= xml_stylesheet 'main'
= xml_stylesheet 'coderay'
%feed{:xmlns => "http://www.w3.org/2005/Atom"}
%title
- if @post
== #{h @post.title} :
Comments : Nex3
%id= objects_url
- unless @comments.empty? && @post.nil?
%updated= (@comments[-1] || @post).created_at.xmlschema
%icon= image_path('favicon.png')
%link{:rel => 'self', :href => objects_path + '.atom'}/
- for comment in current_objects
%entry
%id= comment.uid
%title= h truncate(comment.content.gsub(/\s+/, ' '), 50)
%updated= comment.updated_at.xmlschema
%published= comment.created_at.xmlschema
%author
%name= h comment.user.name
%uri= h comment.user.href
%content{:type => 'xhtml'}
.model{:xmlns => "http://www.w3.org/1999/xhtml"}[comment]
= stylesheet_link_tag 'main'
= stylesheet_link_tag 'coderay'
.content= find_and_preserve(comment.render)
%link{:rel => 'alternate', :href => comment_path(comment.post_id, comment)}/
| Fix links on /comments.atom entries. | Fix links on /comments.atom entries.
| Haml | mit | nex3/nex3-s-blog-engine,nex3/nex3-s-blog-engine | haml | ## Code Before:
!!! XML
= xml_stylesheet 'main'
= xml_stylesheet 'coderay'
%feed{:xmlns => "http://www.w3.org/2005/Atom"}
%title
- if @post
== #{h @post.title} :
Comments : Nex3
%id= objects_url
- unless @comments.empty? && @post.nil?
%updated= (@comments[-1] || @post).created_at.xmlschema
%icon= image_path('favicon.png')
%link{:rel => 'self', :href => objects_path + '.atom'}/
- for comment in current_objects
%entry
%id= comment.uid
%title= h truncate(comment.content.gsub(/\s+/, ' '), 50)
%updated= comment.updated_at.xmlschema
%published= comment.created_at.xmlschema
%author
%name= h comment.user.name
%uri= h comment.user.href
%content{:type => 'xhtml'}
.model{:xmlns => "http://www.w3.org/1999/xhtml"}[comment]
= stylesheet_link_tag 'main'
= stylesheet_link_tag 'coderay'
.content= find_and_preserve(comment.render)
%link{:rel => 'alternate', :href => object_path(comment)}/
## Instruction:
Fix links on /comments.atom entries.
## Code After:
!!! XML
= xml_stylesheet 'main'
= xml_stylesheet 'coderay'
%feed{:xmlns => "http://www.w3.org/2005/Atom"}
%title
- if @post
== #{h @post.title} :
Comments : Nex3
%id= objects_url
- unless @comments.empty? && @post.nil?
%updated= (@comments[-1] || @post).created_at.xmlschema
%icon= image_path('favicon.png')
%link{:rel => 'self', :href => objects_path + '.atom'}/
- for comment in current_objects
%entry
%id= comment.uid
%title= h truncate(comment.content.gsub(/\s+/, ' '), 50)
%updated= comment.updated_at.xmlschema
%published= comment.created_at.xmlschema
%author
%name= h comment.user.name
%uri= h comment.user.href
%content{:type => 'xhtml'}
.model{:xmlns => "http://www.w3.org/1999/xhtml"}[comment]
= stylesheet_link_tag 'main'
= stylesheet_link_tag 'coderay'
.content= find_and_preserve(comment.render)
%link{:rel => 'alternate', :href => comment_path(comment.post_id, comment)}/
| !!! XML
= xml_stylesheet 'main'
= xml_stylesheet 'coderay'
%feed{:xmlns => "http://www.w3.org/2005/Atom"}
%title
- if @post
== #{h @post.title} :
Comments : Nex3
%id= objects_url
- unless @comments.empty? && @post.nil?
%updated= (@comments[-1] || @post).created_at.xmlschema
%icon= image_path('favicon.png')
%link{:rel => 'self', :href => objects_path + '.atom'}/
- for comment in current_objects
%entry
%id= comment.uid
%title= h truncate(comment.content.gsub(/\s+/, ' '), 50)
%updated= comment.updated_at.xmlschema
%published= comment.created_at.xmlschema
%author
%name= h comment.user.name
%uri= h comment.user.href
%content{:type => 'xhtml'}
.model{:xmlns => "http://www.w3.org/1999/xhtml"}[comment]
= stylesheet_link_tag 'main'
= stylesheet_link_tag 'coderay'
.content= find_and_preserve(comment.render)
- %link{:rel => 'alternate', :href => object_path(comment)}/
? ^^ ^
+ %link{:rel => 'alternate', :href => comment_path(comment.post_id, comment)}/
? + ^^ ^ +++++++++++++++++
| 2 | 0.071429 | 1 | 1 |
18957b36ddcd96b41c8897542609c53c166052c3 | config/initializers/middleware.js | config/initializers/middleware.js | express = require('express')
expressValidator = require('express-validator')
module.exports = (function(){
function configure(app) {
app.set('port', process.env.PORT || 3000);
app.set('host', process.env.HOST || '127.0.0.1')
app.use(express.static(__dirname + '/public'));
app.use(express.bodyParser())
app.use(expressValidator());
app.use(express.cookieParser());
app.use(express.session({secret: 'oi09ajsdf09fwlkej33lkjpx'}));
app.use(express.methodOverride())
app.use(function(err, req, res, next) {
res.send({ error: err });
});
}
return { configure: configure }
})()
| express = require('express')
expressValidator = require('express-validator')
module.exports = (function(){
function configure(app) {
app.set('port', process.env.PORT || 3000);
app.set('host', process.env.HOST || '127.0.0.1')
app.use(express.static(__dirname + '/public'));
app.use(function(req,res,next) {
res.header("Access-Control-Allow-Origin", "*");
res.header("Access-Control-Allow-Headers", "X-Requested-With");
})
app.use(express.bodyParser())
app.use(expressValidator());
app.use(express.cookieParser());
app.use(express.session({secret: 'oi09ajsdf09fwlkej33lkjpx'}));
app.use(express.methodOverride())
app.use(function(err, req, res, next) {
res.send({ error: err });
});
}
return { configure: configure }
})()
| Add access control allow origin middlewar | [FEATURE] Add access control allow origin middlewar
| JavaScript | isc | zealord/gatewayd,crazyquark/gatewayd,xdv/gatewayd,xdv/gatewayd,Parkjeahwan/awegeeks,whotooktwarden/gatewayd,Parkjeahwan/awegeeks,crazyquark/gatewayd,whotooktwarden/gatewayd,zealord/gatewayd | javascript | ## Code Before:
express = require('express')
expressValidator = require('express-validator')
module.exports = (function(){
function configure(app) {
app.set('port', process.env.PORT || 3000);
app.set('host', process.env.HOST || '127.0.0.1')
app.use(express.static(__dirname + '/public'));
app.use(express.bodyParser())
app.use(expressValidator());
app.use(express.cookieParser());
app.use(express.session({secret: 'oi09ajsdf09fwlkej33lkjpx'}));
app.use(express.methodOverride())
app.use(function(err, req, res, next) {
res.send({ error: err });
});
}
return { configure: configure }
})()
## Instruction:
[FEATURE] Add access control allow origin middlewar
## Code After:
express = require('express')
expressValidator = require('express-validator')
module.exports = (function(){
function configure(app) {
app.set('port', process.env.PORT || 3000);
app.set('host', process.env.HOST || '127.0.0.1')
app.use(express.static(__dirname + '/public'));
app.use(function(req,res,next) {
res.header("Access-Control-Allow-Origin", "*");
res.header("Access-Control-Allow-Headers", "X-Requested-With");
})
app.use(express.bodyParser())
app.use(expressValidator());
app.use(express.cookieParser());
app.use(express.session({secret: 'oi09ajsdf09fwlkej33lkjpx'}));
app.use(express.methodOverride())
app.use(function(err, req, res, next) {
res.send({ error: err });
});
}
return { configure: configure }
})()
| express = require('express')
expressValidator = require('express-validator')
module.exports = (function(){
function configure(app) {
app.set('port', process.env.PORT || 3000);
app.set('host', process.env.HOST || '127.0.0.1')
app.use(express.static(__dirname + '/public'));
+ app.use(function(req,res,next) {
+ res.header("Access-Control-Allow-Origin", "*");
+ res.header("Access-Control-Allow-Headers", "X-Requested-With");
+ })
app.use(express.bodyParser())
app.use(expressValidator());
app.use(express.cookieParser());
app.use(express.session({secret: 'oi09ajsdf09fwlkej33lkjpx'}));
app.use(express.methodOverride())
app.use(function(err, req, res, next) {
res.send({ error: err });
});
}
return { configure: configure }
})() | 4 | 0.210526 | 4 | 0 |
dfe76dd56642cbb6642a9a3776e407647f6ca1ba | router.js | router.js | var express = require('express');
var router = module.exports = express.Router();
var path = require('path');
var appDir = path.join(__dirname, 'app');
router.get('/', function(req, res) {
res.render('index');
});
| var express = require('express');
var router = module.exports = express.Router();
var path = require('path');
var appDir = path.join(__dirname, 'app');
router.get('/', function(req, res) {
res.render('index');
});
router.get('/:page', function(req, res) {
res.render(req.params.page);
});
router.get('/pages/:page', function(req, res) {
res.render('pages/' + req.params.page);
});
| Add routing for /:page and /pages/:page | Add routing for /:page and /pages/:page
Render them from their corresponding view directories
| JavaScript | mit | szekelyszilv/hackkings2015,SomeRandomTeam/hackkings2015,SomeRandomTeam/hackkings2015,szekelyszilv/hackkings2015,szekelyszilv/hackkings2015,SomeRandomTeam/hackkings2015 | javascript | ## Code Before:
var express = require('express');
var router = module.exports = express.Router();
var path = require('path');
var appDir = path.join(__dirname, 'app');
router.get('/', function(req, res) {
res.render('index');
});
## Instruction:
Add routing for /:page and /pages/:page
Render them from their corresponding view directories
## Code After:
var express = require('express');
var router = module.exports = express.Router();
var path = require('path');
var appDir = path.join(__dirname, 'app');
router.get('/', function(req, res) {
res.render('index');
});
router.get('/:page', function(req, res) {
res.render(req.params.page);
});
router.get('/pages/:page', function(req, res) {
res.render('pages/' + req.params.page);
});
| var express = require('express');
var router = module.exports = express.Router();
var path = require('path');
var appDir = path.join(__dirname, 'app');
router.get('/', function(req, res) {
res.render('index');
});
+ router.get('/:page', function(req, res) {
+ res.render(req.params.page);
+ });
+
+ router.get('/pages/:page', function(req, res) {
+ res.render('pages/' + req.params.page);
+ });
+ | 8 | 0.8 | 8 | 0 |
8770666a2e0420da82573791c69d06d2baa8c99d | FakeSearchView/gradle.properties | FakeSearchView/gradle.properties | VERSION_NAME = 0.4-SNAPSHOT
GROUP = com.github.leonardoxh
POM_DESCRIPTION = A custom SearchView for android
POM_URL = https://github.com/leonardoxh/FakeSearchView
POM_SCM_URL = https://github.com/leonardoxh/FakeSearchView
POM_SCM_CONNECTION = scm:git@github.com:leonardoxh/FakeSearchView.git
POM_SCM_DEV_CONNECTION = scm:git@github.com:leonardoxh/FakeSearchView.git
POM_LICENCE_NAME = The Apache Software License, Version 2.0
POM_LICENCE_URL = http://www.apache.org/licenses/LICENSE-2.0.txt
POM_LICENCE_DIST = repo
| VERSION_NAME = 0.3.1
GROUP = com.github.leonardoxh
POM_DESCRIPTION = A custom SearchView for android
POM_URL = https://github.com/leonardoxh/FakeSearchView
POM_SCM_URL = https://github.com/leonardoxh/FakeSearchView
POM_SCM_CONNECTION = scm:git@github.com:leonardoxh/FakeSearchView.git
POM_SCM_DEV_CONNECTION = scm:git@github.com:leonardoxh/FakeSearchView.git
POM_LICENCE_NAME = The Apache Software License, Version 2.0
POM_LICENCE_URL = http://www.apache.org/licenses/LICENSE-2.0.txt
POM_LICENCE_DIST = repo
| Prepare for next maven fix release | Prepare for next maven fix release
| INI | apache-2.0 | luinnx/FakeSearchView,leonardoxh/FakeSearchView | ini | ## Code Before:
VERSION_NAME = 0.4-SNAPSHOT
GROUP = com.github.leonardoxh
POM_DESCRIPTION = A custom SearchView for android
POM_URL = https://github.com/leonardoxh/FakeSearchView
POM_SCM_URL = https://github.com/leonardoxh/FakeSearchView
POM_SCM_CONNECTION = scm:git@github.com:leonardoxh/FakeSearchView.git
POM_SCM_DEV_CONNECTION = scm:git@github.com:leonardoxh/FakeSearchView.git
POM_LICENCE_NAME = The Apache Software License, Version 2.0
POM_LICENCE_URL = http://www.apache.org/licenses/LICENSE-2.0.txt
POM_LICENCE_DIST = repo
## Instruction:
Prepare for next maven fix release
## Code After:
VERSION_NAME = 0.3.1
GROUP = com.github.leonardoxh
POM_DESCRIPTION = A custom SearchView for android
POM_URL = https://github.com/leonardoxh/FakeSearchView
POM_SCM_URL = https://github.com/leonardoxh/FakeSearchView
POM_SCM_CONNECTION = scm:git@github.com:leonardoxh/FakeSearchView.git
POM_SCM_DEV_CONNECTION = scm:git@github.com:leonardoxh/FakeSearchView.git
POM_LICENCE_NAME = The Apache Software License, Version 2.0
POM_LICENCE_URL = http://www.apache.org/licenses/LICENSE-2.0.txt
POM_LICENCE_DIST = repo
| - VERSION_NAME = 0.4-SNAPSHOT
+ VERSION_NAME = 0.3.1
GROUP = com.github.leonardoxh
POM_DESCRIPTION = A custom SearchView for android
POM_URL = https://github.com/leonardoxh/FakeSearchView
POM_SCM_URL = https://github.com/leonardoxh/FakeSearchView
POM_SCM_CONNECTION = scm:git@github.com:leonardoxh/FakeSearchView.git
POM_SCM_DEV_CONNECTION = scm:git@github.com:leonardoxh/FakeSearchView.git
POM_LICENCE_NAME = The Apache Software License, Version 2.0
POM_LICENCE_URL = http://www.apache.org/licenses/LICENSE-2.0.txt
POM_LICENCE_DIST = repo | 2 | 0.181818 | 1 | 1 |
221d445c0d5cc06c9b994b3ceca61afb197a5f87 | libraries/lets-plot.json | libraries/lets-plot.json | {
"properties": {
"libraryVersion": "1.4.2",
"apiVersion": "0.0.18-SNAPSHOT",
"isolatedFrame": ""
},
"link": "https://github.com/JetBrains/lets-plot-kotlin",
"repositories": [
"https://jetbrains.bintray.com/lets-plot-maven"
],
"dependencies": [
"org.jetbrains.lets-plot:lets-plot-kotlin-api:$apiVersion"
],
"imports": [
"jetbrains.letsPlot.*",
"jetbrains.letsPlot.geom.*",
"jetbrains.letsPlot.stat.*"
],
"init": [
"import jetbrains.letsPlot.LetsPlot",
"import jetbrains.letsPlot.frontend.NotebookFrontendContext",
"val isolatedFrameParam = if(\"$isolatedFrame\".isNotEmpty()) \"$isolatedFrame\".toBoolean() else null",
"val frontendContext = LetsPlot.setupNotebook(\"$libraryVersion\", isolatedFrameParam) {DISPLAY(HTML(it))}",
"LetsPlot.apiVersion = \"$apiVersion\"",
"// Load library JS",
"DISPLAY(HTML(frontendContext.getConfigureHtml()))"
],
"renderers": {
"jetbrains.letsPlot.intern.Plot": "HTML(frontendContext.getHtml($it))",
"jetbrains.letsPlot.GGBunch": "HTML(frontendContext.getHtml($it))"
}
}
| {
"properties": {
"api": "0.0.22-SNAPSHOT",
"js": "1.4.2",
"isolatedFrame": ""
},
"link": "https://github.com/JetBrains/lets-plot-kotlin",
"repositories": [
"https://jetbrains.bintray.com/lets-plot-maven"
],
"dependencies": [
"org.jetbrains.lets-plot:lets-plot-kotlin-api:$api"
],
"imports": [
"jetbrains.letsPlot.*",
"jetbrains.letsPlot.geom.*",
"jetbrains.letsPlot.stat.*",
"jetbrains.letsPlot.label.*",
"jetbrains.letsPlot.scale.*",
"jetbrains.letsPlot.facet.*"
],
"init": [
"import jetbrains.letsPlot.LetsPlot",
"import jetbrains.letsPlot.frontend.NotebookFrontendContext",
"val isolatedFrameParam = if(\"$isolatedFrame\".isNotEmpty()) \"$isolatedFrame\".toBoolean() else null",
"val frontendContext = LetsPlot.setupNotebook(\"$js\", isolatedFrameParam) {DISPLAY(HTML(it))}",
"LetsPlot.apiVersion = \"$api\"",
"// Load library JS",
"DISPLAY(HTML(frontendContext.getConfigureHtml()))"
],
"renderers": {
"jetbrains.letsPlot.intern.Plot": "HTML(frontendContext.getHtml($it))",
"jetbrains.letsPlot.GGBunch": "HTML(frontendContext.getHtml($it))"
}
}
| Upgrade Lets-Plot Kotlin API to v. 0.0.22-SNAPSHOT | Upgrade Lets-Plot Kotlin API to v. 0.0.22-SNAPSHOT
| JSON | apache-2.0 | ligee/kotlin-jupyter | json | ## Code Before:
{
"properties": {
"libraryVersion": "1.4.2",
"apiVersion": "0.0.18-SNAPSHOT",
"isolatedFrame": ""
},
"link": "https://github.com/JetBrains/lets-plot-kotlin",
"repositories": [
"https://jetbrains.bintray.com/lets-plot-maven"
],
"dependencies": [
"org.jetbrains.lets-plot:lets-plot-kotlin-api:$apiVersion"
],
"imports": [
"jetbrains.letsPlot.*",
"jetbrains.letsPlot.geom.*",
"jetbrains.letsPlot.stat.*"
],
"init": [
"import jetbrains.letsPlot.LetsPlot",
"import jetbrains.letsPlot.frontend.NotebookFrontendContext",
"val isolatedFrameParam = if(\"$isolatedFrame\".isNotEmpty()) \"$isolatedFrame\".toBoolean() else null",
"val frontendContext = LetsPlot.setupNotebook(\"$libraryVersion\", isolatedFrameParam) {DISPLAY(HTML(it))}",
"LetsPlot.apiVersion = \"$apiVersion\"",
"// Load library JS",
"DISPLAY(HTML(frontendContext.getConfigureHtml()))"
],
"renderers": {
"jetbrains.letsPlot.intern.Plot": "HTML(frontendContext.getHtml($it))",
"jetbrains.letsPlot.GGBunch": "HTML(frontendContext.getHtml($it))"
}
}
## Instruction:
Upgrade Lets-Plot Kotlin API to v. 0.0.22-SNAPSHOT
## Code After:
{
"properties": {
"api": "0.0.22-SNAPSHOT",
"js": "1.4.2",
"isolatedFrame": ""
},
"link": "https://github.com/JetBrains/lets-plot-kotlin",
"repositories": [
"https://jetbrains.bintray.com/lets-plot-maven"
],
"dependencies": [
"org.jetbrains.lets-plot:lets-plot-kotlin-api:$api"
],
"imports": [
"jetbrains.letsPlot.*",
"jetbrains.letsPlot.geom.*",
"jetbrains.letsPlot.stat.*",
"jetbrains.letsPlot.label.*",
"jetbrains.letsPlot.scale.*",
"jetbrains.letsPlot.facet.*"
],
"init": [
"import jetbrains.letsPlot.LetsPlot",
"import jetbrains.letsPlot.frontend.NotebookFrontendContext",
"val isolatedFrameParam = if(\"$isolatedFrame\".isNotEmpty()) \"$isolatedFrame\".toBoolean() else null",
"val frontendContext = LetsPlot.setupNotebook(\"$js\", isolatedFrameParam) {DISPLAY(HTML(it))}",
"LetsPlot.apiVersion = \"$api\"",
"// Load library JS",
"DISPLAY(HTML(frontendContext.getConfigureHtml()))"
],
"renderers": {
"jetbrains.letsPlot.intern.Plot": "HTML(frontendContext.getHtml($it))",
"jetbrains.letsPlot.GGBunch": "HTML(frontendContext.getHtml($it))"
}
}
| {
"properties": {
- "libraryVersion": "1.4.2",
- "apiVersion": "0.0.18-SNAPSHOT",
? ------- ^^
+ "api": "0.0.22-SNAPSHOT",
? ^^
+ "js": "1.4.2",
"isolatedFrame": ""
},
"link": "https://github.com/JetBrains/lets-plot-kotlin",
"repositories": [
"https://jetbrains.bintray.com/lets-plot-maven"
],
"dependencies": [
- "org.jetbrains.lets-plot:lets-plot-kotlin-api:$apiVersion"
? -------
+ "org.jetbrains.lets-plot:lets-plot-kotlin-api:$api"
],
"imports": [
"jetbrains.letsPlot.*",
"jetbrains.letsPlot.geom.*",
- "jetbrains.letsPlot.stat.*"
+ "jetbrains.letsPlot.stat.*",
? +
+ "jetbrains.letsPlot.label.*",
+ "jetbrains.letsPlot.scale.*",
+ "jetbrains.letsPlot.facet.*"
],
"init": [
"import jetbrains.letsPlot.LetsPlot",
"import jetbrains.letsPlot.frontend.NotebookFrontendContext",
"val isolatedFrameParam = if(\"$isolatedFrame\".isNotEmpty()) \"$isolatedFrame\".toBoolean() else null",
- "val frontendContext = LetsPlot.setupNotebook(\"$libraryVersion\", isolatedFrameParam) {DISPLAY(HTML(it))}",
? ^^^^^^^^^^ ---
+ "val frontendContext = LetsPlot.setupNotebook(\"$js\", isolatedFrameParam) {DISPLAY(HTML(it))}",
? ^
- "LetsPlot.apiVersion = \"$apiVersion\"",
? -------
+ "LetsPlot.apiVersion = \"$api\"",
"// Load library JS",
"DISPLAY(HTML(frontendContext.getConfigureHtml()))"
],
"renderers": {
"jetbrains.letsPlot.intern.Plot": "HTML(frontendContext.getHtml($it))",
"jetbrains.letsPlot.GGBunch": "HTML(frontendContext.getHtml($it))"
}
} | 15 | 0.46875 | 9 | 6 |
78cb2028373db7a650c0c2081f1cf4232d52d3d6 | CHANGELOG.md | CHANGELOG.md |
Initial versions with various patches for:
- Weakly retain expectation (fix tests)
- Token and basic authorisation
- Fix for response retaining `env` before invoking call-backs
- Using the given dispatch queue (fix)
- Response success or failure using `onSuccess` or `onFailure`
- URL session handler sets up configuration and session
- Make `buildRequest` method public
- Invoke `runRequest(request)` from `runRequest(method, path, requestBuilder)`
- Make `buildResponse` behaviour available for `Connection`s
- Variant of `Response`'s `onComplete` method: with dispatch queue
- Ping tests use optional `path` argument
- Request runners accept optional `path` argument
- Travis CI support
- Jazzy set-up
See [Full Change Log](https://github.com/royratcliffe/faraday/compare/0.1.0...0.1.11).
|
Initial versions with various patches for:
- Weakly retain expectation (fix tests)
- Token and basic authorisation
- Fix for response retaining `env` before invoking call-backs
- Using the given dispatch queue (fix)
- Response success or failure using `onSuccess` or `onFailure`
- URL session handler sets up configuration and session
- Make `buildRequest` method public
- Invoke `runRequest(request)` from `runRequest(method, path, requestBuilder)`
- Make `buildResponse` behaviour available for `Connection`s
- Variant of `Response`'s `onComplete` method: with dispatch queue
- Ping tests use optional `path` argument
- Request runners accept optional `path` argument
- Travis CI support; fix tests for Travis
- Jazzy set-up
See [Full Change Log](https://github.com/royratcliffe/faraday/compare/0.1.0...0.1.11).
| Change log: fix tests for Travis | Change log: fix tests for Travis
| Markdown | mit | royratcliffe/Faraday,royratcliffe/Faraday | markdown | ## Code Before:
Initial versions with various patches for:
- Weakly retain expectation (fix tests)
- Token and basic authorisation
- Fix for response retaining `env` before invoking call-backs
- Using the given dispatch queue (fix)
- Response success or failure using `onSuccess` or `onFailure`
- URL session handler sets up configuration and session
- Make `buildRequest` method public
- Invoke `runRequest(request)` from `runRequest(method, path, requestBuilder)`
- Make `buildResponse` behaviour available for `Connection`s
- Variant of `Response`'s `onComplete` method: with dispatch queue
- Ping tests use optional `path` argument
- Request runners accept optional `path` argument
- Travis CI support
- Jazzy set-up
See [Full Change Log](https://github.com/royratcliffe/faraday/compare/0.1.0...0.1.11).
## Instruction:
Change log: fix tests for Travis
## Code After:
Initial versions with various patches for:
- Weakly retain expectation (fix tests)
- Token and basic authorisation
- Fix for response retaining `env` before invoking call-backs
- Using the given dispatch queue (fix)
- Response success or failure using `onSuccess` or `onFailure`
- URL session handler sets up configuration and session
- Make `buildRequest` method public
- Invoke `runRequest(request)` from `runRequest(method, path, requestBuilder)`
- Make `buildResponse` behaviour available for `Connection`s
- Variant of `Response`'s `onComplete` method: with dispatch queue
- Ping tests use optional `path` argument
- Request runners accept optional `path` argument
- Travis CI support; fix tests for Travis
- Jazzy set-up
See [Full Change Log](https://github.com/royratcliffe/faraday/compare/0.1.0...0.1.11).
|
Initial versions with various patches for:
- Weakly retain expectation (fix tests)
- Token and basic authorisation
- Fix for response retaining `env` before invoking call-backs
- Using the given dispatch queue (fix)
- Response success or failure using `onSuccess` or `onFailure`
- URL session handler sets up configuration and session
- Make `buildRequest` method public
- Invoke `runRequest(request)` from `runRequest(method, path, requestBuilder)`
- Make `buildResponse` behaviour available for `Connection`s
- Variant of `Response`'s `onComplete` method: with dispatch queue
- Ping tests use optional `path` argument
- Request runners accept optional `path` argument
- - Travis CI support
+ - Travis CI support; fix tests for Travis
- Jazzy set-up
See [Full Change Log](https://github.com/royratcliffe/faraday/compare/0.1.0...0.1.11). | 2 | 0.105263 | 1 | 1 |
9a42fe38b595c7d7a45f1496f92207923b846b2a | _includes/footer.html | _includes/footer.html | <div class="footer">
<!-- Copyright years -->
{% assign first = site.posts.last.date | date: '%Y' %}
{% assign current = site.time | date: '%Y' %}
{% if first == current %}
{% assign copyright = current %}
{% else %}
{% assign copyright = first | append: '–' | append: current %}</p>
{% endif %}
<hr>
<span class="copy-left">©</span> {{ copyright }} {{ site.author.name }}. No rights reserved.<br>
<small>Steal these posts. Share the knowledge. Power to the people.</small>
</div>
| <div class="footer">
<!-- Copyright years -->
{% assign first = site.posts.last.date | date: '%Y' %}
{% assign current = site.time | date: '%Y' %}
{% if first == current or first == nil %}
{% assign copyright = current %}
{% else %}
{% assign copyright = first | append: '–' | append: current %}</p>
{% endif %}
<hr>
<span class="copy-left">©</span> {{ copyright }} {{ site.author.name }}. No rights reserved.<br>
<small>Steal these posts. Share the knowledge. Power to the people.</small>
</div>
| Check if first post date is nil | Check if first post date is nil
| HTML | mit | epistrephein/blog.epistrephe.in,epistrephein/blog.epistrephe.in,epistrephein/epistrephe.in,epistrephein/epistrephe.in | html | ## Code Before:
<div class="footer">
<!-- Copyright years -->
{% assign first = site.posts.last.date | date: '%Y' %}
{% assign current = site.time | date: '%Y' %}
{% if first == current %}
{% assign copyright = current %}
{% else %}
{% assign copyright = first | append: '–' | append: current %}</p>
{% endif %}
<hr>
<span class="copy-left">©</span> {{ copyright }} {{ site.author.name }}. No rights reserved.<br>
<small>Steal these posts. Share the knowledge. Power to the people.</small>
</div>
## Instruction:
Check if first post date is nil
## Code After:
<div class="footer">
<!-- Copyright years -->
{% assign first = site.posts.last.date | date: '%Y' %}
{% assign current = site.time | date: '%Y' %}
{% if first == current or first == nil %}
{% assign copyright = current %}
{% else %}
{% assign copyright = first | append: '–' | append: current %}</p>
{% endif %}
<hr>
<span class="copy-left">©</span> {{ copyright }} {{ site.author.name }}. No rights reserved.<br>
<small>Steal these posts. Share the knowledge. Power to the people.</small>
</div>
| <div class="footer">
<!-- Copyright years -->
{% assign first = site.posts.last.date | date: '%Y' %}
{% assign current = site.time | date: '%Y' %}
- {% if first == current %}
+ {% if first == current or first == nil %}
? ++++++++++++++++
{% assign copyright = current %}
{% else %}
{% assign copyright = first | append: '–' | append: current %}</p>
{% endif %}
<hr>
<span class="copy-left">©</span> {{ copyright }} {{ site.author.name }}. No rights reserved.<br>
<small>Steal these posts. Share the knowledge. Power to the people.</small>
</div> | 2 | 0.133333 | 1 | 1 |
4598143f0c14a198379b25494dd95a1252c30a1a | templates/theme-footer.html | templates/theme-footer.html | <%! from django.core.urlresolvers import reverse %>
<%namespace name='static' file='/static_content.html'/>
<%! from datetime import datetime %>
<%! from pytz import UTC %>
<div class="wrapper-footer">
<footer>
<div class="colophon">
<nav class="nav-colophon">
<ol>
<li><a href="${reverse('tos')}">Terms of Service</a></li>
<li><a href="${reverse('tos')}#privacy">Privacy Policy</a></li>
<li><a href="${reverse('tos')}#honor">Honor Code</a></li>
<li><a href="${reverse('tos')}#copyright">Copyright</a></li>
<li><a href="${reverse('about_edx')}#careers">Careers</a></li>
<li><a href="${reverse('about_edx')}#contact">Contact</a></li>
<li><a href="${marketing_link('FAQ')}">Help</a></li>
</ol>
</nav>
</div>
<div class="references">
<span>Built on <a href="http://code.edx.org">OpenEdX</a>.</span>
</div>
</footer>
</div>
| <%! from django.core.urlresolvers import reverse %>
<%namespace name='static' file='/static_content.html'/>
<%! from datetime import datetime %>
<%! from pytz import UTC %>
<div class="wrapper-footer">
<footer>
<div class="colophon">
<nav class="nav-colophon">
<ol>
<li><a href="${reverse('tos')}">Terms of Service</a></li>
<li><a href="${reverse('tos')}#privacy">Privacy Policy</a></li>
<li><a href="${reverse('tos')}#honor">Honor Code</a></li>
<li><a href="${reverse('tos')}#copyright">Copyright</a></li>
<li><a href="http://www.labster.com/jobs/" target="_blank">Careers</a></li>
<li><a href="http://www.labster.com/about-contact/" target="_blank">Contact</a></li>
<li><a href="${marketing_link('FAQ')}">Help</a></li>
</ol>
</nav>
</div>
<div class="references">
<span>Built on <a href="http://code.edx.org">OpenEdX</a>.</span>
</div>
</footer>
</div>
| Change link of Contact and Career | Change link of Contact and Career
| HTML | apache-2.0 | Livit/Livit.Learn.EdX.Theme | html | ## Code Before:
<%! from django.core.urlresolvers import reverse %>
<%namespace name='static' file='/static_content.html'/>
<%! from datetime import datetime %>
<%! from pytz import UTC %>
<div class="wrapper-footer">
<footer>
<div class="colophon">
<nav class="nav-colophon">
<ol>
<li><a href="${reverse('tos')}">Terms of Service</a></li>
<li><a href="${reverse('tos')}#privacy">Privacy Policy</a></li>
<li><a href="${reverse('tos')}#honor">Honor Code</a></li>
<li><a href="${reverse('tos')}#copyright">Copyright</a></li>
<li><a href="${reverse('about_edx')}#careers">Careers</a></li>
<li><a href="${reverse('about_edx')}#contact">Contact</a></li>
<li><a href="${marketing_link('FAQ')}">Help</a></li>
</ol>
</nav>
</div>
<div class="references">
<span>Built on <a href="http://code.edx.org">OpenEdX</a>.</span>
</div>
</footer>
</div>
## Instruction:
Change link of Contact and Career
## Code After:
<%! from django.core.urlresolvers import reverse %>
<%namespace name='static' file='/static_content.html'/>
<%! from datetime import datetime %>
<%! from pytz import UTC %>
<div class="wrapper-footer">
<footer>
<div class="colophon">
<nav class="nav-colophon">
<ol>
<li><a href="${reverse('tos')}">Terms of Service</a></li>
<li><a href="${reverse('tos')}#privacy">Privacy Policy</a></li>
<li><a href="${reverse('tos')}#honor">Honor Code</a></li>
<li><a href="${reverse('tos')}#copyright">Copyright</a></li>
<li><a href="http://www.labster.com/jobs/" target="_blank">Careers</a></li>
<li><a href="http://www.labster.com/about-contact/" target="_blank">Contact</a></li>
<li><a href="${marketing_link('FAQ')}">Help</a></li>
</ol>
</nav>
</div>
<div class="references">
<span>Built on <a href="http://code.edx.org">OpenEdX</a>.</span>
</div>
</footer>
</div>
| <%! from django.core.urlresolvers import reverse %>
<%namespace name='static' file='/static_content.html'/>
<%! from datetime import datetime %>
<%! from pytz import UTC %>
<div class="wrapper-footer">
<footer>
<div class="colophon">
<nav class="nav-colophon">
<ol>
<li><a href="${reverse('tos')}">Terms of Service</a></li>
<li><a href="${reverse('tos')}#privacy">Privacy Policy</a></li>
<li><a href="${reverse('tos')}#honor">Honor Code</a></li>
<li><a href="${reverse('tos')}#copyright">Copyright</a></li>
- <li><a href="${reverse('about_edx')}#careers">Careers</a></li>
- <li><a href="${reverse('about_edx')}#contact">Contact</a></li>
+ <li><a href="http://www.labster.com/jobs/" target="_blank">Careers</a></li>
+ <li><a href="http://www.labster.com/about-contact/" target="_blank">Contact</a></li>
<li><a href="${marketing_link('FAQ')}">Help</a></li>
</ol>
</nav>
</div>
<div class="references">
<span>Built on <a href="http://code.edx.org">OpenEdX</a>.</span>
</div>
</footer>
</div> | 4 | 0.16 | 2 | 2 |
8273875bf47a717affd8643bffdb0565b80b794c | system/env.zsh | system/env.zsh | export GITHUB_USER="phamann"
export PATH=/usr/local/bin:/usr/bin:/bin:/usr/sbin:/sbin:$PATH # Reorder PATH so local bin is first
export GREP_OPTIONS='--color=auto'
export GREP_COLOR='1;32'
export MANPAGER="less -X" # Don’t clear the screen after quitting a manual page
export EDITOR="vim"
export TERM="screen-256color"
export CLICOLOR=1
export JAVA_HOME=$(/usr/libexec/java_home)
export PATH="/usr/local/heroku/bin:$PATH"
export PATH=/opt/local/bin:/opt/local/sbin:$PATH
| export GITHUB_USER="phamann"
export PATH=/usr/local/bin:/usr/bin:/bin:/usr/sbin:/sbin:$PATH # Reorder PATH so local bin is first
export GREP_OPTIONS='--color=auto'
export GREP_COLOR='1;32'
export MANPAGER="less -X" # Don’t clear the screen after quitting a manual page
export EDITOR="vim"
export TERM="screen-256color"
export VIM_COLOUR_SCHEME="solarized"
export CLICOLOR=1
export JAVA_HOME=$(/usr/libexec/java_home)
export PATH="/usr/local/heroku/bin:$PATH"
export PATH=/opt/local/bin:/opt/local/sbin:$PATH
export PATH="$HOME/.node/bin:$PATH"
export PATH="node_modules/.bin:$PATH"
export PATH=$HOME/.rbenv/shims:$PATH
export PATH="$HOME/Code/neo4j-2.2.3/bin:$PATH"
export GOPATH=$HOME/Code/Go
export PATH=$PATH:$GOROOT/bin
export PATH="/usr/local/opt/gnu-tar/libexec/gnubin:$PATH"
| Add various things to the $PATH | Add various things to the $PATH
| Shell | mit | phamann/dotfiles,phamann/dotfiles,phamann/dotfiles | shell | ## Code Before:
export GITHUB_USER="phamann"
export PATH=/usr/local/bin:/usr/bin:/bin:/usr/sbin:/sbin:$PATH # Reorder PATH so local bin is first
export GREP_OPTIONS='--color=auto'
export GREP_COLOR='1;32'
export MANPAGER="less -X" # Don’t clear the screen after quitting a manual page
export EDITOR="vim"
export TERM="screen-256color"
export CLICOLOR=1
export JAVA_HOME=$(/usr/libexec/java_home)
export PATH="/usr/local/heroku/bin:$PATH"
export PATH=/opt/local/bin:/opt/local/sbin:$PATH
## Instruction:
Add various things to the $PATH
## Code After:
export GITHUB_USER="phamann"
export PATH=/usr/local/bin:/usr/bin:/bin:/usr/sbin:/sbin:$PATH # Reorder PATH so local bin is first
export GREP_OPTIONS='--color=auto'
export GREP_COLOR='1;32'
export MANPAGER="less -X" # Don’t clear the screen after quitting a manual page
export EDITOR="vim"
export TERM="screen-256color"
export VIM_COLOUR_SCHEME="solarized"
export CLICOLOR=1
export JAVA_HOME=$(/usr/libexec/java_home)
export PATH="/usr/local/heroku/bin:$PATH"
export PATH=/opt/local/bin:/opt/local/sbin:$PATH
export PATH="$HOME/.node/bin:$PATH"
export PATH="node_modules/.bin:$PATH"
export PATH=$HOME/.rbenv/shims:$PATH
export PATH="$HOME/Code/neo4j-2.2.3/bin:$PATH"
export GOPATH=$HOME/Code/Go
export PATH=$PATH:$GOROOT/bin
export PATH="/usr/local/opt/gnu-tar/libexec/gnubin:$PATH"
| export GITHUB_USER="phamann"
export PATH=/usr/local/bin:/usr/bin:/bin:/usr/sbin:/sbin:$PATH # Reorder PATH so local bin is first
export GREP_OPTIONS='--color=auto'
export GREP_COLOR='1;32'
export MANPAGER="less -X" # Don’t clear the screen after quitting a manual page
export EDITOR="vim"
export TERM="screen-256color"
+ export VIM_COLOUR_SCHEME="solarized"
export CLICOLOR=1
export JAVA_HOME=$(/usr/libexec/java_home)
export PATH="/usr/local/heroku/bin:$PATH"
export PATH=/opt/local/bin:/opt/local/sbin:$PATH
-
+ export PATH="$HOME/.node/bin:$PATH"
+ export PATH="node_modules/.bin:$PATH"
+ export PATH=$HOME/.rbenv/shims:$PATH
+ export PATH="$HOME/Code/neo4j-2.2.3/bin:$PATH"
+ export GOPATH=$HOME/Code/Go
+ export PATH=$PATH:$GOROOT/bin
+ export PATH="/usr/local/opt/gnu-tar/libexec/gnubin:$PATH" | 9 | 0.75 | 8 | 1 |
6b0c451eb4b50edb705f82b65811b7d463a526fc | README.md | README.md | Fast Status
====
[](https://travis-ci.org/lazyengineering/faststatus) [](https://waffle.io/lazyengineering/faststatus)
Simple API for sharing if a resource is free, busy, or very busy
| Fast Status
====
[](https://travis-ci.org/lazyengineering/faststatus) [](https://waffle.io/lazyengineering/faststatus)
Simple API for sharing if a resource is free, busy, or very busy
Read more about this project:
- [Introducing Fast Status](http://jessecarl.github.io/blog/2016/05/19/introducing-fast-status/)
| Add reference to intro blog entry | Add reference to intro blog entry
Adds link to blog entry introducing Fast Status.
| Markdown | mit | lazyengineering/faststatus | markdown | ## Code Before:
Fast Status
====
[](https://travis-ci.org/lazyengineering/faststatus) [](https://waffle.io/lazyengineering/faststatus)
Simple API for sharing if a resource is free, busy, or very busy
## Instruction:
Add reference to intro blog entry
Adds link to blog entry introducing Fast Status.
## Code After:
Fast Status
====
[](https://travis-ci.org/lazyengineering/faststatus) [](https://waffle.io/lazyengineering/faststatus)
Simple API for sharing if a resource is free, busy, or very busy
Read more about this project:
- [Introducing Fast Status](http://jessecarl.github.io/blog/2016/05/19/introducing-fast-status/)
| Fast Status
====
[](https://travis-ci.org/lazyengineering/faststatus) [](https://waffle.io/lazyengineering/faststatus)
Simple API for sharing if a resource is free, busy, or very busy
+
+ Read more about this project:
+
+ - [Introducing Fast Status](http://jessecarl.github.io/blog/2016/05/19/introducing-fast-status/) | 4 | 0.666667 | 4 | 0 |
9b76da782f41716fa2ec2a84c95f8d66e18b9038 | bin/start-crawler.bat | bin/start-crawler.bat | @echo off
SETLOCAL ENABLEDELAYEDEXPANSION
:: Initialize system variables and check configuration
call "%~dp0.etc\init.cmd"
IF %ERRORLEVEL% NEQ 0 exit /B %ERRORLEVEL%
pushd "!REGAIN_HOME!/runtime/crawler"
TITLE Regain crawler
java -Xmx1024m -Dfile.encoding=UTF-8 -cp !IK_EXT_HOME!/bin;!IK_HOME!/IKAnalyzer2012_u6.jar;!REGAIN_HOME!/runtime/crawler/regain-crawler.jar net.sf.regain.crawler.Main -config "!PROFILE_HOME!/.work/CrawlerConfiguration.xml" -logConfig "!PROFILE_HOME!/.work/log4j.properties"
popd
ENDLOCAL | @echo off
SETLOCAL ENABLEDELAYEDEXPANSION
:: Initialize system variables and check configuration
call "%~dp0.etc\init.cmd"
IF %ERRORLEVEL% NEQ 0 exit /B %ERRORLEVEL%
pushd "!REGAIN_HOME!/runtime/crawler"
TITLE Regain crawler
java %JAVA_OPTS% -Dfile.encoding=UTF-8 -cp !IK_EXT_HOME!/bin;!IK_HOME!/IKAnalyzer2012_u6.jar;!REGAIN_HOME!/runtime/crawler/regain-crawler.jar net.sf.regain.crawler.Main -config "!PROFILE_HOME!/.work/CrawlerConfiguration.xml" -logConfig "!PROFILE_HOME!/.work/log4j.properties"
popd
ENDLOCAL | Add %JAVA_OPTS% to crawler, so some java options such as -Xmx can be adjusted by this variable. | Add %JAVA_OPTS% to crawler, so some java options such as -Xmx can be adjusted by this variable. | Batchfile | lgpl-2.1 | thinkbase/PortableRegain,thinkbase/PortableRegain,thinkbase/PortableRegain,thinkbase/PortableRegain | batchfile | ## Code Before:
@echo off
SETLOCAL ENABLEDELAYEDEXPANSION
:: Initialize system variables and check configuration
call "%~dp0.etc\init.cmd"
IF %ERRORLEVEL% NEQ 0 exit /B %ERRORLEVEL%
pushd "!REGAIN_HOME!/runtime/crawler"
TITLE Regain crawler
java -Xmx1024m -Dfile.encoding=UTF-8 -cp !IK_EXT_HOME!/bin;!IK_HOME!/IKAnalyzer2012_u6.jar;!REGAIN_HOME!/runtime/crawler/regain-crawler.jar net.sf.regain.crawler.Main -config "!PROFILE_HOME!/.work/CrawlerConfiguration.xml" -logConfig "!PROFILE_HOME!/.work/log4j.properties"
popd
ENDLOCAL
## Instruction:
Add %JAVA_OPTS% to crawler, so some java options such as -Xmx can be adjusted by this variable.
## Code After:
@echo off
SETLOCAL ENABLEDELAYEDEXPANSION
:: Initialize system variables and check configuration
call "%~dp0.etc\init.cmd"
IF %ERRORLEVEL% NEQ 0 exit /B %ERRORLEVEL%
pushd "!REGAIN_HOME!/runtime/crawler"
TITLE Regain crawler
java %JAVA_OPTS% -Dfile.encoding=UTF-8 -cp !IK_EXT_HOME!/bin;!IK_HOME!/IKAnalyzer2012_u6.jar;!REGAIN_HOME!/runtime/crawler/regain-crawler.jar net.sf.regain.crawler.Main -config "!PROFILE_HOME!/.work/CrawlerConfiguration.xml" -logConfig "!PROFILE_HOME!/.work/log4j.properties"
popd
ENDLOCAL | @echo off
SETLOCAL ENABLEDELAYEDEXPANSION
:: Initialize system variables and check configuration
call "%~dp0.etc\init.cmd"
IF %ERRORLEVEL% NEQ 0 exit /B %ERRORLEVEL%
pushd "!REGAIN_HOME!/runtime/crawler"
TITLE Regain crawler
- java -Xmx1024m -Dfile.encoding=UTF-8 -cp !IK_EXT_HOME!/bin;!IK_HOME!/IKAnalyzer2012_u6.jar;!REGAIN_HOME!/runtime/crawler/regain-crawler.jar net.sf.regain.crawler.Main -config "!PROFILE_HOME!/.work/CrawlerConfiguration.xml" -logConfig "!PROFILE_HOME!/.work/log4j.properties"
? ^^^^^^^^^
+ java %JAVA_OPTS% -Dfile.encoding=UTF-8 -cp !IK_EXT_HOME!/bin;!IK_HOME!/IKAnalyzer2012_u6.jar;!REGAIN_HOME!/runtime/crawler/regain-crawler.jar net.sf.regain.crawler.Main -config "!PROFILE_HOME!/.work/CrawlerConfiguration.xml" -logConfig "!PROFILE_HOME!/.work/log4j.properties"
? ^^^^^^^^^^^
popd
ENDLOCAL | 2 | 0.153846 | 1 | 1 |
f7a3483b1e39affb7fa2d5777f0d57cc77a3f0d3 | HOWDOI.md | HOWDOI.md | Collection of common things that I always forget how to deal with
## Fix vim issues
```
Error detected while processing function <SNR>13_InitializePythonBuiltin:
line 23:
E887: Sorry, this command is disabled, the Python's site module could not be loaded.
Press ENTER or type command to continue
```
Uninstall and recompile vim with `brew remove vim && brew install --build-from-source vim`.
```
_arguments:448: _vim_files: function definition file not found
```
https://github.com/robbyrussell/oh-my-zsh/issues/518
Fix:
```
rm ~/.zcompdump*
exec zsh
```
Needs to be run in every open terminal.
## fix the session crashing after the first command after `zot`
Did you reinstall the zsh-users/autosuggestions module? It breaks your terminal
every time stop trying to use it you don't even like it.
## slow git prompt in vm
if [[ -d "/vagrant" ]]; then
export GIT_PS1_SHOWDIRTYSTATE=
export GIT_PS1_SHOWUNTRACKEDFILES=
fi
| Collection of common things that I always forget how to deal with
## Fix vim issues
```
Error detected while processing function <SNR>13_InitializePythonBuiltin:
line 23:
E887: Sorry, this command is disabled, the Python's site module could not be loaded.
Press ENTER or type command to continue
```
Uninstall and recompile vim with `brew remove vim && brew install --build-from-source vim`.
```
_arguments:448: _vim_files: function definition file not found
```
https://github.com/robbyrussell/oh-my-zsh/issues/518
Fix:
```
rm ~/.zcompdump*
exec zsh
```
Needs to be run in every open terminal.
Missing backups or swaps?
```
mkdir ~/.vim/backups
mkdir ~/.vim/swaps
```
## fix the session crashing after the first command after `zot`
Did you reinstall the zsh-users/autosuggestions module? It breaks your terminal
every time stop trying to use it you don't even like it.
## slow git prompt in vm
if [[ -d "/vagrant" ]]; then
export GIT_PS1_SHOWDIRTYSTATE=
export GIT_PS1_SHOWUNTRACKEDFILES=
fi
| Add note on missing vim dirs. | Add note on missing vim dirs. | Markdown | mit | dersam/zotfiles | markdown | ## Code Before:
Collection of common things that I always forget how to deal with
## Fix vim issues
```
Error detected while processing function <SNR>13_InitializePythonBuiltin:
line 23:
E887: Sorry, this command is disabled, the Python's site module could not be loaded.
Press ENTER or type command to continue
```
Uninstall and recompile vim with `brew remove vim && brew install --build-from-source vim`.
```
_arguments:448: _vim_files: function definition file not found
```
https://github.com/robbyrussell/oh-my-zsh/issues/518
Fix:
```
rm ~/.zcompdump*
exec zsh
```
Needs to be run in every open terminal.
## fix the session crashing after the first command after `zot`
Did you reinstall the zsh-users/autosuggestions module? It breaks your terminal
every time stop trying to use it you don't even like it.
## slow git prompt in vm
if [[ -d "/vagrant" ]]; then
export GIT_PS1_SHOWDIRTYSTATE=
export GIT_PS1_SHOWUNTRACKEDFILES=
fi
## Instruction:
Add note on missing vim dirs.
## Code After:
Collection of common things that I always forget how to deal with
## Fix vim issues
```
Error detected while processing function <SNR>13_InitializePythonBuiltin:
line 23:
E887: Sorry, this command is disabled, the Python's site module could not be loaded.
Press ENTER or type command to continue
```
Uninstall and recompile vim with `brew remove vim && brew install --build-from-source vim`.
```
_arguments:448: _vim_files: function definition file not found
```
https://github.com/robbyrussell/oh-my-zsh/issues/518
Fix:
```
rm ~/.zcompdump*
exec zsh
```
Needs to be run in every open terminal.
Missing backups or swaps?
```
mkdir ~/.vim/backups
mkdir ~/.vim/swaps
```
## fix the session crashing after the first command after `zot`
Did you reinstall the zsh-users/autosuggestions module? It breaks your terminal
every time stop trying to use it you don't even like it.
## slow git prompt in vm
if [[ -d "/vagrant" ]]; then
export GIT_PS1_SHOWDIRTYSTATE=
export GIT_PS1_SHOWUNTRACKEDFILES=
fi
| Collection of common things that I always forget how to deal with
## Fix vim issues
```
Error detected while processing function <SNR>13_InitializePythonBuiltin:
line 23:
E887: Sorry, this command is disabled, the Python's site module could not be loaded.
Press ENTER or type command to continue
```
Uninstall and recompile vim with `brew remove vim && brew install --build-from-source vim`.
```
_arguments:448: _vim_files: function definition file not found
```
https://github.com/robbyrussell/oh-my-zsh/issues/518
Fix:
```
rm ~/.zcompdump*
exec zsh
```
Needs to be run in every open terminal.
+ Missing backups or swaps?
+ ```
+ mkdir ~/.vim/backups
+ mkdir ~/.vim/swaps
+ ```
+
## fix the session crashing after the first command after `zot`
Did you reinstall the zsh-users/autosuggestions module? It breaks your terminal
every time stop trying to use it you don't even like it.
## slow git prompt in vm
if [[ -d "/vagrant" ]]; then
export GIT_PS1_SHOWDIRTYSTATE=
export GIT_PS1_SHOWUNTRACKEDFILES=
fi | 6 | 0.181818 | 6 | 0 |
5fe53a31bd7f37f8d9bd4fbe3796c8a0fa85019a | storm/db.py | storm/db.py | import motor
import error
from tornado import gen
class Connection(object):
def __init__(self, host='localhost', port=None, db=None):
self.host = host
self.port = port
self.db = db
class Database(object):
def __init__(self, connection):
if not isinstance(connection, Connection):
raise error.StormError('connection must be instance of storm.db.Connection')
self.connection = connection
self.is_connected = False
class MongoDb(Database):
def connect(self):
if self.is_connected:
return
self.motor_client = motor.MotorClient(
self.connection.host,
self.connection.port
).open_sync()
self.db = self.motor_client[self.connection.db]
self.is_connected = True
@gen.coroutine
def select_one(self, table, **args):
self.connect()
result = yield motor.Op(getattr(self.db, table).find_one, args)
if result is None:
raise error.StormNotFoundError("Object of type: %s not found with args: %s" % (table, args))
callback = args.get('callback')
if callback is None:
raise gen.Return(result)
callback(result)
@gen.coroutine
def insert(self, table, data, callback=None):
self.connect()
result = yield motor.Op(self.db[table].insert, data)
if callback is None:
raise gen.Return(result)
callback(result)
| import motor
import error
from tornado import gen
from bson.objectid import ObjectId
class Connection(object):
def __init__(self, host='localhost', port=None, db=None):
self.host = host
self.port = port
self.db = db
class Database(object):
def __init__(self, connection):
if not isinstance(connection, Connection):
raise error.StormError('connection must be instance of storm.db.Connection')
self.connection = connection
self.is_connected = False
class MongoDb(Database):
def connect(self):
if self.is_connected:
return
self.motor_client = motor.MotorClient(
self.connection.host,
self.connection.port
).open_sync()
self.db = self.motor_client[self.connection.db]
self.is_connected = True
@gen.coroutine
def select_one(self, table, **kwargs):
self.connect()
if '_id' in kwargs:
kwargs['_id'] = ObjectId(kwargs['_id'])
result = yield motor.Op(getattr(self.db, table).find_one, **kwargs)
if result is None:
raise error.StormNotFoundError("Object of type: %s not found with args: %s" % (table, kwargs))
callback = kwargs.get('callback')
if callback is None:
raise gen.Return(result)
callback(result)
@gen.coroutine
def insert(self, table, data, callback=None):
self.connect()
result = yield motor.Op(self.db[table].insert, data)
if callback is None:
raise gen.Return(result)
callback(result)
| Make sure looking up by id works correctly | Make sure looking up by id works correctly
| Python | mit | ccampbell/storm,liujiantong/storm | python | ## Code Before:
import motor
import error
from tornado import gen
class Connection(object):
def __init__(self, host='localhost', port=None, db=None):
self.host = host
self.port = port
self.db = db
class Database(object):
def __init__(self, connection):
if not isinstance(connection, Connection):
raise error.StormError('connection must be instance of storm.db.Connection')
self.connection = connection
self.is_connected = False
class MongoDb(Database):
def connect(self):
if self.is_connected:
return
self.motor_client = motor.MotorClient(
self.connection.host,
self.connection.port
).open_sync()
self.db = self.motor_client[self.connection.db]
self.is_connected = True
@gen.coroutine
def select_one(self, table, **args):
self.connect()
result = yield motor.Op(getattr(self.db, table).find_one, args)
if result is None:
raise error.StormNotFoundError("Object of type: %s not found with args: %s" % (table, args))
callback = args.get('callback')
if callback is None:
raise gen.Return(result)
callback(result)
@gen.coroutine
def insert(self, table, data, callback=None):
self.connect()
result = yield motor.Op(self.db[table].insert, data)
if callback is None:
raise gen.Return(result)
callback(result)
## Instruction:
Make sure looking up by id works correctly
## Code After:
import motor
import error
from tornado import gen
from bson.objectid import ObjectId
class Connection(object):
def __init__(self, host='localhost', port=None, db=None):
self.host = host
self.port = port
self.db = db
class Database(object):
def __init__(self, connection):
if not isinstance(connection, Connection):
raise error.StormError('connection must be instance of storm.db.Connection')
self.connection = connection
self.is_connected = False
class MongoDb(Database):
def connect(self):
if self.is_connected:
return
self.motor_client = motor.MotorClient(
self.connection.host,
self.connection.port
).open_sync()
self.db = self.motor_client[self.connection.db]
self.is_connected = True
@gen.coroutine
def select_one(self, table, **kwargs):
self.connect()
if '_id' in kwargs:
kwargs['_id'] = ObjectId(kwargs['_id'])
result = yield motor.Op(getattr(self.db, table).find_one, **kwargs)
if result is None:
raise error.StormNotFoundError("Object of type: %s not found with args: %s" % (table, kwargs))
callback = kwargs.get('callback')
if callback is None:
raise gen.Return(result)
callback(result)
@gen.coroutine
def insert(self, table, data, callback=None):
self.connect()
result = yield motor.Op(self.db[table].insert, data)
if callback is None:
raise gen.Return(result)
callback(result)
| import motor
import error
from tornado import gen
+ from bson.objectid import ObjectId
class Connection(object):
def __init__(self, host='localhost', port=None, db=None):
self.host = host
self.port = port
self.db = db
class Database(object):
def __init__(self, connection):
if not isinstance(connection, Connection):
raise error.StormError('connection must be instance of storm.db.Connection')
self.connection = connection
self.is_connected = False
class MongoDb(Database):
def connect(self):
if self.is_connected:
return
self.motor_client = motor.MotorClient(
self.connection.host,
self.connection.port
).open_sync()
self.db = self.motor_client[self.connection.db]
self.is_connected = True
@gen.coroutine
- def select_one(self, table, **args):
+ def select_one(self, table, **kwargs):
? ++
self.connect()
+ if '_id' in kwargs:
+ kwargs['_id'] = ObjectId(kwargs['_id'])
+
- result = yield motor.Op(getattr(self.db, table).find_one, args)
+ result = yield motor.Op(getattr(self.db, table).find_one, **kwargs)
? ++++
if result is None:
- raise error.StormNotFoundError("Object of type: %s not found with args: %s" % (table, args))
+ raise error.StormNotFoundError("Object of type: %s not found with args: %s" % (table, kwargs))
? ++
- callback = args.get('callback')
+ callback = kwargs.get('callback')
? ++
if callback is None:
raise gen.Return(result)
callback(result)
@gen.coroutine
def insert(self, table, data, callback=None):
self.connect()
result = yield motor.Op(self.db[table].insert, data)
if callback is None:
raise gen.Return(result)
callback(result) | 12 | 0.2 | 8 | 4 |
d8b56597fe8b5e34abd1499eb9f5252447bba563 | testsuite/tests/polykinds/T5798.hs | testsuite/tests/polykinds/T5798.hs | {-# LANGUAGE PolyKinds #-}
module T5798 where
data Proxy t = ProxyC
test :: Proxy '[Int, Bool]
test = ProxyC
| {-# LANGUAGE PolyKinds, DataKinds #-}
module T5798 where
data Proxy t = ProxyC
test :: Proxy '[Int, Bool]
test = ProxyC
| Add DataKinds flag to test | Add DataKinds flag to test
| Haskell | bsd-3-clause | forked-upstream-packages-for-ghcjs/ghc,TomMD/ghc,hferreiro/replay,christiaanb/ghc,sdiehl/ghc,tjakway/ghcjvm,sdiehl/ghc,nushio3/ghc,spacekitteh/smcghc,mcschroeder/ghc,mcschroeder/ghc,acowley/ghc,gcampax/ghc,forked-upstream-packages-for-ghcjs/ghc,elieux/ghc,vikraman/ghc,snoyberg/ghc,fmthoma/ghc,jstolarek/ghc,lukexi/ghc,christiaanb/ghc,anton-dessiatov/ghc,gcampax/ghc,lukexi/ghc-7.8-arm64,bitemyapp/ghc,tjakway/ghcjvm,shlevy/ghc,sgillespie/ghc,hferreiro/replay,fmthoma/ghc,spacekitteh/smcghc,vikraman/ghc,nathyong/microghc-ghc,AlexanderPankiv/ghc,nathyong/microghc-ghc,wxwxwwxxx/ghc,acowley/ghc,anton-dessiatov/ghc,nkaretnikov/ghc,tibbe/ghc,ml9951/ghc,hferreiro/replay,lukexi/ghc-7.8-arm64,nushio3/ghc,TomMD/ghc,olsner/ghc,GaloisInc/halvm-ghc,sdiehl/ghc,gridaphobe/ghc,GaloisInc/halvm-ghc,snoyberg/ghc,snoyberg/ghc,gridaphobe/ghc,sdiehl/ghc,elieux/ghc,acowley/ghc,ml9951/ghc,ghc-android/ghc,siddhanathan/ghc,bitemyapp/ghc,TomMD/ghc,elieux/ghc,fmthoma/ghc,urbanslug/ghc,olsner/ghc,olsner/ghc,lukexi/ghc-7.8-arm64,vTurbine/ghc,nathyong/microghc-ghc,oldmanmike/ghc,ezyang/ghc,ezyang/ghc,olsner/ghc,siddhanathan/ghc,lukexi/ghc,hferreiro/replay,vTurbine/ghc,mettekou/ghc,frantisekfarka/ghc-dsi,hferreiro/replay,wxwxwwxxx/ghc,christiaanb/ghc,mfine/ghc,siddhanathan/ghc,tibbe/ghc,gcampax/ghc,siddhanathan/ghc,nkaretnikov/ghc,shlevy/ghc,anton-dessiatov/ghc,mfine/ghc,urbanslug/ghc,nushio3/ghc,mettekou/ghc,fmthoma/ghc,ml9951/ghc,ezyang/ghc,shlevy/ghc,vikraman/ghc,acowley/ghc,urbanslug/ghc,urbanslug/ghc,GaloisInc/halvm-ghc,da-x/ghc,vTurbine/ghc,wxwxwwxxx/ghc,ghc-android/ghc,TomMD/ghc,AlexanderPankiv/ghc,GaloisInc/halvm-ghc,da-x/ghc,wxwxwwxxx/ghc,vikraman/ghc,mettekou/ghc,frantisekfarka/ghc-dsi,jstolarek/ghc,mfine/ghc,nkaretnikov/ghc,siddhanathan/ghc,ryantm/ghc,anton-dessiatov/ghc,forked-upstream-packages-for-ghcjs/ghc,snoyberg/ghc,tjakway/ghcjvm,sdiehl/ghc,tjakway/ghcjvm,anton-dessiatov/ghc,shlevy/ghc,green-haskell/ghc,shlevy/ghc,snoyberg/ghc,AlexanderPankiv/ghc,forked-upstream-packages-for-ghcjs/ghc,frantisekfarka/ghc-dsi,vTurbine/ghc,mcschroeder/ghc,mettekou/ghc,bitemyapp/ghc,sdiehl/ghc,GaloisInc/halvm-ghc,ryantm/ghc,gridaphobe/ghc,TomMD/ghc,ezyang/ghc,wxwxwwxxx/ghc,sgillespie/ghc,wxwxwwxxx/ghc,gcampax/ghc,ml9951/ghc,urbanslug/ghc,sgillespie/ghc,AlexanderPankiv/ghc,frantisekfarka/ghc-dsi,mcschroeder/ghc,siddhanathan/ghc,nathyong/microghc-ghc,ghc-android/ghc,snoyberg/ghc,ml9951/ghc,hferreiro/replay,mcschroeder/ghc,sgillespie/ghc,snoyberg/ghc,mettekou/ghc,ryantm/ghc,nathyong/microghc-ghc,lukexi/ghc-7.8-arm64,acowley/ghc,da-x/ghc,da-x/ghc,tibbe/ghc,gridaphobe/ghc,fmthoma/ghc,TomMD/ghc,mcschroeder/ghc,da-x/ghc,fmthoma/ghc,bitemyapp/ghc,nushio3/ghc,tjakway/ghcjvm,vTurbine/ghc,elieux/ghc,siddhanathan/ghc,tjakway/ghcjvm,AlexanderPankiv/ghc,shlevy/ghc,urbanslug/ghc,ml9951/ghc,sdiehl/ghc,frantisekfarka/ghc-dsi,ghc-android/ghc,nkaretnikov/ghc,AlexanderPankiv/ghc,oldmanmike/ghc,da-x/ghc,mettekou/ghc,oldmanmike/ghc,tibbe/ghc,gcampax/ghc,nushio3/ghc,green-haskell/ghc,acowley/ghc,mfine/ghc,shlevy/ghc,christiaanb/ghc,ghc-android/ghc,ghc-android/ghc,mfine/ghc,christiaanb/ghc,oldmanmike/ghc,mettekou/ghc,vTurbine/ghc,sgillespie/ghc,olsner/ghc,spacekitteh/smcghc,ryantm/ghc,GaloisInc/halvm-ghc,olsner/ghc,vTurbine/ghc,ezyang/ghc,nushio3/ghc,nushio3/ghc,green-haskell/ghc,ml9951/ghc,elieux/ghc,forked-upstream-packages-for-ghcjs/ghc,TomMD/ghc,gridaphobe/ghc,bitemyapp/ghc,oldmanmike/ghc,nkaretnikov/ghc,gcampax/ghc,sgillespie/ghc,gcampax/ghc,holzensp/ghc,vikraman/ghc,holzensp/ghc,nkaretnikov/ghc,christiaanb/ghc,ezyang/ghc,nathyong/microghc-ghc,lukexi/ghc-7.8-arm64,oldmanmike/ghc,forked-upstream-packages-for-ghcjs/ghc,ryantm/ghc,acowley/ghc,hferreiro/replay,spacekitteh/smcghc,forked-upstream-packages-for-ghcjs/ghc,lukexi/ghc,oldmanmike/ghc,gridaphobe/ghc,lukexi/ghc,mfine/ghc,tjakway/ghcjvm,elieux/ghc,elieux/ghc,christiaanb/ghc,jstolarek/ghc,spacekitteh/smcghc,ml9951/ghc,sgillespie/ghc,lukexi/ghc,ezyang/ghc,jstolarek/ghc,anton-dessiatov/ghc,olsner/ghc,da-x/ghc,mcschroeder/ghc,AlexanderPankiv/ghc,anton-dessiatov/ghc,holzensp/ghc,nkaretnikov/ghc,holzensp/ghc,ghc-android/ghc,GaloisInc/halvm-ghc,holzensp/ghc,mfine/ghc,fmthoma/ghc,nathyong/microghc-ghc,jstolarek/ghc,green-haskell/ghc,tibbe/ghc,urbanslug/ghc,green-haskell/ghc,vikraman/ghc,gridaphobe/ghc,wxwxwwxxx/ghc,vikraman/ghc | haskell | ## Code Before:
{-# LANGUAGE PolyKinds #-}
module T5798 where
data Proxy t = ProxyC
test :: Proxy '[Int, Bool]
test = ProxyC
## Instruction:
Add DataKinds flag to test
## Code After:
{-# LANGUAGE PolyKinds, DataKinds #-}
module T5798 where
data Proxy t = ProxyC
test :: Proxy '[Int, Bool]
test = ProxyC
| - {-# LANGUAGE PolyKinds #-}
+ {-# LANGUAGE PolyKinds, DataKinds #-}
? +++++++++++
module T5798 where
data Proxy t = ProxyC
test :: Proxy '[Int, Bool]
test = ProxyC | 2 | 0.25 | 1 | 1 |
6d56186d8b24aec350fcd390bfc68300e579a0ec | .travis.yml | .travis.yml | before_install:
- sudo apt-get update -qq
- sudo apt-get install -qq exim4
- sudo service exim4 start
- sleep 1
- sudo netstat -tulpen
language: go
go:
- 1.6
| before_install:
- sudo apt-get update -qq
- sudo apt-get install -qq exim4
- sudo service exim4 start
- sleep 1
- sudo iptables -v -L
- sudo iptables -F INPUT
- sudo iptables -F OUTPUT
- sudo iptables -F FORWARD
- sudo netstat -tulpen
language: go
go:
- 1.6
| Allow connection to :25 for tests | Travis: Allow connection to :25 for tests
| YAML | apache-2.0 | TNG/openpgp-validation-server,TNG/openpgp-validation-server,TNG/gpg-validation-server,TNG/gpg-validation-server | yaml | ## Code Before:
before_install:
- sudo apt-get update -qq
- sudo apt-get install -qq exim4
- sudo service exim4 start
- sleep 1
- sudo netstat -tulpen
language: go
go:
- 1.6
## Instruction:
Travis: Allow connection to :25 for tests
## Code After:
before_install:
- sudo apt-get update -qq
- sudo apt-get install -qq exim4
- sudo service exim4 start
- sleep 1
- sudo iptables -v -L
- sudo iptables -F INPUT
- sudo iptables -F OUTPUT
- sudo iptables -F FORWARD
- sudo netstat -tulpen
language: go
go:
- 1.6
| before_install:
- sudo apt-get update -qq
- sudo apt-get install -qq exim4
- sudo service exim4 start
- sleep 1
+ - sudo iptables -v -L
+ - sudo iptables -F INPUT
+ - sudo iptables -F OUTPUT
+ - sudo iptables -F FORWARD
- sudo netstat -tulpen
language: go
go:
- 1.6 | 4 | 0.363636 | 4 | 0 |
318846b58ad173404c25c911c84a7618a7f8baa5 | 723e_web/app/scripts/models/userModel.js | 723e_web/app/scripts/models/userModel.js | define(["jquery", "backbone", "ws"], function($, Backbone, WebServices) {
var user = Backbone.Model.extend({
urlRoot: WebServices.v1.users,
defaults: {
},
initialize: function() {
//alert("Welcome to this world");
},
currency: function(){
var accounts = this.get('accounts');
if(accounts.length === 0){
return null;
}
return this.get('accounts')[0].currency;
}
});
return user;
});
| define(["jquery", "backbone", "ws"], function($, Backbone, WebServices) {
var user = Backbone.Model.extend({
urlRoot: WebServices.v1.users,
defaults: {
},
initialize: function() {
//alert("Welcome to this world");
},
currency: function(){
var accounts = this.get('accounts');
return this.get('accounts')[0].currency;
}
});
return user;
});
| Fix bug with accounts.length with web interface | Fix bug with accounts.length with web interface
| JavaScript | mit | sebastienbarbier/723e_server,sebastienbarbier/723e,Natim/723e,Natim/723e,sebastienbarbier/723e_server,sebastienbarbier/723e,Natim/723e | javascript | ## Code Before:
define(["jquery", "backbone", "ws"], function($, Backbone, WebServices) {
var user = Backbone.Model.extend({
urlRoot: WebServices.v1.users,
defaults: {
},
initialize: function() {
//alert("Welcome to this world");
},
currency: function(){
var accounts = this.get('accounts');
if(accounts.length === 0){
return null;
}
return this.get('accounts')[0].currency;
}
});
return user;
});
## Instruction:
Fix bug with accounts.length with web interface
## Code After:
define(["jquery", "backbone", "ws"], function($, Backbone, WebServices) {
var user = Backbone.Model.extend({
urlRoot: WebServices.v1.users,
defaults: {
},
initialize: function() {
//alert("Welcome to this world");
},
currency: function(){
var accounts = this.get('accounts');
return this.get('accounts')[0].currency;
}
});
return user;
});
| define(["jquery", "backbone", "ws"], function($, Backbone, WebServices) {
var user = Backbone.Model.extend({
urlRoot: WebServices.v1.users,
defaults: {
},
initialize: function() {
//alert("Welcome to this world");
},
currency: function(){
var accounts = this.get('accounts');
- if(accounts.length === 0){
- return null;
- }
return this.get('accounts')[0].currency;
}
});
return user;
}); | 3 | 0.136364 | 0 | 3 |
9cbe31ee975e7223b11ff61626828ea5b095f99f | PrehensilePonyTail/PPTail/Properties/launchSettings.json | PrehensilePonyTail/PPTail/Properties/launchSettings.json | {
"profiles": {
"PPTail": {
"commandName": "Project",
"environmentVariables": {
"outputPath": "C:\\Users\\barry\\Documents\\My Web Sites\\CognitiveInheritance",
"sourceDataPath": "C:\\Users\\barry\\Documents\\My Web Sites\\CognitiveInheritanceData"
}
},
"win7-dev-box": {
"commandName": "Project",
"environmentVariables": {
"outputPath": "C:\\Users\\bstahl\\Documents\\My Web Sites\\CognitiveInheritance\\Generated",
"sourceDataPath": "C:\\Users\\bstahl\\Documents\\CognitiveInheritanceData"
}
}
}
} | {
"profiles": {
"PPTail": {
"commandName": "Project",
"environmentVariables": {
"outputPath": "C:\\OutputDataLocation",
"sourceDataPath": "C:\\SourceDataLocation"
}
},
"mobile-dev-box": {
"commandName": "Project",
"environmentVariables": {
"outputPath": "E:\\GeneratedWebsite",
"sourceDataPath": "E:\\CognitiveInheritance\\App_Data"
}
},
"win7-dev-box": {
"commandName": "Project",
"environmentVariables": {
"outputPath": "C:\\Users\\bstahl\\Documents\\My Web Sites\\CognitiveInheritance\\Generated",
"sourceDataPath": "C:\\Users\\bstahl\\Documents\\CognitiveInheritanceData"
}
}
}
} | Add environment settings for Mobile development box. | Add environment settings for Mobile development box.
| JSON | mit | bsstahl/PPTail,bsstahl/PPTail | json | ## Code Before:
{
"profiles": {
"PPTail": {
"commandName": "Project",
"environmentVariables": {
"outputPath": "C:\\Users\\barry\\Documents\\My Web Sites\\CognitiveInheritance",
"sourceDataPath": "C:\\Users\\barry\\Documents\\My Web Sites\\CognitiveInheritanceData"
}
},
"win7-dev-box": {
"commandName": "Project",
"environmentVariables": {
"outputPath": "C:\\Users\\bstahl\\Documents\\My Web Sites\\CognitiveInheritance\\Generated",
"sourceDataPath": "C:\\Users\\bstahl\\Documents\\CognitiveInheritanceData"
}
}
}
}
## Instruction:
Add environment settings for Mobile development box.
## Code After:
{
"profiles": {
"PPTail": {
"commandName": "Project",
"environmentVariables": {
"outputPath": "C:\\OutputDataLocation",
"sourceDataPath": "C:\\SourceDataLocation"
}
},
"mobile-dev-box": {
"commandName": "Project",
"environmentVariables": {
"outputPath": "E:\\GeneratedWebsite",
"sourceDataPath": "E:\\CognitiveInheritance\\App_Data"
}
},
"win7-dev-box": {
"commandName": "Project",
"environmentVariables": {
"outputPath": "C:\\Users\\bstahl\\Documents\\My Web Sites\\CognitiveInheritance\\Generated",
"sourceDataPath": "C:\\Users\\bstahl\\Documents\\CognitiveInheritanceData"
}
}
}
} | {
"profiles": {
"PPTail": {
"commandName": "Project",
"environmentVariables": {
- "outputPath": "C:\\Users\\barry\\Documents\\My Web Sites\\CognitiveInheritance",
- "sourceDataPath": "C:\\Users\\barry\\Documents\\My Web Sites\\CognitiveInheritanceData"
+ "outputPath": "C:\\OutputDataLocation",
+ "sourceDataPath": "C:\\SourceDataLocation"
+ }
+ },
+ "mobile-dev-box": {
+ "commandName": "Project",
+ "environmentVariables": {
+ "outputPath": "E:\\GeneratedWebsite",
+ "sourceDataPath": "E:\\CognitiveInheritance\\App_Data"
}
},
"win7-dev-box": {
"commandName": "Project",
"environmentVariables": {
"outputPath": "C:\\Users\\bstahl\\Documents\\My Web Sites\\CognitiveInheritance\\Generated",
"sourceDataPath": "C:\\Users\\bstahl\\Documents\\CognitiveInheritanceData"
}
}
}
} | 11 | 0.611111 | 9 | 2 |
5851a686bd84a60a1d7e14afd605ccb00f2988cd | tasks/loadHerokuEnv.js | tasks/loadHerokuEnv.js | import { herokuConfig } from './util'
import { each, includes } from 'lodash'
const keys = [
'ASSET_HOST',
'ASSET_PATH',
'AWS_ACCESS_KEY_ID',
'AWS_S3_BUCKET',
'AWS_S3_HOST',
'AWS_SECRET_ACCESS_KEY',
'FACEBOOK_APP_ID',
'FILEPICKER_API_KEY',
'GOOGLE_BROWSER_KEY',
'GOOGLE_CLIENT_ID',
'HOST',
'LOG_LEVEL',
'NODE_ENV',
'SEGMENT_KEY',
'SLACK_APP_CLIENT_ID',
'SOCKET_HOST',
'UPSTREAM_HOST'
]
export default function (done) {
herokuConfig().info((err, vars) => {
if (err) done(err)
each(vars, (v, k) => {
if (includes(keys, k) || k.startsWith('FEATURE_FLAG_')) {
process.env[k] = v
}
})
done()
})
}
| import { herokuConfig } from './util'
import { extend, pick } from 'lodash'
const keys = [
'ASSET_HOST',
'ASSET_PATH',
'AWS_ACCESS_KEY_ID',
'AWS_S3_BUCKET',
'AWS_S3_HOST',
'AWS_SECRET_ACCESS_KEY',
'FACEBOOK_APP_ID',
'FILEPICKER_API_KEY',
'GOOGLE_BROWSER_KEY',
'GOOGLE_CLIENT_ID',
'HOST',
'LOG_LEVEL',
'NODE_ENV',
'SEGMENT_KEY',
'SLACK_APP_CLIENT_ID',
'SOCKET_HOST',
'UPSTREAM_HOST'
]
export default function (done) {
herokuConfig().info((err, vars) => {
if (err) done(err)
extend(process.env, pick(vars, keys))
done()
})
}
| Revert "add all feature flag env variables during deployment" | Revert "add all feature flag env variables during deployment"
This reverts commit 30054de694abf26b8871ff06421a930012019390.
This is unnecessary because we’re not setting feature flag env
variables on the client side using envify (see previous commit).
| JavaScript | agpl-3.0 | Hylozoic/hylo-redux,Hylozoic/hylo-redux | javascript | ## Code Before:
import { herokuConfig } from './util'
import { each, includes } from 'lodash'
const keys = [
'ASSET_HOST',
'ASSET_PATH',
'AWS_ACCESS_KEY_ID',
'AWS_S3_BUCKET',
'AWS_S3_HOST',
'AWS_SECRET_ACCESS_KEY',
'FACEBOOK_APP_ID',
'FILEPICKER_API_KEY',
'GOOGLE_BROWSER_KEY',
'GOOGLE_CLIENT_ID',
'HOST',
'LOG_LEVEL',
'NODE_ENV',
'SEGMENT_KEY',
'SLACK_APP_CLIENT_ID',
'SOCKET_HOST',
'UPSTREAM_HOST'
]
export default function (done) {
herokuConfig().info((err, vars) => {
if (err) done(err)
each(vars, (v, k) => {
if (includes(keys, k) || k.startsWith('FEATURE_FLAG_')) {
process.env[k] = v
}
})
done()
})
}
## Instruction:
Revert "add all feature flag env variables during deployment"
This reverts commit 30054de694abf26b8871ff06421a930012019390.
This is unnecessary because we’re not setting feature flag env
variables on the client side using envify (see previous commit).
## Code After:
import { herokuConfig } from './util'
import { extend, pick } from 'lodash'
const keys = [
'ASSET_HOST',
'ASSET_PATH',
'AWS_ACCESS_KEY_ID',
'AWS_S3_BUCKET',
'AWS_S3_HOST',
'AWS_SECRET_ACCESS_KEY',
'FACEBOOK_APP_ID',
'FILEPICKER_API_KEY',
'GOOGLE_BROWSER_KEY',
'GOOGLE_CLIENT_ID',
'HOST',
'LOG_LEVEL',
'NODE_ENV',
'SEGMENT_KEY',
'SLACK_APP_CLIENT_ID',
'SOCKET_HOST',
'UPSTREAM_HOST'
]
export default function (done) {
herokuConfig().info((err, vars) => {
if (err) done(err)
extend(process.env, pick(vars, keys))
done()
})
}
| import { herokuConfig } from './util'
- import { each, includes } from 'lodash'
? ^^^ - ^^^^^
+ import { extend, pick } from 'lodash'
? ^^^^^ + ^
const keys = [
'ASSET_HOST',
'ASSET_PATH',
'AWS_ACCESS_KEY_ID',
'AWS_S3_BUCKET',
'AWS_S3_HOST',
'AWS_SECRET_ACCESS_KEY',
'FACEBOOK_APP_ID',
'FILEPICKER_API_KEY',
'GOOGLE_BROWSER_KEY',
'GOOGLE_CLIENT_ID',
'HOST',
'LOG_LEVEL',
'NODE_ENV',
'SEGMENT_KEY',
'SLACK_APP_CLIENT_ID',
'SOCKET_HOST',
'UPSTREAM_HOST'
]
export default function (done) {
herokuConfig().info((err, vars) => {
if (err) done(err)
- each(vars, (v, k) => {
- if (includes(keys, k) || k.startsWith('FEATURE_FLAG_')) {
- process.env[k] = v
- }
- })
+ extend(process.env, pick(vars, keys))
done()
})
} | 8 | 0.228571 | 2 | 6 |
2b88f8f458781bd88f559f1a5a966fd5050414a0 | tests/merchandise/music/test_models.py | tests/merchandise/music/test_models.py | import pytest
from components.merchandise.music.models import Album, Single
from components.merchandise.music.factories import (AlbumFactory,
BaseFactory, SingleFactory)
@pytest.mark.django_db
class TestAlbums(object):
def test_album_factory(self):
album = AlbumFactory()
assert isinstance(album, Album)
assert 'album' in album.romanized_name
assert album.identifier == 'album'
@pytest.mark.django_db
class TestSingles(object):
def test_single_factory(self):
single = SingleFactory()
assert isinstance(single, Single)
assert 'single' in single.romanized_name
assert single.identifier == 'single'
| import pytest
from components.merchandise.music.models import Album, Single
from components.merchandise.music.factories import (AlbumFactory,
BaseFactory, SingleFactory)
@pytest.mark.django_db
def test_album_factory():
factory = AlbumFactory()
assert isinstance(factory, Album)
assert 'album' in factory.romanized_name
assert factory.identifier == 'album'
@pytest.mark.django_db
def test_single_factory():
factory = SingleFactory()
assert isinstance(factory, Single)
assert 'single' in factory.romanized_name
assert factory.identifier == 'single'
| Remove the class surrounding the music tests. Staying strictly functional. | Remove the class surrounding the music tests. Staying strictly functional.
| Python | apache-2.0 | hello-base/web,hello-base/web,hello-base/web,hello-base/web | python | ## Code Before:
import pytest
from components.merchandise.music.models import Album, Single
from components.merchandise.music.factories import (AlbumFactory,
BaseFactory, SingleFactory)
@pytest.mark.django_db
class TestAlbums(object):
def test_album_factory(self):
album = AlbumFactory()
assert isinstance(album, Album)
assert 'album' in album.romanized_name
assert album.identifier == 'album'
@pytest.mark.django_db
class TestSingles(object):
def test_single_factory(self):
single = SingleFactory()
assert isinstance(single, Single)
assert 'single' in single.romanized_name
assert single.identifier == 'single'
## Instruction:
Remove the class surrounding the music tests. Staying strictly functional.
## Code After:
import pytest
from components.merchandise.music.models import Album, Single
from components.merchandise.music.factories import (AlbumFactory,
BaseFactory, SingleFactory)
@pytest.mark.django_db
def test_album_factory():
factory = AlbumFactory()
assert isinstance(factory, Album)
assert 'album' in factory.romanized_name
assert factory.identifier == 'album'
@pytest.mark.django_db
def test_single_factory():
factory = SingleFactory()
assert isinstance(factory, Single)
assert 'single' in factory.romanized_name
assert factory.identifier == 'single'
| import pytest
from components.merchandise.music.models import Album, Single
from components.merchandise.music.factories import (AlbumFactory,
BaseFactory, SingleFactory)
@pytest.mark.django_db
- class TestAlbums(object):
- def test_album_factory(self):
? ---- ----
+ def test_album_factory():
- album = AlbumFactory()
? ^^^^ ^^^^
+ factory = AlbumFactory()
? ^ ^^^^^
- assert isinstance(album, Album)
? ---- ^^^^
+ assert isinstance(factory, Album)
? + ^^^^^
- assert 'album' in album.romanized_name
? ---- ^^^^
+ assert 'album' in factory.romanized_name
? + ^^^^^
- assert album.identifier == 'album'
? ---- ^^^^
+ assert factory.identifier == 'album'
? + ^^^^^
@pytest.mark.django_db
- class TestSingles(object):
- def test_single_factory(self):
? ---- ----
+ def test_single_factory():
- single = SingleFactory()
+ factory = SingleFactory()
- assert isinstance(single, Single)
? ---- ^^^^^^
+ assert isinstance(factory, Single)
? ^^^^^^^
- assert 'single' in single.romanized_name
? ---- ^^^^^^
+ assert 'single' in factory.romanized_name
? ^^^^^^^
- assert single.identifier == 'single'
? ---- ^^^^^^
+ assert factory.identifier == 'single'
? ^^^^^^^
| 22 | 0.956522 | 10 | 12 |
58874bbedc4afe8d2fb9318e965513f1b8219441 | DiskSampling/PoissonDiskSampling.h | DiskSampling/PoissonDiskSampling.h |
class PoissonDiskSampling
{
public:
PoissonDiskSampling() = default;
~PoissonDiskSampling() = default;
PoissonDiskSampling(const PoissonDiskSampling& rhs) = default;
PoissonDiskSampling(const PoissonDiskSampling&& rhs) = default;
PoissonDiskSampling& operator=(const PoissonDiskSampling& rhs) = default;
PoissonDiskSampling& operator=(const PoissonDiskSampling&& rhs) = default;
public:
PoissonDiskSampling(int pointWidth, int pointHeight, double pointMinDist, double pointCount);
std::vector<std::pair<double, double>> Generate();
};
#endif |
class PoissonDiskSampling
{
public:
PoissonDiskSampling() = default;
PoissonDiskSampling(int pointWidth, int pointHeight, double pointMinDist, double pointCount);
~PoissonDiskSampling() = default;
PoissonDiskSampling(const PoissonDiskSampling& pds) = delete;
PoissonDiskSampling(PoissonDiskSampling&& pds) = delete;
PoissonDiskSampling& operator=(const PoissonDiskSampling& rhs) = delete;
PoissonDiskSampling& operator=(PoissonDiskSampling&& rhs) = delete;
std::vector<std::pair<double, double>> Generate();
struct Point
{
Point() : x(0.0), y(0.0) { }
Point(double _x, double _y) : x(_x), y(_y) { }
~Point() = default;
Point(const Point& p) : x(p.x), y(p.y) { }
Point(Point&& p) : x(p.x), y(p.y) { }
Point& operator=(const Point& rhs)
{
if (this == &rhs)
{
return;
}
x = rhs.x;
y = rhs.y;
return *this;
}
Point& operator=(Point&& rhs)
{
if (this == &rhs)
{
return;
}
x = rhs.x;
y = rhs.y;
return *this;
}
double x, y;
};
};
#endif | Fix grammer error and implement Point structure | Fix grammer error and implement Point structure
| C | mit | utilForever/PolyMapGenerator | c | ## Code Before:
class PoissonDiskSampling
{
public:
PoissonDiskSampling() = default;
~PoissonDiskSampling() = default;
PoissonDiskSampling(const PoissonDiskSampling& rhs) = default;
PoissonDiskSampling(const PoissonDiskSampling&& rhs) = default;
PoissonDiskSampling& operator=(const PoissonDiskSampling& rhs) = default;
PoissonDiskSampling& operator=(const PoissonDiskSampling&& rhs) = default;
public:
PoissonDiskSampling(int pointWidth, int pointHeight, double pointMinDist, double pointCount);
std::vector<std::pair<double, double>> Generate();
};
#endif
## Instruction:
Fix grammer error and implement Point structure
## Code After:
class PoissonDiskSampling
{
public:
PoissonDiskSampling() = default;
PoissonDiskSampling(int pointWidth, int pointHeight, double pointMinDist, double pointCount);
~PoissonDiskSampling() = default;
PoissonDiskSampling(const PoissonDiskSampling& pds) = delete;
PoissonDiskSampling(PoissonDiskSampling&& pds) = delete;
PoissonDiskSampling& operator=(const PoissonDiskSampling& rhs) = delete;
PoissonDiskSampling& operator=(PoissonDiskSampling&& rhs) = delete;
std::vector<std::pair<double, double>> Generate();
struct Point
{
Point() : x(0.0), y(0.0) { }
Point(double _x, double _y) : x(_x), y(_y) { }
~Point() = default;
Point(const Point& p) : x(p.x), y(p.y) { }
Point(Point&& p) : x(p.x), y(p.y) { }
Point& operator=(const Point& rhs)
{
if (this == &rhs)
{
return;
}
x = rhs.x;
y = rhs.y;
return *this;
}
Point& operator=(Point&& rhs)
{
if (this == &rhs)
{
return;
}
x = rhs.x;
y = rhs.y;
return *this;
}
double x, y;
};
};
#endif |
class PoissonDiskSampling
{
public:
PoissonDiskSampling() = default;
+ PoissonDiskSampling(int pointWidth, int pointHeight, double pointMinDist, double pointCount);
~PoissonDiskSampling() = default;
- PoissonDiskSampling(const PoissonDiskSampling& rhs) = default;
? ^^ ---
+ PoissonDiskSampling(const PoissonDiskSampling& pds) = delete;
? ^^ + +
- PoissonDiskSampling(const PoissonDiskSampling&& rhs) = default;
? ------ ^^ ---
+ PoissonDiskSampling(PoissonDiskSampling&& pds) = delete;
? ^^ + +
- PoissonDiskSampling& operator=(const PoissonDiskSampling& rhs) = default;
? ---
+ PoissonDiskSampling& operator=(const PoissonDiskSampling& rhs) = delete;
? + +
- PoissonDiskSampling& operator=(const PoissonDiskSampling&& rhs) = default;
? ------ ---
+ PoissonDiskSampling& operator=(PoissonDiskSampling&& rhs) = delete;
? + +
-
- public:
- PoissonDiskSampling(int pointWidth, int pointHeight, double pointMinDist, double pointCount);
std::vector<std::pair<double, double>> Generate();
+
+ struct Point
+ {
+ Point() : x(0.0), y(0.0) { }
+ Point(double _x, double _y) : x(_x), y(_y) { }
+
+ ~Point() = default;
+
+ Point(const Point& p) : x(p.x), y(p.y) { }
+ Point(Point&& p) : x(p.x), y(p.y) { }
+
+ Point& operator=(const Point& rhs)
+ {
+ if (this == &rhs)
+ {
+ return;
+ }
+
+ x = rhs.x;
+ y = rhs.y;
+
+ return *this;
+ }
+ Point& operator=(Point&& rhs)
+ {
+ if (this == &rhs)
+ {
+ return;
+ }
+
+ x = rhs.x;
+ y = rhs.y;
+
+ return *this;
+ }
+
+ double x, y;
+ };
};
#endif | 50 | 2.380952 | 43 | 7 |
1d9c48d23d82882c7b40138253e35044ba10cf0e | requirements.txt | requirements.txt | argparse; python_version<'3.2'
defusedxml
ordereddict; python_version<'3.1'
requests-oauthlib>=0.6.1
requests>=2.10.0
requests_toolbelt
setuptools>=20.10.1
six>=1.10.0
| argparse; python_version<'3.2'
defusedxml
ordereddict; python_version<'3.1'
pbr>=3.0.0
requests-oauthlib>=0.6.1
requests>=2.10.0
requests_toolbelt
setuptools>=20.10.1
six>=1.10.0
| Add pbr >= 3.0.0 as a module dependency | Add pbr >= 3.0.0 as a module dependency
Fixes #501.
| Text | bsd-2-clause | dwmarshall/pycontribs-jira,pycontribs/jira,pycontribs/jira,dwmarshall/pycontribs-jira | text | ## Code Before:
argparse; python_version<'3.2'
defusedxml
ordereddict; python_version<'3.1'
requests-oauthlib>=0.6.1
requests>=2.10.0
requests_toolbelt
setuptools>=20.10.1
six>=1.10.0
## Instruction:
Add pbr >= 3.0.0 as a module dependency
Fixes #501.
## Code After:
argparse; python_version<'3.2'
defusedxml
ordereddict; python_version<'3.1'
pbr>=3.0.0
requests-oauthlib>=0.6.1
requests>=2.10.0
requests_toolbelt
setuptools>=20.10.1
six>=1.10.0
| argparse; python_version<'3.2'
defusedxml
ordereddict; python_version<'3.1'
+ pbr>=3.0.0
requests-oauthlib>=0.6.1
requests>=2.10.0
requests_toolbelt
setuptools>=20.10.1
six>=1.10.0 | 1 | 0.125 | 1 | 0 |
81f7a72faef0c82bf746b3a562f8ed7aab67b69a | js/views/sidebar.js | js/views/sidebar.js | ( function( window, Backbone, $, _, Scrivener, undefined ) {
"use strict";
var document = window.document;
Scrivener.Views.Sidebar = Backbone.View.extend( {
className : 'scrivener-customizer-sidebar',
events : {
'click .button.close' : 'onCloseCustomizerClick'
},
initialize : function( attributes ) {
this.render( attributes.ajaxData );
},
render : function( ajaxData ) {
this.$el.html( '<div class="content">' + ajaxData.sidebarHTML + '</div>' );
},
onCloseCustomizerClick : function( event ) {
this.model.closeCustomizer();
},
close : function() {
this.remove();
}
} );
} )( window, Backbone, jQuery, _, Scrivener ); | ( function( window, Backbone, $, _, Scrivener, undefined ) {
"use strict";
var document = window.document;
Scrivener.Views.Sidebar = Backbone.View.extend( {
className : 'scrivener-customizer-sidebar',
events : {
'click .button.close' : 'onCloseCustomizerClick',
'click #set-post-thumbnail' : 'openMediaManager',
'click #remove-post-thumbnail' : 'removePostThumbnail',
},
initialize : function( attributes ) {
this.render( attributes.ajaxData );
},
render : function( ajaxData ) {
this.$el.html( '<div class="content">' + ajaxData.sidebarHTML + '</div>' );
},
onCloseCustomizerClick : function( event ) {
this.model.closeCustomizer();
},
close : function() {
this.remove();
},
openMediaManager: function( event ) {
event.stopPropagation();
event.preventDefault();
wp.media.featuredImage.frame().open();
},
removePostThumbnail: function( event ) {
//wp.media.view.settings.post.featuredImageId = -1;
}
} );
} )( window, Backbone, jQuery, _, Scrivener ); | Add media manager open JS method | Add media manager open JS method
| JavaScript | mit | resarahman/Post-Customizer,chirox/Post-Customizer,10up/Post-Customizer,chirox/Post-Customizer,resarahman/Post-Customizer,resarahman/Post-Customizer,tw2113/Post-Customizer,10up/Post-Customizer,chirox/Post-Customizer,tw2113/Post-Customizer | javascript | ## Code Before:
( function( window, Backbone, $, _, Scrivener, undefined ) {
"use strict";
var document = window.document;
Scrivener.Views.Sidebar = Backbone.View.extend( {
className : 'scrivener-customizer-sidebar',
events : {
'click .button.close' : 'onCloseCustomizerClick'
},
initialize : function( attributes ) {
this.render( attributes.ajaxData );
},
render : function( ajaxData ) {
this.$el.html( '<div class="content">' + ajaxData.sidebarHTML + '</div>' );
},
onCloseCustomizerClick : function( event ) {
this.model.closeCustomizer();
},
close : function() {
this.remove();
}
} );
} )( window, Backbone, jQuery, _, Scrivener );
## Instruction:
Add media manager open JS method
## Code After:
( function( window, Backbone, $, _, Scrivener, undefined ) {
"use strict";
var document = window.document;
Scrivener.Views.Sidebar = Backbone.View.extend( {
className : 'scrivener-customizer-sidebar',
events : {
'click .button.close' : 'onCloseCustomizerClick',
'click #set-post-thumbnail' : 'openMediaManager',
'click #remove-post-thumbnail' : 'removePostThumbnail',
},
initialize : function( attributes ) {
this.render( attributes.ajaxData );
},
render : function( ajaxData ) {
this.$el.html( '<div class="content">' + ajaxData.sidebarHTML + '</div>' );
},
onCloseCustomizerClick : function( event ) {
this.model.closeCustomizer();
},
close : function() {
this.remove();
},
openMediaManager: function( event ) {
event.stopPropagation();
event.preventDefault();
wp.media.featuredImage.frame().open();
},
removePostThumbnail: function( event ) {
//wp.media.view.settings.post.featuredImageId = -1;
}
} );
} )( window, Backbone, jQuery, _, Scrivener ); | ( function( window, Backbone, $, _, Scrivener, undefined ) {
"use strict";
var document = window.document;
Scrivener.Views.Sidebar = Backbone.View.extend( {
className : 'scrivener-customizer-sidebar',
events : {
- 'click .button.close' : 'onCloseCustomizerClick'
+ 'click .button.close' : 'onCloseCustomizerClick',
? +
+ 'click #set-post-thumbnail' : 'openMediaManager',
+ 'click #remove-post-thumbnail' : 'removePostThumbnail',
},
initialize : function( attributes ) {
this.render( attributes.ajaxData );
},
render : function( ajaxData ) {
this.$el.html( '<div class="content">' + ajaxData.sidebarHTML + '</div>' );
},
onCloseCustomizerClick : function( event ) {
this.model.closeCustomizer();
},
close : function() {
this.remove();
+ },
+
+ openMediaManager: function( event ) {
+ event.stopPropagation();
+ event.preventDefault();
+ wp.media.featuredImage.frame().open();
+ },
+
+ removePostThumbnail: function( event ) {
+ //wp.media.view.settings.post.featuredImageId = -1;
}
} );
} )( window, Backbone, jQuery, _, Scrivener ); | 14 | 0.4375 | 13 | 1 |
a811aba5ce25c7b5a75e76739951dc6a448198d1 | third_party/mlir/test/lib/IR/CMakeLists.txt | third_party/mlir/test/lib/IR/CMakeLists.txt | add_llvm_library(MLIRTestIR
TestFunc.cpp
TestSymbolUses.cpp
ADDITIONAL_HEADER_DIRS
)
target_link_libraries(MLIRTestIR
MLIRPass
)
| add_llvm_library(MLIRTestIR
TestFunc.cpp
TestSymbolUses.cpp
ADDITIONAL_HEADER_DIRS
)
include_directories(${CMAKE_CURRENT_SOURCE_DIR}/../TestDialect)
include_directories(${CMAKE_CURRENT_BINARY_DIR}/../TestDialect)
target_link_libraries(MLIRTestIR
MLIRPass
)
| Add include path to the TestDialect to fix broken build. | Add include path to the TestDialect to fix broken build.
PiperOrigin-RevId: 284067891
Change-Id: I4014753aed53b6b90dc4f8bd141e2eceff2b9f49
| Text | apache-2.0 | frreiss/tensorflow-fred,tensorflow/tensorflow-experimental_link_static_libraries_once,jhseu/tensorflow,xzturn/tensorflow,aam-at/tensorflow,xzturn/tensorflow,paolodedios/tensorflow,tensorflow/tensorflow-pywrap_tf_optimizer,petewarden/tensorflow,jhseu/tensorflow,cxxgtxy/tensorflow,renyi533/tensorflow,yongtang/tensorflow,frreiss/tensorflow-fred,aam-at/tensorflow,Intel-Corporation/tensorflow,aldian/tensorflow,davidzchen/tensorflow,annarev/tensorflow,paolodedios/tensorflow,karllessard/tensorflow,petewarden/tensorflow,davidzchen/tensorflow,tensorflow/tensorflow-experimental_link_static_libraries_once,renyi533/tensorflow,Intel-Corporation/tensorflow,freedomtan/tensorflow,Intel-tensorflow/tensorflow,gautam1858/tensorflow,gunan/tensorflow,cxxgtxy/tensorflow,freedomtan/tensorflow,tensorflow/tensorflow-experimental_link_static_libraries_once,gunan/tensorflow,sarvex/tensorflow,yongtang/tensorflow,petewarden/tensorflow,karllessard/tensorflow,gunan/tensorflow,aam-at/tensorflow,gautam1858/tensorflow,gunan/tensorflow,aam-at/tensorflow,tensorflow/tensorflow,tensorflow/tensorflow-experimental_link_static_libraries_once,xzturn/tensorflow,tensorflow/tensorflow-pywrap_tf_optimizer,karllessard/tensorflow,jhseu/tensorflow,tensorflow/tensorflow-pywrap_tf_optimizer,freedomtan/tensorflow,xzturn/tensorflow,gunan/tensorflow,aldian/tensorflow,tensorflow/tensorflow,gautam1858/tensorflow,jhseu/tensorflow,jhseu/tensorflow,petewarden/tensorflow,freedomtan/tensorflow,renyi533/tensorflow,freedomtan/tensorflow,karllessard/tensorflow,Intel-tensorflow/tensorflow,yongtang/tensorflow,tensorflow/tensorflow,renyi533/tensorflow,gunan/tensorflow,aldian/tensorflow,annarev/tensorflow,paolodedios/tensorflow,freedomtan/tensorflow,freedomtan/tensorflow,tensorflow/tensorflow-pywrap_saved_model,tensorflow/tensorflow-pywrap_tf_optimizer,sarvex/tensorflow,frreiss/tensorflow-fred,yongtang/tensorflow,tensorflow/tensorflow-experimental_link_static_libraries_once,karllessard/tensorflow,annarev/tensorflow,tensorflow/tensorflow-experimental_link_static_libraries_once,paolodedios/tensorflow,aam-at/tensorflow,frreiss/tensorflow-fred,sarvex/tensorflow,yongtang/tensorflow,tensorflow/tensorflow-experimental_link_static_libraries_once,frreiss/tensorflow-fred,tensorflow/tensorflow-pywrap_saved_model,gunan/tensorflow,gautam1858/tensorflow,paolodedios/tensorflow,renyi533/tensorflow,frreiss/tensorflow-fred,sarvex/tensorflow,petewarden/tensorflow,davidzchen/tensorflow,gautam1858/tensorflow,tensorflow/tensorflow,xzturn/tensorflow,petewarden/tensorflow,sarvex/tensorflow,petewarden/tensorflow,paolodedios/tensorflow,Intel-tensorflow/tensorflow,karllessard/tensorflow,xzturn/tensorflow,tensorflow/tensorflow-pywrap_tf_optimizer,Intel-tensorflow/tensorflow,davidzchen/tensorflow,aldian/tensorflow,jhseu/tensorflow,karllessard/tensorflow,karllessard/tensorflow,tensorflow/tensorflow-pywrap_saved_model,xzturn/tensorflow,tensorflow/tensorflow-pywrap_tf_optimizer,frreiss/tensorflow-fred,tensorflow/tensorflow,Intel-Corporation/tensorflow,jhseu/tensorflow,cxxgtxy/tensorflow,davidzchen/tensorflow,aam-at/tensorflow,annarev/tensorflow,freedomtan/tensorflow,davidzchen/tensorflow,aldian/tensorflow,gunan/tensorflow,paolodedios/tensorflow,karllessard/tensorflow,aam-at/tensorflow,tensorflow/tensorflow-pywrap_tf_optimizer,jhseu/tensorflow,Intel-Corporation/tensorflow,tensorflow/tensorflow-pywrap_saved_model,freedomtan/tensorflow,davidzchen/tensorflow,petewarden/tensorflow,gautam1858/tensorflow,aam-at/tensorflow,renyi533/tensorflow,renyi533/tensorflow,frreiss/tensorflow-fred,annarev/tensorflow,petewarden/tensorflow,yongtang/tensorflow,tensorflow/tensorflow,tensorflow/tensorflow-pywrap_saved_model,paolodedios/tensorflow,paolodedios/tensorflow,Intel-tensorflow/tensorflow,xzturn/tensorflow,renyi533/tensorflow,jhseu/tensorflow,tensorflow/tensorflow-pywrap_tf_optimizer,annarev/tensorflow,tensorflow/tensorflow-pywrap_saved_model,tensorflow/tensorflow-pywrap_saved_model,tensorflow/tensorflow-experimental_link_static_libraries_once,annarev/tensorflow,freedomtan/tensorflow,tensorflow/tensorflow-experimental_link_static_libraries_once,jhseu/tensorflow,davidzchen/tensorflow,gunan/tensorflow,yongtang/tensorflow,tensorflow/tensorflow,Intel-Corporation/tensorflow,jhseu/tensorflow,petewarden/tensorflow,petewarden/tensorflow,yongtang/tensorflow,Intel-tensorflow/tensorflow,freedomtan/tensorflow,davidzchen/tensorflow,xzturn/tensorflow,davidzchen/tensorflow,xzturn/tensorflow,xzturn/tensorflow,tensorflow/tensorflow-pywrap_saved_model,aldian/tensorflow,sarvex/tensorflow,Intel-Corporation/tensorflow,annarev/tensorflow,jhseu/tensorflow,aam-at/tensorflow,Intel-Corporation/tensorflow,xzturn/tensorflow,cxxgtxy/tensorflow,paolodedios/tensorflow,tensorflow/tensorflow-experimental_link_static_libraries_once,Intel-Corporation/tensorflow,cxxgtxy/tensorflow,tensorflow/tensorflow-experimental_link_static_libraries_once,renyi533/tensorflow,karllessard/tensorflow,aam-at/tensorflow,gunan/tensorflow,aldian/tensorflow,Intel-tensorflow/tensorflow,frreiss/tensorflow-fred,davidzchen/tensorflow,aam-at/tensorflow,renyi533/tensorflow,gautam1858/tensorflow,gautam1858/tensorflow,yongtang/tensorflow,gautam1858/tensorflow,gautam1858/tensorflow,aldian/tensorflow,renyi533/tensorflow,tensorflow/tensorflow-pywrap_saved_model,yongtang/tensorflow,tensorflow/tensorflow,tensorflow/tensorflow-pywrap_tf_optimizer,annarev/tensorflow,Intel-tensorflow/tensorflow,freedomtan/tensorflow,cxxgtxy/tensorflow,sarvex/tensorflow,cxxgtxy/tensorflow,Intel-tensorflow/tensorflow,tensorflow/tensorflow,annarev/tensorflow,gunan/tensorflow,Intel-tensorflow/tensorflow,petewarden/tensorflow,tensorflow/tensorflow-pywrap_tf_optimizer,frreiss/tensorflow-fred,gautam1858/tensorflow,tensorflow/tensorflow-pywrap_tf_optimizer,Intel-tensorflow/tensorflow,renyi533/tensorflow,frreiss/tensorflow-fred,tensorflow/tensorflow-pywrap_saved_model,davidzchen/tensorflow,karllessard/tensorflow,sarvex/tensorflow,paolodedios/tensorflow,cxxgtxy/tensorflow,tensorflow/tensorflow,tensorflow/tensorflow,frreiss/tensorflow-fred,tensorflow/tensorflow-pywrap_saved_model,yongtang/tensorflow,gunan/tensorflow,annarev/tensorflow,aam-at/tensorflow,gautam1858/tensorflow | text | ## Code Before:
add_llvm_library(MLIRTestIR
TestFunc.cpp
TestSymbolUses.cpp
ADDITIONAL_HEADER_DIRS
)
target_link_libraries(MLIRTestIR
MLIRPass
)
## Instruction:
Add include path to the TestDialect to fix broken build.
PiperOrigin-RevId: 284067891
Change-Id: I4014753aed53b6b90dc4f8bd141e2eceff2b9f49
## Code After:
add_llvm_library(MLIRTestIR
TestFunc.cpp
TestSymbolUses.cpp
ADDITIONAL_HEADER_DIRS
)
include_directories(${CMAKE_CURRENT_SOURCE_DIR}/../TestDialect)
include_directories(${CMAKE_CURRENT_BINARY_DIR}/../TestDialect)
target_link_libraries(MLIRTestIR
MLIRPass
)
| add_llvm_library(MLIRTestIR
TestFunc.cpp
TestSymbolUses.cpp
ADDITIONAL_HEADER_DIRS
)
+ include_directories(${CMAKE_CURRENT_SOURCE_DIR}/../TestDialect)
+ include_directories(${CMAKE_CURRENT_BINARY_DIR}/../TestDialect)
target_link_libraries(MLIRTestIR
MLIRPass
) | 2 | 0.222222 | 2 | 0 |
f3090ee1adce9109b968423b67916669a29983e7 | src/Item.php | src/Item.php | <?php
/**
* Box packing (3D bin packing, knapsack problem).
*
* @author Doug Wright
*/
declare(strict_types=1);
namespace DVDoug\BoxPacker;
/**
* An item to be packed.
*
* @author Doug Wright
*/
interface Item
{
public const ROTATION_NEVER = 1;
public const ROTATION_KEEP_FLAT = 2;
public const ROTATION_BEST_FIT = 6;
/**
* Item SKU etc.
*/
public function getDescription(): string;
/**
* Item width in mm.
*/
public function getWidth(): int;
/**
* Item length in mm.
*/
public function getLength(): int;
/**
* Item depth in mm.
*/
public function getDepth(): int;
/**
* Item weight in g.
*/
public function getWeight(): int;
/**
* Possible item rotations allowed. One of the ROTATION_* constants.
*/
public function getAllowedRotations(): int;
}
| <?php
/**
* Box packing (3D bin packing, knapsack problem).
*
* @author Doug Wright
*/
declare(strict_types=1);
namespace DVDoug\BoxPacker;
/**
* An item to be packed.
*
* @author Doug Wright
*/
interface Item
{
/** @var int must be placed in it's defined orientation only */
public const ROTATION_NEVER = 1;
/** @var int can be turned sideways 90°, but cannot be placed *on* it's side e.g. fragile "↑this way up" items */
public const ROTATION_KEEP_FLAT = 2;
/** @var int no handling restrictions, item can be placed in any orientation */
public const ROTATION_BEST_FIT = 6;
/**
* Item SKU etc.
*/
public function getDescription(): string;
/**
* Item width in mm.
*/
public function getWidth(): int;
/**
* Item length in mm.
*/
public function getLength(): int;
/**
* Item depth in mm.
*/
public function getDepth(): int;
/**
* Item weight in g.
*/
public function getWeight(): int;
/**
* Possible item rotations allowed. One of the ROTATION_* constants.
*/
public function getAllowedRotations(): int;
}
| Add descriptions for the ROTATION_* constants | Add descriptions for the ROTATION_* constants
| PHP | mit | dvdoug/BoxPacker | php | ## Code Before:
<?php
/**
* Box packing (3D bin packing, knapsack problem).
*
* @author Doug Wright
*/
declare(strict_types=1);
namespace DVDoug\BoxPacker;
/**
* An item to be packed.
*
* @author Doug Wright
*/
interface Item
{
public const ROTATION_NEVER = 1;
public const ROTATION_KEEP_FLAT = 2;
public const ROTATION_BEST_FIT = 6;
/**
* Item SKU etc.
*/
public function getDescription(): string;
/**
* Item width in mm.
*/
public function getWidth(): int;
/**
* Item length in mm.
*/
public function getLength(): int;
/**
* Item depth in mm.
*/
public function getDepth(): int;
/**
* Item weight in g.
*/
public function getWeight(): int;
/**
* Possible item rotations allowed. One of the ROTATION_* constants.
*/
public function getAllowedRotations(): int;
}
## Instruction:
Add descriptions for the ROTATION_* constants
## Code After:
<?php
/**
* Box packing (3D bin packing, knapsack problem).
*
* @author Doug Wright
*/
declare(strict_types=1);
namespace DVDoug\BoxPacker;
/**
* An item to be packed.
*
* @author Doug Wright
*/
interface Item
{
/** @var int must be placed in it's defined orientation only */
public const ROTATION_NEVER = 1;
/** @var int can be turned sideways 90°, but cannot be placed *on* it's side e.g. fragile "↑this way up" items */
public const ROTATION_KEEP_FLAT = 2;
/** @var int no handling restrictions, item can be placed in any orientation */
public const ROTATION_BEST_FIT = 6;
/**
* Item SKU etc.
*/
public function getDescription(): string;
/**
* Item width in mm.
*/
public function getWidth(): int;
/**
* Item length in mm.
*/
public function getLength(): int;
/**
* Item depth in mm.
*/
public function getDepth(): int;
/**
* Item weight in g.
*/
public function getWeight(): int;
/**
* Possible item rotations allowed. One of the ROTATION_* constants.
*/
public function getAllowedRotations(): int;
}
| <?php
/**
* Box packing (3D bin packing, knapsack problem).
*
* @author Doug Wright
*/
declare(strict_types=1);
namespace DVDoug\BoxPacker;
/**
* An item to be packed.
*
* @author Doug Wright
*/
interface Item
{
+ /** @var int must be placed in it's defined orientation only */
public const ROTATION_NEVER = 1;
+ /** @var int can be turned sideways 90°, but cannot be placed *on* it's side e.g. fragile "↑this way up" items */
public const ROTATION_KEEP_FLAT = 2;
+ /** @var int no handling restrictions, item can be placed in any orientation */
public const ROTATION_BEST_FIT = 6;
/**
* Item SKU etc.
*/
public function getDescription(): string;
/**
* Item width in mm.
*/
public function getWidth(): int;
/**
* Item length in mm.
*/
public function getLength(): int;
/**
* Item depth in mm.
*/
public function getDepth(): int;
/**
* Item weight in g.
*/
public function getWeight(): int;
/**
* Possible item rotations allowed. One of the ROTATION_* constants.
*/
public function getAllowedRotations(): int;
} | 3 | 0.056604 | 3 | 0 |
6d079581814faa030c0e73a3c3622bdb0f7eeab2 | ansible/playbooks/apps/stack-data.yaml | ansible/playbooks/apps/stack-data.yaml | ---
- name: AEM Author CloudFormation Stack
hosts: all
gather_facts: no
connection: local
tasks:
- name: Create random credentials for system users
system_users_credentials:
register: credentials
tags:
- create
- name: Ensure data bucket exists
s3_bucket:
name: "{{ s3.data_bucket_name }}"
state: present
tags:
- create
- name: Upload stack init script
s3:
bucket: "{{ s3.data_bucket_name }}"
object: "{{ stack_prefix }}/stack-init.sh"
src: ../../../scripts/stack-init.sh
mode: put
tags:
- create
- name: Upload stack provisioner
s3:
bucket: "{{ s3.data_bucket_name }}"
object: "{{ stack_prefix }}/aem-aws-stack-provisioner.tar.gz"
src: ../../../stage/aem-aws-stack-provisioner.tar.gz
mode: put
tags:
- create
- name: Delete stack data
s3:
bucket: "{{ s3.data_bucket_name }}"
object: "{{ stack_prefix }}"
mode: delobj
tags:
- delete
| ---
- name: AEM Author CloudFormation Stack
hosts: all
gather_facts: no
connection: local
tasks:
- name: Generate random credentials for system users
system_users_credentials:
register: credentials
tags:
- create
- name: Ensure stage directory exists
file:
path: ../../../stage
state: directory
mode: 0755
tags:
- create
- name: Create temporary file containing generated credentials
copy:
content: "{{ credentials.meta }}"
dest: ../../../stage/system_users_credentials.json
tags:
- create
- name: Ensure data bucket exists
s3_bucket:
name: "{{ s3.data_bucket_name }}"
state: present
tags:
- create
- name: Upload stack init script
s3:
bucket: "{{ s3.data_bucket_name }}"
object: "{{ stack_prefix }}/stack-init.sh"
src: ../../../scripts/stack-init.sh
mode: put
tags:
- create
- name: Upload stack provisioner
s3:
bucket: "{{ s3.data_bucket_name }}"
object: "{{ stack_prefix }}/aem-aws-stack-provisioner.tar.gz"
src: ../../../stage/aem-aws-stack-provisioner.tar.gz
mode: put
tags:
- create
- name: Upload system user credentials
s3:
bucket: "{{ s3.data_bucket_name }}"
object: "{{ stack_prefix }}/system-users-credentials.json"
src: ../../../stage/system-users-credentials.json
mode: put
encrypt: yes
tags:
- create
- name: Delete stack data
s3:
bucket: "{{ s3.data_bucket_name }}"
object: "{{ stack_prefix }}"
mode: delobj
tags:
- delete
| Add system user credentials upload to S3. | Add system user credentials upload to S3.
| YAML | apache-2.0 | shinesolutions/aem-aws-stack-builder,shinesolutions/aem-aws-stack-builder | yaml | ## Code Before:
---
- name: AEM Author CloudFormation Stack
hosts: all
gather_facts: no
connection: local
tasks:
- name: Create random credentials for system users
system_users_credentials:
register: credentials
tags:
- create
- name: Ensure data bucket exists
s3_bucket:
name: "{{ s3.data_bucket_name }}"
state: present
tags:
- create
- name: Upload stack init script
s3:
bucket: "{{ s3.data_bucket_name }}"
object: "{{ stack_prefix }}/stack-init.sh"
src: ../../../scripts/stack-init.sh
mode: put
tags:
- create
- name: Upload stack provisioner
s3:
bucket: "{{ s3.data_bucket_name }}"
object: "{{ stack_prefix }}/aem-aws-stack-provisioner.tar.gz"
src: ../../../stage/aem-aws-stack-provisioner.tar.gz
mode: put
tags:
- create
- name: Delete stack data
s3:
bucket: "{{ s3.data_bucket_name }}"
object: "{{ stack_prefix }}"
mode: delobj
tags:
- delete
## Instruction:
Add system user credentials upload to S3.
## Code After:
---
- name: AEM Author CloudFormation Stack
hosts: all
gather_facts: no
connection: local
tasks:
- name: Generate random credentials for system users
system_users_credentials:
register: credentials
tags:
- create
- name: Ensure stage directory exists
file:
path: ../../../stage
state: directory
mode: 0755
tags:
- create
- name: Create temporary file containing generated credentials
copy:
content: "{{ credentials.meta }}"
dest: ../../../stage/system_users_credentials.json
tags:
- create
- name: Ensure data bucket exists
s3_bucket:
name: "{{ s3.data_bucket_name }}"
state: present
tags:
- create
- name: Upload stack init script
s3:
bucket: "{{ s3.data_bucket_name }}"
object: "{{ stack_prefix }}/stack-init.sh"
src: ../../../scripts/stack-init.sh
mode: put
tags:
- create
- name: Upload stack provisioner
s3:
bucket: "{{ s3.data_bucket_name }}"
object: "{{ stack_prefix }}/aem-aws-stack-provisioner.tar.gz"
src: ../../../stage/aem-aws-stack-provisioner.tar.gz
mode: put
tags:
- create
- name: Upload system user credentials
s3:
bucket: "{{ s3.data_bucket_name }}"
object: "{{ stack_prefix }}/system-users-credentials.json"
src: ../../../stage/system-users-credentials.json
mode: put
encrypt: yes
tags:
- create
- name: Delete stack data
s3:
bucket: "{{ s3.data_bucket_name }}"
object: "{{ stack_prefix }}"
mode: delobj
tags:
- delete
| ---
- name: AEM Author CloudFormation Stack
hosts: all
gather_facts: no
connection: local
tasks:
- - name: Create random credentials for system users
? ^ -
+ - name: Generate random credentials for system users
? ^^^^
system_users_credentials:
register: credentials
+ tags:
+ - create
+
+ - name: Ensure stage directory exists
+ file:
+ path: ../../../stage
+ state: directory
+ mode: 0755
+ tags:
+ - create
+
+ - name: Create temporary file containing generated credentials
+ copy:
+ content: "{{ credentials.meta }}"
+ dest: ../../../stage/system_users_credentials.json
tags:
- create
- name: Ensure data bucket exists
s3_bucket:
name: "{{ s3.data_bucket_name }}"
state: present
tags:
- create
- name: Upload stack init script
s3:
bucket: "{{ s3.data_bucket_name }}"
object: "{{ stack_prefix }}/stack-init.sh"
src: ../../../scripts/stack-init.sh
mode: put
tags:
- create
- name: Upload stack provisioner
s3:
bucket: "{{ s3.data_bucket_name }}"
object: "{{ stack_prefix }}/aem-aws-stack-provisioner.tar.gz"
src: ../../../stage/aem-aws-stack-provisioner.tar.gz
mode: put
tags:
- create
+ - name: Upload system user credentials
+ s3:
+ bucket: "{{ s3.data_bucket_name }}"
+ object: "{{ stack_prefix }}/system-users-credentials.json"
+ src: ../../../stage/system-users-credentials.json
+ mode: put
+ encrypt: yes
+ tags:
+ - create
+
- name: Delete stack data
s3:
bucket: "{{ s3.data_bucket_name }}"
object: "{{ stack_prefix }}"
mode: delobj
tags:
- delete | 27 | 0.586957 | 26 | 1 |
51ad5db13df68765fecfdbacb0fa3dcf9e81014d | zsh/starship.toml | zsh/starship.toml | [cmd_duration]
disabled = true
| [cmd_duration]
disabled = true
[aws]
symbol = " "
[conda]
symbol = " "
[dart]
symbol = " "
[directory]
read_only = " "
[docker_context]
symbol = " "
[elixir]
symbol = " "
[elm]
symbol = " "
[git_branch]
symbol = " "
[golang]
symbol = " "
[hg_branch]
symbol = " "
[java]
symbol = " "
[julia]
symbol = " "
[memory_usage]
symbol = " "
[nim]
symbol = " "
[nix_shell]
symbol = " "
[package]
symbol = " "
[perl]
symbol = " "
[php]
symbol = " "
[python]
symbol = " "
[ruby]
symbol = " "
[rust]
symbol = " "
[scala]
symbol = " "
[shlvl]
symbol = " "
[swift]
symbol = "ﯣ "
| Use nerd font glyphs in prompt | Use nerd font glyphs in prompt
Use nerd font glyphs in prompt.
| TOML | mit | routeaccess/dotfiles,nevinvalsaraj/dotfiles,nevinvalsaraj/dotfiles,routeaccess/dotfiles,routeaccess/dotfiles,nevinvalsaraj/dotfiles | toml | ## Code Before:
[cmd_duration]
disabled = true
## Instruction:
Use nerd font glyphs in prompt
Use nerd font glyphs in prompt.
## Code After:
[cmd_duration]
disabled = true
[aws]
symbol = " "
[conda]
symbol = " "
[dart]
symbol = " "
[directory]
read_only = " "
[docker_context]
symbol = " "
[elixir]
symbol = " "
[elm]
symbol = " "
[git_branch]
symbol = " "
[golang]
symbol = " "
[hg_branch]
symbol = " "
[java]
symbol = " "
[julia]
symbol = " "
[memory_usage]
symbol = " "
[nim]
symbol = " "
[nix_shell]
symbol = " "
[package]
symbol = " "
[perl]
symbol = " "
[php]
symbol = " "
[python]
symbol = " "
[ruby]
symbol = " "
[rust]
symbol = " "
[scala]
symbol = " "
[shlvl]
symbol = " "
[swift]
symbol = "ﯣ "
| [cmd_duration]
disabled = true
+
+ [aws]
+ symbol = " "
+
+ [conda]
+ symbol = " "
+
+ [dart]
+ symbol = " "
+
+ [directory]
+ read_only = " "
+
+ [docker_context]
+ symbol = " "
+
+ [elixir]
+ symbol = " "
+
+ [elm]
+ symbol = " "
+
+ [git_branch]
+ symbol = " "
+
+ [golang]
+ symbol = " "
+
+ [hg_branch]
+ symbol = " "
+
+ [java]
+ symbol = " "
+
+ [julia]
+ symbol = " "
+
+ [memory_usage]
+ symbol = " "
+
+ [nim]
+ symbol = " "
+
+ [nix_shell]
+ symbol = " "
+
+ [package]
+ symbol = " "
+
+ [perl]
+ symbol = " "
+
+ [php]
+ symbol = " "
+
+ [python]
+ symbol = " "
+
+ [ruby]
+ symbol = " "
+
+ [rust]
+ symbol = " "
+
+ [scala]
+ symbol = " "
+
+ [shlvl]
+ symbol = " "
+
+ [swift]
+ symbol = "ﯣ "
+ | 73 | 36.5 | 73 | 0 |
4e71c95c4195c8f1fa402cf9283f09e709865d6a | lib/smart_answer_flows/state-pension-age/state_pension_age.govspeak.erb | lib/smart_answer_flows/state-pension-age/state_pension_age.govspeak.erb | <% content_for :title do %>
State Pension calculator
<% end %>
<% content_for :meta_description do %>
Work out your State Pension age and Pension Credit qualifying age or estimate how much basic State Pension you may get
<% end %>
<% content_for :body do %>
Use the calculator to:
* work out when you’ll reach State Pension age
* work out your Pension Credit qualifying age
* find out when you’ll be eligible for free bus travel
##If you’re under 55
The calculator also gives an estimate of your basic State Pension under the current rules - not the [new State Pension](/new-state-pension) rules.
It doesn’t estimate any [Additional State Pension](/additional-state-pension).
##If you’re 55 or over
Don’t use the calculator to work out your State Pension. Get a [State Pension Statement](/state-pension-statement) instead.
<% end %>
<% content_for :post_body do %>
##What you need
You need to enter the number of years you:
* worked and paid National Insurance
* got unemployment, sickness or disability benefits
* claimed Child Benefit or were a carer or foster carer
Don’t count the current tax year. Count tax years from 6 April to 5 April and don’t count any years twice, eg when you were working and getting benefits.
This calculator uses a simplified calculation based on the current law. It can’t take into account every circumstance that might affect you.
^Don’t use the calculator to make future financial decisions.^
<% end %>
| <% content_for :title do %>
Check your State Pension age
<% end %>
<% content_for :meta_description do %>
Work out your State Pension age and Pension Credit qualifying age
<% end %>
<% content_for :body do %>
Check when you’ll reach State Pension age, Pension Credit qualifying age.
You can also find out when you’ll be eligible for free bus travel.
<% end %>
| Update state-pension-age flow start page | Update state-pension-age flow start page
This commit updates the start page to show that state-pension-age calculates your pension start age & date, and your bus pass qualification age.
| HTML+ERB | mit | stwalsh/smart-answers,alphagov/smart-answers,aledelcueto/smart-answers,alphagov/smart-answers,stwalsh/smart-answers,alphagov/smart-answers,alphagov/smart-answers,aledelcueto/smart-answers,aledelcueto/smart-answers,stwalsh/smart-answers,aledelcueto/smart-answers,stwalsh/smart-answers | html+erb | ## Code Before:
<% content_for :title do %>
State Pension calculator
<% end %>
<% content_for :meta_description do %>
Work out your State Pension age and Pension Credit qualifying age or estimate how much basic State Pension you may get
<% end %>
<% content_for :body do %>
Use the calculator to:
* work out when you’ll reach State Pension age
* work out your Pension Credit qualifying age
* find out when you’ll be eligible for free bus travel
##If you’re under 55
The calculator also gives an estimate of your basic State Pension under the current rules - not the [new State Pension](/new-state-pension) rules.
It doesn’t estimate any [Additional State Pension](/additional-state-pension).
##If you’re 55 or over
Don’t use the calculator to work out your State Pension. Get a [State Pension Statement](/state-pension-statement) instead.
<% end %>
<% content_for :post_body do %>
##What you need
You need to enter the number of years you:
* worked and paid National Insurance
* got unemployment, sickness or disability benefits
* claimed Child Benefit or were a carer or foster carer
Don’t count the current tax year. Count tax years from 6 April to 5 April and don’t count any years twice, eg when you were working and getting benefits.
This calculator uses a simplified calculation based on the current law. It can’t take into account every circumstance that might affect you.
^Don’t use the calculator to make future financial decisions.^
<% end %>
## Instruction:
Update state-pension-age flow start page
This commit updates the start page to show that state-pension-age calculates your pension start age & date, and your bus pass qualification age.
## Code After:
<% content_for :title do %>
Check your State Pension age
<% end %>
<% content_for :meta_description do %>
Work out your State Pension age and Pension Credit qualifying age
<% end %>
<% content_for :body do %>
Check when you’ll reach State Pension age, Pension Credit qualifying age.
You can also find out when you’ll be eligible for free bus travel.
<% end %>
| <% content_for :title do %>
- State Pension calculator
+ Check your State Pension age
<% end %>
<% content_for :meta_description do %>
- Work out your State Pension age and Pension Credit qualifying age or estimate how much basic State Pension you may get
+ Work out your State Pension age and Pension Credit qualifying age
<% end %>
<% content_for :body do %>
- Use the calculator to:
+ Check when you’ll reach State Pension age, Pension Credit qualifying age.
- * work out when you’ll reach State Pension age
- * work out your Pension Credit qualifying age
- * find out when you’ll be eligible for free bus travel
? ^
+ You can also find out when you’ll be eligible for free bus travel.
? ^^^^^^^^^^^^ +
-
- ##If you’re under 55
-
- The calculator also gives an estimate of your basic State Pension under the current rules - not the [new State Pension](/new-state-pension) rules.
-
- It doesn’t estimate any [Additional State Pension](/additional-state-pension).
-
- ##If you’re 55 or over
-
- Don’t use the calculator to work out your State Pension. Get a [State Pension Statement](/state-pension-statement) instead.
<% end %>
-
- <% content_for :post_body do %>
- ##What you need
-
- You need to enter the number of years you:
-
- * worked and paid National Insurance
- * got unemployment, sickness or disability benefits
- * claimed Child Benefit or were a carer or foster carer
-
- Don’t count the current tax year. Count tax years from 6 April to 5 April and don’t count any years twice, eg when you were working and getting benefits.
-
- This calculator uses a simplified calculation based on the current law. It can’t take into account every circumstance that might affect you.
-
- ^Don’t use the calculator to make future financial decisions.^
-
- <% end %> | 37 | 0.860465 | 4 | 33 |
df565dab5c8c23c952646a02732abcb3f26431e8 | app/controllers/application_controller.rb | app/controllers/application_controller.rb | class ApplicationController < ActionController::Base
# Prevent CSRF attacks by raising an exception.
# For APIs, you may want to use :null_session instead.
protect_from_forgery with: :exception
end
| class ApplicationController < ActionController::Base
# Prevent CSRF attacks by raising an exception.
# For APIs, you may want to use :null_session instead.
protect_from_forgery with: :exception
rescue_from ActiveRecord::RecordNotFound do
respond_to do |type|
type.all {render :nothing => true, :status => 404}
end
end
end
| Add rescue from record not found clause | Add rescue from record not found clause
This will make the app return 404 when user tries
to act on non-existing id
| Ruby | mit | morynicz/Ippon,morynicz/tournament-service,morynicz/tournament-service,morynicz/Ippon,morynicz/Ippon,morynicz/tournament-service | ruby | ## Code Before:
class ApplicationController < ActionController::Base
# Prevent CSRF attacks by raising an exception.
# For APIs, you may want to use :null_session instead.
protect_from_forgery with: :exception
end
## Instruction:
Add rescue from record not found clause
This will make the app return 404 when user tries
to act on non-existing id
## Code After:
class ApplicationController < ActionController::Base
# Prevent CSRF attacks by raising an exception.
# For APIs, you may want to use :null_session instead.
protect_from_forgery with: :exception
rescue_from ActiveRecord::RecordNotFound do
respond_to do |type|
type.all {render :nothing => true, :status => 404}
end
end
end
| class ApplicationController < ActionController::Base
# Prevent CSRF attacks by raising an exception.
# For APIs, you may want to use :null_session instead.
protect_from_forgery with: :exception
+
+ rescue_from ActiveRecord::RecordNotFound do
+ respond_to do |type|
+ type.all {render :nothing => true, :status => 404}
+ end
+ end
end | 6 | 1.2 | 6 | 0 |
83c9e422efd18ad16d4400c5e58a168e43c2ddb5 | .travis.yml | .travis.yml | language: android
sudo: required
jdk:
- oraclejdk8
android:
components:
- tools
- build-tools-23.0.2
- android-23
- extra
# https://github.com/travis-ci/travis-ci/issues/3259
script:
- sudo apt-get update && sudo apt-get install oracle-java8-installer
- java -version | language: android
sudo: required
dist: trusty # https://github.com/travis-ci/travis-ci/issues/3259
jdk:
- oraclejdk8
android:
components:
- tools
- build-tools-23.0.2
- android-23
- extra | Use trusty dist for updated Java 8 version | Use trusty dist for updated Java 8 version
| YAML | apache-2.0 | sakuna63/requery,requery/requery,sakuna63/requery,requery/requery | yaml | ## Code Before:
language: android
sudo: required
jdk:
- oraclejdk8
android:
components:
- tools
- build-tools-23.0.2
- android-23
- extra
# https://github.com/travis-ci/travis-ci/issues/3259
script:
- sudo apt-get update && sudo apt-get install oracle-java8-installer
- java -version
## Instruction:
Use trusty dist for updated Java 8 version
## Code After:
language: android
sudo: required
dist: trusty # https://github.com/travis-ci/travis-ci/issues/3259
jdk:
- oraclejdk8
android:
components:
- tools
- build-tools-23.0.2
- android-23
- extra | language: android
sudo: required
+ dist: trusty # https://github.com/travis-ci/travis-ci/issues/3259
jdk:
- oraclejdk8
android:
components:
- tools
- build-tools-23.0.2
- android-23
- extra
- # https://github.com/travis-ci/travis-ci/issues/3259
- script:
- - sudo apt-get update && sudo apt-get install oracle-java8-installer
- - java -version | 5 | 0.357143 | 1 | 4 |
7a3c2e8a1a5c1e4da70f8351e89a6690af4f7d2f | .circleci/config.yml | .circleci/config.yml | ---
version: 2
jobs:
serverTest:
docker:
- image: isic/isic_test:latest
- image: circleci/mongo:3.6-ram
command: ["mongod", "--storageEngine", "ephemeralForTest"]
working_directory: /home/circleci/project
steps:
- checkout:
path: /home/circleci/project
- run:
name: Run Tox
command: tox
webBuild_webTest:
docker:
- image: isic/isic_test:latest
working_directory: /home/circleci/project
steps:
- checkout:
path: /home/circleci/project
- restore_cache:
key: yarn-{{ arch }}-{{ checksum "isic-archive-gui/yarn.lock" }}
- run:
name: Install ISIC GUI dependencies
command: yarn install
working_directory: isic-archive-gui
- run:
name: Build ISIC Admin GUI
command: yarn run build
working_directory: isic-archive-gui
- run:
name: Build ISIC Integration GUI
command: yarn run build:integration
working_directory: isic-archive-gui
- run:
name: Lint ISIC GUI
command: yarn run lint --no-fix
working_directory: isic-archive-gui
- save_cache:
paths: /home/circleci/.cache/yarn
key: yarn-{{ arch }}-{{ checksum "isic-archive-gui/yarn.lock" }}
workflows:
version: 2
test_all:
jobs:
- serverTest
- webBuild_webTest
| ---
version: 2
jobs:
serverTest:
docker:
- image: isic/isic_test:latest
- image: circleci/mongo:3.6-ram
command: ["mongod", "--storageEngine", "ephemeralForTest", "--dbpath", "/dev/shm/mongo"]
working_directory: /home/circleci/project
steps:
- checkout:
path: /home/circleci/project
- run:
name: Run Tox
command: tox
webBuild_webTest:
docker:
- image: isic/isic_test:latest
working_directory: /home/circleci/project
steps:
- checkout:
path: /home/circleci/project
- restore_cache:
key: yarn-{{ arch }}-{{ checksum "isic-archive-gui/yarn.lock" }}
- run:
name: Install ISIC GUI dependencies
command: yarn install
working_directory: isic-archive-gui
- run:
name: Build ISIC Admin GUI
command: yarn run build
working_directory: isic-archive-gui
- run:
name: Build ISIC Integration GUI
command: yarn run build:integration
working_directory: isic-archive-gui
- run:
name: Lint ISIC GUI
command: yarn run lint --no-fix
working_directory: isic-archive-gui
- save_cache:
paths: /home/circleci/.cache/yarn
key: yarn-{{ arch }}-{{ checksum "isic-archive-gui/yarn.lock" }}
workflows:
version: 2
test_all:
jobs:
- serverTest
- webBuild_webTest
| Revert "Remove unnecessary mongo flag for CI" | Revert "Remove unnecessary mongo flag for CI"
| YAML | apache-2.0 | ImageMarkup/isic-archive,ImageMarkup/isic-archive,ImageMarkup/isic-archive,ImageMarkup/isic-archive | yaml | ## Code Before:
---
version: 2
jobs:
serverTest:
docker:
- image: isic/isic_test:latest
- image: circleci/mongo:3.6-ram
command: ["mongod", "--storageEngine", "ephemeralForTest"]
working_directory: /home/circleci/project
steps:
- checkout:
path: /home/circleci/project
- run:
name: Run Tox
command: tox
webBuild_webTest:
docker:
- image: isic/isic_test:latest
working_directory: /home/circleci/project
steps:
- checkout:
path: /home/circleci/project
- restore_cache:
key: yarn-{{ arch }}-{{ checksum "isic-archive-gui/yarn.lock" }}
- run:
name: Install ISIC GUI dependencies
command: yarn install
working_directory: isic-archive-gui
- run:
name: Build ISIC Admin GUI
command: yarn run build
working_directory: isic-archive-gui
- run:
name: Build ISIC Integration GUI
command: yarn run build:integration
working_directory: isic-archive-gui
- run:
name: Lint ISIC GUI
command: yarn run lint --no-fix
working_directory: isic-archive-gui
- save_cache:
paths: /home/circleci/.cache/yarn
key: yarn-{{ arch }}-{{ checksum "isic-archive-gui/yarn.lock" }}
workflows:
version: 2
test_all:
jobs:
- serverTest
- webBuild_webTest
## Instruction:
Revert "Remove unnecessary mongo flag for CI"
## Code After:
---
version: 2
jobs:
serverTest:
docker:
- image: isic/isic_test:latest
- image: circleci/mongo:3.6-ram
command: ["mongod", "--storageEngine", "ephemeralForTest", "--dbpath", "/dev/shm/mongo"]
working_directory: /home/circleci/project
steps:
- checkout:
path: /home/circleci/project
- run:
name: Run Tox
command: tox
webBuild_webTest:
docker:
- image: isic/isic_test:latest
working_directory: /home/circleci/project
steps:
- checkout:
path: /home/circleci/project
- restore_cache:
key: yarn-{{ arch }}-{{ checksum "isic-archive-gui/yarn.lock" }}
- run:
name: Install ISIC GUI dependencies
command: yarn install
working_directory: isic-archive-gui
- run:
name: Build ISIC Admin GUI
command: yarn run build
working_directory: isic-archive-gui
- run:
name: Build ISIC Integration GUI
command: yarn run build:integration
working_directory: isic-archive-gui
- run:
name: Lint ISIC GUI
command: yarn run lint --no-fix
working_directory: isic-archive-gui
- save_cache:
paths: /home/circleci/.cache/yarn
key: yarn-{{ arch }}-{{ checksum "isic-archive-gui/yarn.lock" }}
workflows:
version: 2
test_all:
jobs:
- serverTest
- webBuild_webTest
| ---
version: 2
jobs:
serverTest:
docker:
- image: isic/isic_test:latest
- image: circleci/mongo:3.6-ram
- command: ["mongod", "--storageEngine", "ephemeralForTest"]
+ command: ["mongod", "--storageEngine", "ephemeralForTest", "--dbpath", "/dev/shm/mongo"]
? ++++++++++++++++++++++++++++++
working_directory: /home/circleci/project
steps:
- checkout:
path: /home/circleci/project
- run:
name: Run Tox
command: tox
webBuild_webTest:
docker:
- image: isic/isic_test:latest
working_directory: /home/circleci/project
steps:
- checkout:
path: /home/circleci/project
- restore_cache:
key: yarn-{{ arch }}-{{ checksum "isic-archive-gui/yarn.lock" }}
- run:
name: Install ISIC GUI dependencies
command: yarn install
working_directory: isic-archive-gui
- run:
name: Build ISIC Admin GUI
command: yarn run build
working_directory: isic-archive-gui
- run:
name: Build ISIC Integration GUI
command: yarn run build:integration
working_directory: isic-archive-gui
- run:
name: Lint ISIC GUI
command: yarn run lint --no-fix
working_directory: isic-archive-gui
- save_cache:
paths: /home/circleci/.cache/yarn
key: yarn-{{ arch }}-{{ checksum "isic-archive-gui/yarn.lock" }}
workflows:
version: 2
test_all:
jobs:
- serverTest
- webBuild_webTest | 2 | 0.035714 | 1 | 1 |
e200bbe06b408e1ca3773ddeef81bb50e29cd432 | .atom/packages.cson | .atom/packages.cson | packages: [
"autocomplete-python"
"fold-functions"
"language-ini"
"linter"
"linter-pep8"
"linter-python-pep8"
"markdown-preview-plus"
"package-sync"
"python-indent"
]
| packages: [
"auto-detect-indentation"
"autocomplete-python"
"fold-functions"
"language-ini"
"linter"
"linter-pep8"
"linter-python-pep8"
"markdown-preview-plus"
"package-sync"
"python-indent"
]
| Add indentation auto detection to Atom | Add indentation auto detection to Atom
| CoffeeScript | mit | janwh/dotfiles | coffeescript | ## Code Before:
packages: [
"autocomplete-python"
"fold-functions"
"language-ini"
"linter"
"linter-pep8"
"linter-python-pep8"
"markdown-preview-plus"
"package-sync"
"python-indent"
]
## Instruction:
Add indentation auto detection to Atom
## Code After:
packages: [
"auto-detect-indentation"
"autocomplete-python"
"fold-functions"
"language-ini"
"linter"
"linter-pep8"
"linter-python-pep8"
"markdown-preview-plus"
"package-sync"
"python-indent"
]
| packages: [
+ "auto-detect-indentation"
"autocomplete-python"
"fold-functions"
"language-ini"
"linter"
"linter-pep8"
"linter-python-pep8"
"markdown-preview-plus"
"package-sync"
"python-indent"
] | 1 | 0.090909 | 1 | 0 |
c03879ac0227fab4a840b22c848552464b513591 | lib/generators/friendly_id_generator.rb | lib/generators/friendly_id_generator.rb | require 'rails/generators'
require 'rails/generators/migration'
# This generator adds a migration for the {FriendlyId::History
# FriendlyId::History} addon.
class FriendlyIdGenerator < Rails::Generators::Base
include Rails::Generators::Migration
source_root File.expand_path('../../friendly_id', __FILE__)
# Copies the migration template to db/migrate.
def copy_files(*args)
migration_template 'migration.rb', 'db/migrate/create_friendly_id_slugs.rb'
end
# TODO: use the module provided with Rails, no need to do this
# any more
# Taken from ActiveRecord's migration generator
def self.next_migration_number(dirname) #:nodoc:
if ActiveRecord::Base.timestamped_migrations
Time.now.utc.strftime("%Y%m%d%H%M%S")
else
"%.3d" % (current_migration_number(dirname) + 1)
end
end
end
| require 'rails/generators'
require "rails/generators/active_record"
# This generator adds a migration for the {FriendlyId::History
# FriendlyId::History} addon.
class FriendlyIdGenerator < Rails::Generators::Base
include Rails::Generators::Migration
extend ActiveRecord::Generators::Migration
source_root File.expand_path('../../friendly_id', __FILE__)
# Copies the migration template to db/migrate.
def copy_files(*args)
migration_template 'migration.rb', 'db/migrate/create_friendly_id_slugs.rb'
end
end
| Use Rails-provided module for migration numbering | Use Rails-provided module for migration numbering
| Ruby | mit | ashagnanasekar/friendly_id,kangkyu/friendly_id,morsedigital/friendly_id,onursarikaya/friendly_id,zenhacks/friendly_id,Pathgather/friendly_id,sideci-sample/sideci-sample-friendly_id,joshsoftware/friendly_id,norman/friendly_id,ThanhKhoaIT/friendly_id,MatthewRDodds/friendly_id,Empact/friendly_id,tekin/friendly_id,MAubreyK/friendly_id | ruby | ## Code Before:
require 'rails/generators'
require 'rails/generators/migration'
# This generator adds a migration for the {FriendlyId::History
# FriendlyId::History} addon.
class FriendlyIdGenerator < Rails::Generators::Base
include Rails::Generators::Migration
source_root File.expand_path('../../friendly_id', __FILE__)
# Copies the migration template to db/migrate.
def copy_files(*args)
migration_template 'migration.rb', 'db/migrate/create_friendly_id_slugs.rb'
end
# TODO: use the module provided with Rails, no need to do this
# any more
# Taken from ActiveRecord's migration generator
def self.next_migration_number(dirname) #:nodoc:
if ActiveRecord::Base.timestamped_migrations
Time.now.utc.strftime("%Y%m%d%H%M%S")
else
"%.3d" % (current_migration_number(dirname) + 1)
end
end
end
## Instruction:
Use Rails-provided module for migration numbering
## Code After:
require 'rails/generators'
require "rails/generators/active_record"
# This generator adds a migration for the {FriendlyId::History
# FriendlyId::History} addon.
class FriendlyIdGenerator < Rails::Generators::Base
include Rails::Generators::Migration
extend ActiveRecord::Generators::Migration
source_root File.expand_path('../../friendly_id', __FILE__)
# Copies the migration template to db/migrate.
def copy_files(*args)
migration_template 'migration.rb', 'db/migrate/create_friendly_id_slugs.rb'
end
end
| require 'rails/generators'
- require 'rails/generators/migration'
? ^ ---- ^^
+ require "rails/generators/active_record"
? ^ + ++++++ ^^^
# This generator adds a migration for the {FriendlyId::History
# FriendlyId::History} addon.
class FriendlyIdGenerator < Rails::Generators::Base
include Rails::Generators::Migration
+ extend ActiveRecord::Generators::Migration
source_root File.expand_path('../../friendly_id', __FILE__)
# Copies the migration template to db/migrate.
def copy_files(*args)
migration_template 'migration.rb', 'db/migrate/create_friendly_id_slugs.rb'
end
- # TODO: use the module provided with Rails, no need to do this
- # any more
- # Taken from ActiveRecord's migration generator
- def self.next_migration_number(dirname) #:nodoc:
- if ActiveRecord::Base.timestamped_migrations
- Time.now.utc.strftime("%Y%m%d%H%M%S")
- else
- "%.3d" % (current_migration_number(dirname) + 1)
- end
- end
end | 13 | 0.5 | 2 | 11 |
9c0e0f605910ac6c51bb522eef30014814759a56 | app.js | app.js | var express = require('express');
var bodyParser = require('body-parser');
var http = require('http');
var path = require('path');
var fs = require('fs');
var quotes = require(path.join(__dirname, 'lib', 'quotes'))
var app = express();
app.set('port', 3000);
app.set('views', path.join(__dirname, 'views'));
app.set('view engine', 'jade');
app.use(express.static(path.join(__dirname, 'public')));
app.get('/', function(req, res) {
res.render('index');
});
app.get('/api/listAllOggFiles/', function(req, res) {
fs.readdir('public/sounds/', function(err, files){
if (err) {
throw err;
}
var filenames = [];
for (var i = 0; i < files.length; i++) {
filenames.push(files[i].split('.')[0]);
}
res.json(filenames);
});
});
app.get('/api/getQuotes/', function(req, res){
res.json(quotes);
});
http.createServer(app).listen(app.get('port'), function(){
console.log('Express server up and running. Listening on port ' + app.get('port'));
});
| var express = require('express');
var bodyParser = require('body-parser');
var http = require('http');
var path = require('path');
var fs = require('fs');
var quotes = require(path.join(__dirname, 'lib', 'quotes'))
var app = express();
app.set('port', (process.env.PORT || 5000));
app.set('views', path.join(__dirname, 'views'));
app.set('view engine', 'jade');
app.use(express.static(path.join(__dirname, 'public')));
app.get('/', function(req, res) {
res.render('index');
});
app.get('/api/listAllOggFiles/', function(req, res) {
fs.readdir('public/sounds/', function(err, files){
if (err) {
throw err;
}
var filenames = [];
for (var i = 0; i < files.length; i++) {
filenames.push(files[i].split('.')[0]);
}
res.json(filenames);
});
});
app.get('/api/getQuotes/', function(req, res){
res.json(quotes);
});
http.createServer(app).listen(app.get('port'), function(){
console.log('Express server up and running. Listening on port ' + app.get('port'));
});
| Change listening port to env port with default. | Change listening port to env port with default.
| JavaScript | mit | Gweylorth/Poulett.es,Gweylorth/Poulett.es | javascript | ## Code Before:
var express = require('express');
var bodyParser = require('body-parser');
var http = require('http');
var path = require('path');
var fs = require('fs');
var quotes = require(path.join(__dirname, 'lib', 'quotes'))
var app = express();
app.set('port', 3000);
app.set('views', path.join(__dirname, 'views'));
app.set('view engine', 'jade');
app.use(express.static(path.join(__dirname, 'public')));
app.get('/', function(req, res) {
res.render('index');
});
app.get('/api/listAllOggFiles/', function(req, res) {
fs.readdir('public/sounds/', function(err, files){
if (err) {
throw err;
}
var filenames = [];
for (var i = 0; i < files.length; i++) {
filenames.push(files[i].split('.')[0]);
}
res.json(filenames);
});
});
app.get('/api/getQuotes/', function(req, res){
res.json(quotes);
});
http.createServer(app).listen(app.get('port'), function(){
console.log('Express server up and running. Listening on port ' + app.get('port'));
});
## Instruction:
Change listening port to env port with default.
## Code After:
var express = require('express');
var bodyParser = require('body-parser');
var http = require('http');
var path = require('path');
var fs = require('fs');
var quotes = require(path.join(__dirname, 'lib', 'quotes'))
var app = express();
app.set('port', (process.env.PORT || 5000));
app.set('views', path.join(__dirname, 'views'));
app.set('view engine', 'jade');
app.use(express.static(path.join(__dirname, 'public')));
app.get('/', function(req, res) {
res.render('index');
});
app.get('/api/listAllOggFiles/', function(req, res) {
fs.readdir('public/sounds/', function(err, files){
if (err) {
throw err;
}
var filenames = [];
for (var i = 0; i < files.length; i++) {
filenames.push(files[i].split('.')[0]);
}
res.json(filenames);
});
});
app.get('/api/getQuotes/', function(req, res){
res.json(quotes);
});
http.createServer(app).listen(app.get('port'), function(){
console.log('Express server up and running. Listening on port ' + app.get('port'));
});
| var express = require('express');
var bodyParser = require('body-parser');
var http = require('http');
var path = require('path');
var fs = require('fs');
var quotes = require(path.join(__dirname, 'lib', 'quotes'))
var app = express();
- app.set('port', 3000);
+ app.set('port', (process.env.PORT || 5000));
app.set('views', path.join(__dirname, 'views'));
app.set('view engine', 'jade');
app.use(express.static(path.join(__dirname, 'public')));
app.get('/', function(req, res) {
res.render('index');
});
app.get('/api/listAllOggFiles/', function(req, res) {
fs.readdir('public/sounds/', function(err, files){
if (err) {
throw err;
}
var filenames = [];
for (var i = 0; i < files.length; i++) {
filenames.push(files[i].split('.')[0]);
}
res.json(filenames);
});
});
app.get('/api/getQuotes/', function(req, res){
res.json(quotes);
});
http.createServer(app).listen(app.get('port'), function(){
console.log('Express server up and running. Listening on port ' + app.get('port'));
}); | 2 | 0.05 | 1 | 1 |
9decb86e8dc832c37df05c0bbe84f81384bc0e91 | .travis.yml | .travis.yml | language: objective-c
before_install:
- brew update
- gem instal cocoapods
- cd Tests && pod install && cd $TRAVIS_BUILD_DIR
script: rake test
| language: objective-c
env:
- OS=6.1
- OS=7.0
before_install:
- brew update
- bundle install
- cd Tests && pod install && cd $TRAVIS_BUILD_DIR
script: rake test
| Add OS ENV variable for running test under iOS 6.1 and 7.0 | Add OS ENV variable for running test under iOS 6.1 and 7.0 | YAML | mit | xing/XNGAPIClient,Ruenzuo/XNGAPIClient,lrrrs/XNGAPIClient | yaml | ## Code Before:
language: objective-c
before_install:
- brew update
- gem instal cocoapods
- cd Tests && pod install && cd $TRAVIS_BUILD_DIR
script: rake test
## Instruction:
Add OS ENV variable for running test under iOS 6.1 and 7.0
## Code After:
language: objective-c
env:
- OS=6.1
- OS=7.0
before_install:
- brew update
- bundle install
- cd Tests && pod install && cd $TRAVIS_BUILD_DIR
script: rake test
| language: objective-c
+ env:
+ - OS=6.1
+ - OS=7.0
before_install:
- brew update
- - gem instal cocoapods
+ - bundle install
- cd Tests && pod install && cd $TRAVIS_BUILD_DIR
script: rake test | 5 | 0.833333 | 4 | 1 |
587f490d0b81c06a88fe9777d306c277ad688925 | src/sql/insert_lock.sql | src/sql/insert_lock.sql | -- $Header$
DROP TABLE pgpool_catalog.insert_lock;
CREATE SCHEMA pgpool_catalog;
CREATE TABLE pgpool_catalog.insert_lock(reloid OID PRIMARY KEY);
-- this row is used as the row lock target when pgpool inserts new oid
INSERT INTO pgpool_catalog.insert_lock VALUES (0);
-- allow "SELECT ... FOR UPDATE" and "INSERT ..." to all roles
GRANT SELECT ON pgpool_catalog.insert_lock TO PUBLIC;
GRANT UPDATE ON pgpool_catalog.insert_lock TO PUBLIC;
GRANT INSERT ON pgpool_catalog.insert_lock TO PUBLIC;
| -- Create lock control table for tables using sequence in native replication mode.
DROP TABLE pgpool_catalog.insert_lock;
CREATE SCHEMA pgpool_catalog;
CREATE TABLE pgpool_catalog.insert_lock(reloid OID PRIMARY KEY);
-- this row is used as the row lock target when pgpool inserts new oid
INSERT INTO pgpool_catalog.insert_lock VALUES (0);
-- allow "SELECT ... FOR UPDATE" and "INSERT ..." to all roles
GRANT USAGE ON SCHEMA pgpool_catalog TO PUBLIC;
GRANT SELECT ON pgpool_catalog.insert_lock TO PUBLIC;
GRANT UPDATE ON pgpool_catalog.insert_lock TO PUBLIC;
GRANT INSERT ON pgpool_catalog.insert_lock TO PUBLIC;
| Fix that the script forgets to allow public access to pgpool_catalog. | Fix that the script forgets to allow public access to pgpool_catalog.
The bug prevents inserting data into user tables if pgpool_catalog is
created in native replication mode. The bug was there from day 1. I
wonder why nobody noticed until today. Per [pgpool-general-jp: 1229].
| SQL | apache-2.0 | treasure-data/prestogres,treasure-data/prestogres,treasure-data/prestogres,treasure-data/prestogres,treasure-data/prestogres,treasure-data/prestogres,treasure-data/prestogres | sql | ## Code Before:
-- $Header$
DROP TABLE pgpool_catalog.insert_lock;
CREATE SCHEMA pgpool_catalog;
CREATE TABLE pgpool_catalog.insert_lock(reloid OID PRIMARY KEY);
-- this row is used as the row lock target when pgpool inserts new oid
INSERT INTO pgpool_catalog.insert_lock VALUES (0);
-- allow "SELECT ... FOR UPDATE" and "INSERT ..." to all roles
GRANT SELECT ON pgpool_catalog.insert_lock TO PUBLIC;
GRANT UPDATE ON pgpool_catalog.insert_lock TO PUBLIC;
GRANT INSERT ON pgpool_catalog.insert_lock TO PUBLIC;
## Instruction:
Fix that the script forgets to allow public access to pgpool_catalog.
The bug prevents inserting data into user tables if pgpool_catalog is
created in native replication mode. The bug was there from day 1. I
wonder why nobody noticed until today. Per [pgpool-general-jp: 1229].
## Code After:
-- Create lock control table for tables using sequence in native replication mode.
DROP TABLE pgpool_catalog.insert_lock;
CREATE SCHEMA pgpool_catalog;
CREATE TABLE pgpool_catalog.insert_lock(reloid OID PRIMARY KEY);
-- this row is used as the row lock target when pgpool inserts new oid
INSERT INTO pgpool_catalog.insert_lock VALUES (0);
-- allow "SELECT ... FOR UPDATE" and "INSERT ..." to all roles
GRANT USAGE ON SCHEMA pgpool_catalog TO PUBLIC;
GRANT SELECT ON pgpool_catalog.insert_lock TO PUBLIC;
GRANT UPDATE ON pgpool_catalog.insert_lock TO PUBLIC;
GRANT INSERT ON pgpool_catalog.insert_lock TO PUBLIC;
| - -- $Header$
+ -- Create lock control table for tables using sequence in native replication mode.
DROP TABLE pgpool_catalog.insert_lock;
CREATE SCHEMA pgpool_catalog;
CREATE TABLE pgpool_catalog.insert_lock(reloid OID PRIMARY KEY);
-- this row is used as the row lock target when pgpool inserts new oid
INSERT INTO pgpool_catalog.insert_lock VALUES (0);
-- allow "SELECT ... FOR UPDATE" and "INSERT ..." to all roles
+ GRANT USAGE ON SCHEMA pgpool_catalog TO PUBLIC;
GRANT SELECT ON pgpool_catalog.insert_lock TO PUBLIC;
GRANT UPDATE ON pgpool_catalog.insert_lock TO PUBLIC;
GRANT INSERT ON pgpool_catalog.insert_lock TO PUBLIC; | 3 | 0.214286 | 2 | 1 |
85844cb14186da50bddd6009b29746c5bb276fc2 | app/assets/javascripts/models/user.js.coffee | app/assets/javascripts/models/user.js.coffee | define [
'chaplin'
'models/base/model'
'models/tags'
], (Chaplin, Model, Tags) ->
'use strict'
class User extends Model
validation:
username:
required: true
email:
pattern: 'email',
required: true
email_confirmed:
fn: 'emailConfirmed'
required: true
password: required: false
defaults: {
accepted_currencies: []
}
initialize: (options) ->
super
@acceptedCurrenciesCollection = new Tags
@listenTo @, "change:accepted_currencies", @updateAcceptedCurrenciesCollection
@updateAcceptedCurrenciesCollection()
urlRoot: ->
Chaplin.mediator.api_base_url + "/users"
sign_in: (options) ->
options ||= {}
options.url = @urlRoot() + '/sign_in'
@save(@attributes, options)
full_name: =>
name = ''
if @get('given_name')?
name += @get('given_name')
if @get('surname')?
name += ' ' + @get('surname')
if name.length == 0
name = "unknown"
$.trim(name)
emailConfirmed: (value, attr, computed) ->
if computed.email? && computed.email_confirmation?
unless computed.email == computed.email_confirmation
return "Does not match email"
updateAcceptedCurrenciesCollection: ->
@acceptedCurrenciesCollection.set(@get('accepted_currencies'), parse: true)
| define [
'chaplin'
'models/base/model'
'models/tags'
], (Chaplin, Model, Tags) ->
'use strict'
class User extends Model
validation:
username:
required: true
email:
pattern: 'email',
required: true
email_confirmation:
fn: 'emailConfirmed'
password: required: false
defaults: {
accepted_currencies: []
}
initialize: (options) ->
super
@acceptedCurrenciesCollection = new Tags
@listenTo @, "change:accepted_currencies", @updateAcceptedCurrenciesCollection
@updateAcceptedCurrenciesCollection()
urlRoot: ->
Chaplin.mediator.api_base_url + "/users"
sign_in: (options) ->
options ||= {}
options.url = @urlRoot() + '/sign_in'
@save(@attributes, options)
full_name: =>
name = ''
if @get('given_name')?
name += @get('given_name')
if @get('surname')?
name += ' ' + @get('surname')
if name.length == 0
name = "unknown"
$.trim(name)
emailConfirmed: (value, attr, computed) ->
if computed.email? && computed.email_confirmation?
unless computed.email == computed.email_confirmation
return "Does not match email"
updateAcceptedCurrenciesCollection: ->
@acceptedCurrenciesCollection.set(@get('accepted_currencies'), parse: true)
| Fix broken client side validation. | Fix broken client side validation.
| CoffeeScript | agpl-3.0 | joatuapp/joatu-app,joatuapp/joatu-app | coffeescript | ## Code Before:
define [
'chaplin'
'models/base/model'
'models/tags'
], (Chaplin, Model, Tags) ->
'use strict'
class User extends Model
validation:
username:
required: true
email:
pattern: 'email',
required: true
email_confirmed:
fn: 'emailConfirmed'
required: true
password: required: false
defaults: {
accepted_currencies: []
}
initialize: (options) ->
super
@acceptedCurrenciesCollection = new Tags
@listenTo @, "change:accepted_currencies", @updateAcceptedCurrenciesCollection
@updateAcceptedCurrenciesCollection()
urlRoot: ->
Chaplin.mediator.api_base_url + "/users"
sign_in: (options) ->
options ||= {}
options.url = @urlRoot() + '/sign_in'
@save(@attributes, options)
full_name: =>
name = ''
if @get('given_name')?
name += @get('given_name')
if @get('surname')?
name += ' ' + @get('surname')
if name.length == 0
name = "unknown"
$.trim(name)
emailConfirmed: (value, attr, computed) ->
if computed.email? && computed.email_confirmation?
unless computed.email == computed.email_confirmation
return "Does not match email"
updateAcceptedCurrenciesCollection: ->
@acceptedCurrenciesCollection.set(@get('accepted_currencies'), parse: true)
## Instruction:
Fix broken client side validation.
## Code After:
define [
'chaplin'
'models/base/model'
'models/tags'
], (Chaplin, Model, Tags) ->
'use strict'
class User extends Model
validation:
username:
required: true
email:
pattern: 'email',
required: true
email_confirmation:
fn: 'emailConfirmed'
password: required: false
defaults: {
accepted_currencies: []
}
initialize: (options) ->
super
@acceptedCurrenciesCollection = new Tags
@listenTo @, "change:accepted_currencies", @updateAcceptedCurrenciesCollection
@updateAcceptedCurrenciesCollection()
urlRoot: ->
Chaplin.mediator.api_base_url + "/users"
sign_in: (options) ->
options ||= {}
options.url = @urlRoot() + '/sign_in'
@save(@attributes, options)
full_name: =>
name = ''
if @get('given_name')?
name += @get('given_name')
if @get('surname')?
name += ' ' + @get('surname')
if name.length == 0
name = "unknown"
$.trim(name)
emailConfirmed: (value, attr, computed) ->
if computed.email? && computed.email_confirmation?
unless computed.email == computed.email_confirmation
return "Does not match email"
updateAcceptedCurrenciesCollection: ->
@acceptedCurrenciesCollection.set(@get('accepted_currencies'), parse: true)
| define [
'chaplin'
'models/base/model'
'models/tags'
], (Chaplin, Model, Tags) ->
'use strict'
class User extends Model
validation:
username:
required: true
email:
pattern: 'email',
required: true
- email_confirmed:
? ^^
+ email_confirmation:
? ^^^^^
fn: 'emailConfirmed'
- required: true
password: required: false
defaults: {
accepted_currencies: []
}
initialize: (options) ->
super
@acceptedCurrenciesCollection = new Tags
@listenTo @, "change:accepted_currencies", @updateAcceptedCurrenciesCollection
@updateAcceptedCurrenciesCollection()
urlRoot: ->
Chaplin.mediator.api_base_url + "/users"
sign_in: (options) ->
options ||= {}
options.url = @urlRoot() + '/sign_in'
@save(@attributes, options)
full_name: =>
name = ''
if @get('given_name')?
name += @get('given_name')
if @get('surname')?
name += ' ' + @get('surname')
if name.length == 0
name = "unknown"
$.trim(name)
emailConfirmed: (value, attr, computed) ->
if computed.email? && computed.email_confirmation?
unless computed.email == computed.email_confirmation
return "Does not match email"
updateAcceptedCurrenciesCollection: ->
@acceptedCurrenciesCollection.set(@get('accepted_currencies'), parse: true) | 3 | 0.051724 | 1 | 2 |
4c86cf0841e4160fb36d896705a1483da834c92c | .travis.yml | .travis.yml | language: node_js
node_js:
- 0.10
notifications:
email:
- medikoo+eregistrations@medikoo.com
script: "npm run lint-ci && npm run css-lint-ci"
| language: node_js
node_js:
- 0.10
notifications:
email:
- medikoo+eregistrations@medikoo.com
script: "npm test && npm run lint-ci && npm run css-lint-ci"
| Add npm test to Travis flow | Add npm test to Travis flow
| YAML | mit | egovernment/eregistrations,egovernment/eregistrations,egovernment/eregistrations | yaml | ## Code Before:
language: node_js
node_js:
- 0.10
notifications:
email:
- medikoo+eregistrations@medikoo.com
script: "npm run lint-ci && npm run css-lint-ci"
## Instruction:
Add npm test to Travis flow
## Code After:
language: node_js
node_js:
- 0.10
notifications:
email:
- medikoo+eregistrations@medikoo.com
script: "npm test && npm run lint-ci && npm run css-lint-ci"
| language: node_js
node_js:
- 0.10
notifications:
email:
- medikoo+eregistrations@medikoo.com
- script: "npm run lint-ci && npm run css-lint-ci"
+ script: "npm test && npm run lint-ci && npm run css-lint-ci"
? ++++++++++++
| 2 | 0.222222 | 1 | 1 |
62c63dceed47d68cdc3cc3f0b8e2928df6bc9265 | sto-areas/index.js | sto-areas/index.js | import major from './major'
import concentration from './concentration'
import emphasis from './emphasis'
import degree from './degree'
import _ from 'lodash'
let areas = [major, concentration, emphasis, degree]
let allAreas = _(areas).map(_.toArray).flatten().value()
// console.log(allAreas)
/**
* Finds an area of study.
*
* @param {Object} args - a lodash filter object.
* @returns {Object} - An area of study.
*/
let find = (args={}) => {
return _.find(allAreas, args)
}
export {major, concentration, emphasis, degree, find}
| import _ from 'lodash'
import major from './major'
import concentration from './concentration'
import emphasis from './emphasis'
import degree from './degree'
let areas = [major, concentration, emphasis, degree]
let allAreas = _(areas)
.map(_.toArray)
.flatten()
.value()
let areaNotFound = {
title: 'Not Found',
years: [null, null],
id: 'a-notfound',
type: 'not-found',
departmentAbbr: 'NOTFOUND',
check: null,
}
Object.freeze(areaNotFound)
/**
* Finds an area of study.
*
* @param {String} id - the id to find.
* @param {Number} yearOfGraduation - the year the student matriculated.
* @returns {Object} - an area of study.
*/
let findAreaOfStudy = (id, yearOfGraduation) => {
let area = _.find(allAreas, (area) => {
if (!area.id || area.id !== id)
return false
if (!area.years)
return false
let [startYear, endYear] = area.years
let yearIsBetween = false
if (_.isNull(endYear) && startYear <= yearOfGraduation)
yearIsBetween = true
else if (endYear >= yearOfGraduation && startYear <= yearOfGraduation)
yearIsBetween = true
return yearIsBetween
})
return area || areaNotFound
}
export default findAreaOfStudy
| Check for years when finding areas | Check for years when finding areas
| JavaScript | agpl-3.0 | hawkrives/gobbldygook,hawkrives/gobbldygook,hawkrives/gobbldygook | javascript | ## Code Before:
import major from './major'
import concentration from './concentration'
import emphasis from './emphasis'
import degree from './degree'
import _ from 'lodash'
let areas = [major, concentration, emphasis, degree]
let allAreas = _(areas).map(_.toArray).flatten().value()
// console.log(allAreas)
/**
* Finds an area of study.
*
* @param {Object} args - a lodash filter object.
* @returns {Object} - An area of study.
*/
let find = (args={}) => {
return _.find(allAreas, args)
}
export {major, concentration, emphasis, degree, find}
## Instruction:
Check for years when finding areas
## Code After:
import _ from 'lodash'
import major from './major'
import concentration from './concentration'
import emphasis from './emphasis'
import degree from './degree'
let areas = [major, concentration, emphasis, degree]
let allAreas = _(areas)
.map(_.toArray)
.flatten()
.value()
let areaNotFound = {
title: 'Not Found',
years: [null, null],
id: 'a-notfound',
type: 'not-found',
departmentAbbr: 'NOTFOUND',
check: null,
}
Object.freeze(areaNotFound)
/**
* Finds an area of study.
*
* @param {String} id - the id to find.
* @param {Number} yearOfGraduation - the year the student matriculated.
* @returns {Object} - an area of study.
*/
let findAreaOfStudy = (id, yearOfGraduation) => {
let area = _.find(allAreas, (area) => {
if (!area.id || area.id !== id)
return false
if (!area.years)
return false
let [startYear, endYear] = area.years
let yearIsBetween = false
if (_.isNull(endYear) && startYear <= yearOfGraduation)
yearIsBetween = true
else if (endYear >= yearOfGraduation && startYear <= yearOfGraduation)
yearIsBetween = true
return yearIsBetween
})
return area || areaNotFound
}
export default findAreaOfStudy
| + import _ from 'lodash'
+
import major from './major'
import concentration from './concentration'
import emphasis from './emphasis'
import degree from './degree'
- import _ from 'lodash'
let areas = [major, concentration, emphasis, degree]
- let allAreas = _(areas).map(_.toArray).flatten().value()
- // console.log(allAreas)
+ let allAreas = _(areas)
+ .map(_.toArray)
+ .flatten()
+ .value()
+
+ let areaNotFound = {
+ title: 'Not Found',
+ years: [null, null],
+ id: 'a-notfound',
+ type: 'not-found',
+ departmentAbbr: 'NOTFOUND',
+ check: null,
+ }
+ Object.freeze(areaNotFound)
/**
* Finds an area of study.
*
- * @param {Object} args - a lodash filter object.
+ * @param {String} id - the id to find.
+ * @param {Number} yearOfGraduation - the year the student matriculated.
- * @returns {Object} - An area of study.
? ^
+ * @returns {Object} - an area of study.
? ^
*/
- let find = (args={}) => {
- return _.find(allAreas, args)
+ let findAreaOfStudy = (id, yearOfGraduation) => {
+ let area = _.find(allAreas, (area) => {
+ if (!area.id || area.id !== id)
+ return false
+
+ if (!area.years)
+ return false
+
+ let [startYear, endYear] = area.years
+
+ let yearIsBetween = false
+ if (_.isNull(endYear) && startYear <= yearOfGraduation)
+ yearIsBetween = true
+ else if (endYear >= yearOfGraduation && startYear <= yearOfGraduation)
+ yearIsBetween = true
+
+ return yearIsBetween
+ })
+
+ return area || areaNotFound
}
- export {major, concentration, emphasis, degree, find}
+ export default findAreaOfStudy | 48 | 2.285714 | 40 | 8 |
ed6a29a5f04734b5bb6f8c707715de42b14a2aec | src/extract.js | src/extract.js |
var esprima = require('esprima');
/**
* Position in the source file.
*
* @see [SpiderMonkey Parser API]
* {@link https://developer.mozilla.org/en-US/docs/Mozilla/Projects/SpiderMonkey/Parser_API}
*
* @typedef {Object} Position
* @property {number} line - Line number, 1-indexed.
* @property {?number} column - Column number, 0-indexed.
*/
/**
* Location - position range.
*
* @typedef {Object} Location
* @property {Position} start
* @property {Position} end
*/
/**
* Extract comments from code.
* Each comment is represented by a string within its usual comment delimiters.
* Returned location always contains both line and column info.
*
* @arg {string} code - JavaScript code.
* @return {{text: string, loc: Location}[]}
*/
var extract = function (code) {
return esprima.parse(code, {
comment: true,
loc: true
}).comments.map(function (comment) {
return {
text: (comment.type == 'Line')
? '//' + comment.value
: '/*' + comment.value + '*/',
loc: comment.loc
};
});
};
module.exports = extract;
| var esprima = require('esprima');
/**
* Position in the source file.
*
* @see [SpiderMonkey Parser API]
* {@link https://developer.mozilla.org/en-US/docs/Mozilla/Projects/SpiderMonkey/Parser_API}
*
* @typedef {Object} Position
* @property {number} line - Line number, 1-indexed.
* @property {?number} column - Column number, 0-indexed.
*/
/**
* Location - position range.
*
* @typedef {Object} Location
* @property {Position} start - Column number must be set.
* @property {Position} end - Column number must be set.
*/
/**
* Extract comments from code.
* Each comment is represented by a string within its usual comment delimiters.
*
* @arg {string} code - JavaScript code.
* @return {{text: string, loc: Location}[]}
*/
var extract = function (code) {
return esprima.parse(code, {
comment: true,
loc: true
}).comments.map(function (comment) {
return {
text: (comment.type == 'Line')
? '//' + comment.value
: '/*' + comment.value + '*/',
loc: comment.loc
};
});
};
module.exports = extract;
| Make column numbers in Location mandatory | Make column numbers in Location mandatory
| JavaScript | mit | eush77/js-comment-check | javascript | ## Code Before:
var esprima = require('esprima');
/**
* Position in the source file.
*
* @see [SpiderMonkey Parser API]
* {@link https://developer.mozilla.org/en-US/docs/Mozilla/Projects/SpiderMonkey/Parser_API}
*
* @typedef {Object} Position
* @property {number} line - Line number, 1-indexed.
* @property {?number} column - Column number, 0-indexed.
*/
/**
* Location - position range.
*
* @typedef {Object} Location
* @property {Position} start
* @property {Position} end
*/
/**
* Extract comments from code.
* Each comment is represented by a string within its usual comment delimiters.
* Returned location always contains both line and column info.
*
* @arg {string} code - JavaScript code.
* @return {{text: string, loc: Location}[]}
*/
var extract = function (code) {
return esprima.parse(code, {
comment: true,
loc: true
}).comments.map(function (comment) {
return {
text: (comment.type == 'Line')
? '//' + comment.value
: '/*' + comment.value + '*/',
loc: comment.loc
};
});
};
module.exports = extract;
## Instruction:
Make column numbers in Location mandatory
## Code After:
var esprima = require('esprima');
/**
* Position in the source file.
*
* @see [SpiderMonkey Parser API]
* {@link https://developer.mozilla.org/en-US/docs/Mozilla/Projects/SpiderMonkey/Parser_API}
*
* @typedef {Object} Position
* @property {number} line - Line number, 1-indexed.
* @property {?number} column - Column number, 0-indexed.
*/
/**
* Location - position range.
*
* @typedef {Object} Location
* @property {Position} start - Column number must be set.
* @property {Position} end - Column number must be set.
*/
/**
* Extract comments from code.
* Each comment is represented by a string within its usual comment delimiters.
*
* @arg {string} code - JavaScript code.
* @return {{text: string, loc: Location}[]}
*/
var extract = function (code) {
return esprima.parse(code, {
comment: true,
loc: true
}).comments.map(function (comment) {
return {
text: (comment.type == 'Line')
? '//' + comment.value
: '/*' + comment.value + '*/',
loc: comment.loc
};
});
};
module.exports = extract;
| -
var esprima = require('esprima');
/**
* Position in the source file.
*
* @see [SpiderMonkey Parser API]
* {@link https://developer.mozilla.org/en-US/docs/Mozilla/Projects/SpiderMonkey/Parser_API}
*
* @typedef {Object} Position
* @property {number} line - Line number, 1-indexed.
* @property {?number} column - Column number, 0-indexed.
*/
/**
* Location - position range.
*
* @typedef {Object} Location
- * @property {Position} start
- * @property {Position} end
+ * @property {Position} start - Column number must be set.
+ * @property {Position} end - Column number must be set.
*/
/**
* Extract comments from code.
* Each comment is represented by a string within its usual comment delimiters.
- * Returned location always contains both line and column info.
*
* @arg {string} code - JavaScript code.
* @return {{text: string, loc: Location}[]}
*/
var extract = function (code) {
return esprima.parse(code, {
comment: true,
loc: true
}).comments.map(function (comment) {
return {
text: (comment.type == 'Line')
? '//' + comment.value
: '/*' + comment.value + '*/',
loc: comment.loc
};
});
};
module.exports = extract; | 6 | 0.122449 | 2 | 4 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.